Unnamed: 0
int64 0
7.24k
| id
int64 1
7.28k
| raw_text
stringlengths 9
124k
| vw_text
stringlengths 12
15k
|
---|---|---|---|
400 | 1,365 | 2D Observers for Human 3D Object Recognition?
Zili Liu
NEC Research Institute
Daniel Kersten
University of Minnesota
. Abstract
Converging evidence has shown that human object recognition
depends on familiarity with the images of an object. Further,
the greater the similarity between objects, the stronger is the
dependence on object appearance, and the more important twodimensional (2D) image information becomes. These findings, however, do not rule out the use of 3D structural information in recognition, and the degree to which 3D information is used in visual
memory is an important issue. Liu, Knill, & Kersten (1995) showed
that any model that is restricted to rotations in the image plane
of independent 2D templates could not account for human performance in discriminating novel object views. We now present results
from models of generalized radial basis functions (GRBF), 2D nearest neighbor matching that allows 2D affine transformations, and
a Bayesian statistical estimator that integrates over all possible 2D
affine transformations. The performance of the human observers
relative to each of the models is better for the novel views than
for the familiar template views, suggesting that humans generalize
better to novel views from template views. The Bayesian estimator yields the optimal performance with 2D affine transformations
and independent 2D templates. Therefore, models of 2D affine
matching operations with independent 2D templates are unlikely
to account for human recognition performance.
1 Introduction
Object recognition is one of the most important functions in human vision. To
understand human object recognition, it is essential to understand how objects are
represented in human visual memory. A central component in object recognition
is the matching of the stored object representation with that derived from the image input. But the nature of the object representation has to be inferred from
recognition performance, by taking into account the contribution from the image
information. When evaluating human performance, how can one separate the con-
830
Z Liu and D. Kersten
tributions to performance of the image information from the representation? Ideal
observer analysis provides a precise computational tool to answer this question. An
ideal observer's recognition performance is restricted only by the available image
information and is otherwise optimal, in the sense of statistical decision theory,
irrespective of how the model is implemented. A comparison of human to ideal
performance (often in terms of efficiency) serves to normalize performance with respect to the image information for the task. We consider the problem of viewpoint
dependence in human recognition.
A recent debate in human object recognition has focused on the dependence of recognition performance on viewpoint [1 , 6]. Depending on the experimental conditions,
an observer's ability to recognize a familiar object from novel viewpoints is impaired
to varying degrees. A central assumption in the debate is the equivalence in viewpoint dependence and recognition performance. In other words, the assumption is
that viewpoint dependent performance implies a viewpoint dependent representation, and that viewpoint independent performance implies a viewpoint independent
representation. However, given that any recognition performance depends on the
input image information, which is necessarily viewpoint dependent, the viewpoint
dependence of the performance is neither necessary nor sufficient for the viewpoint
dependence of the representation. Image information has to be factored out first,
and the ideal observer provides the means to do this.
The second aspect of an ideal observer is that it is implementation free. Consider the GRBF model [5], as compared with human object recognition (see below). The model stores a number of 2D templates {Ti} of a 3D object 0,
and reco~nizes or rejects a stimulus image S by the following similarity measure
~iCi exp UITi - SI1 2 j2( 2 ), where Ci and a are constants. The model's performance
as a function of viewpoint parallels that of human observers. This observation has
led to the conclusion that the human visual system may indeed, as does the model,
use 2D stored views with GRBF interpolation to recognize 3D objects [2]. Such a
conclusion, however, overlooks implementational constraints in the model, because
the model's performance also depends on its implementations. Conceivably, a model
with some 3D information of the objects can also mimic human performance, so
long as it is appropriately implemented. There are typically too many possible
models that can produce the same pattern of results.
In contrast, an ideal observer computes the optimal performance that is only limited
by the stimulus information and the task. We can define constrained ideals that are
also limited by explicitly specified assumptions (e.g., a class of matching operations).
Such a model observer therefore yields the best possible performance among the
class of models with the same stimulus input and assumptions. In this paper,
we are particularly interested in constrained ideal observers that are restricted in
functionally Significant aspects (e.g., a 2D ideal observer that stores independent
2D templates and has access only to 2D affine transformations) . The key idea is
that a constrained ideal observer is the best in its class. So if humans outperform
this ideal observer, they must have used more than what is available to the ideal.
The conclusion that follows is strong: not only does the constrained ideal fail to
account for human performance, but the whole class of its implementations are also
falsified.
A crucial question in object recognition is the extent to which human observers
model the geometric variation in images due to the projection of a 3D object onto a
2D image. At one extreme, we have shown that any model that compares the image
to independent views (even if we allow for 2D rigid transformations of the input
image) is insufficient to account for human performance. At the other extreme, it
is unlikely that variation is modeled in terms of rigid transformation of a 3D object
2D Observers/or Hwnan 3D Object Recognition?
831
template in memory. A possible intermediate solution is to match the input image
to stored views, subject to 2D affine deformations. This is reasonable because 2D
affine transformations approximate 3D variation over a limited range of viewpoint
change.
In this study, we test whether any model limited to the independent comparison
of 2D views, but with 2D affine flexibility, is sufficient to account for viewpoint
dependence in human recognition. In the following section, we first define our experimental task, in which the computational models yield the provably best possible
performance under their specified conditions. We then review the 2D ideal observer
and GRBF model derived in [4], and the 2D affine nearest neighbor model in [8].
Our principal theoretical result is a closed-form solution of a Bayesian 2D affine ideal
observer. We then compare human performance with the 2D affine ideal model, as
well as the other three models. In particular, if humans can classify novel views of
an object better than the 2D affine ideal, then our human observers must have used
more information than that embodied by that ideal.
2 The observers
Let us first define the task. An observer looks at the 2D images of a 3D wire
frame object from a number of viewpoints. These images will be called templates
{Td. Then two distorted copies of the original 3D object are displayed. They
are obtained by adding 3D Gaussian positional noise (i.i.d.) to the vertices of the
original object. One distorted object is called the target, whose Gaussian noise has
a constant variance. The other is the distract or , whose noise has a larger variance
that can be adjusted to achieve a criterion level of performance. The two objects
are displayed from the same viewpoint in parallel projection, which is either from
one of the template views, or a novel view due to 3D rotation. The task is to choose
the one that is more similar to the original object. The observer's performance is
measured by the variance (threshold) that gives rise to 75% correct performance.
The optimal strategy is to choose the stimulus S with a larger probability p (OIS).
From Bayes' rule, this is to choose the larger of p (SIO).
Assume that the models are restricted to 2D transformations of the image, and
cannot reconstruct the 3D structure of the object from its independent templates
{Ti}. Assume also that the prior probability p(Td is constant. Let us represent S
and Ti by their (x, y) vertex coordinates: (X Y )T, where X = (Xl, x2, ... , x n ),
y = (yl, y2 , ... , yn). We assume that the correspondence between S and T i is
solved up to a reflection ambiguity, which is equivalent to an additional template:
Ti = (xr yr )T, where X r = (x n , ... ,x2,xl ), yr = (yn, ... ,y2,yl). We still
denote the template set as {Td. Therefore,
(1)
In what follows, we will compute p(SITi)p(T i ), with the assumption that S =
F (Ti) + N (0, crI 2n ), where N is the Gaussian distribution, 12n the 2n x 2n identity
matrix, and :F a 2D transformation. For the 2D ideal observer, :F is a rigid 2D
rotation. For the GRBF model, F assigns a linear coefficient to each template
T i , in addition to a 2D rotation. For the 2D affine nearest neighbor model, :F
represents the 2D affine transformation that minimizes liS - Ti11 2 , after Sand Ti
are normalized in size. For the 2D affine ideal observer, :F represents all possible
2D affine transformations applicable to T i.
Z Liu and D. Kersten
832
2.1
The 2D ideal observer
The templates are the original 2D images, their mirror reflections, and 2D rotations
(in angle ?) in the image plane. Assume that the stimulus S is generated by adding
Gaussian noise to a template, the probability p(SIO) is an integration over all
templates and their reflections and rotations. The detailed derivation for the 2D
ideal and the GRBF model can be found in [4].
Ep(SITi)p(Ti) ex: E
J
d?exp (-liS - Ti(?)112 /2( 2 )
?
(2)
2.2 The GRBF model
The model has the same template set as the 2D ideal observer does. Its training
requires that EiJ;7r d?Ci(?)N(IITj - Ti(?)II,a) = 1, j = 1,2, ... , with which {cd
can be obtained optimally using singular value decomposition. When a pair of new
stimuli is} are presented, the optimal decision is to choose the one that is closer
to the learned prototype, in other words, the one with a smaller value of
111- E 127r d?ci(?)exp
(_liS - 2:~(?)1I2) II.
(3)
2.3 The 2D affine nearest neighbor model
It has been proved in [8] that the smallest Euclidean distance D(S, T) between S
and T is, when T is allowed a 2D affine transformation, S ~ S/IISII, T ~ T/IITII,
D2(S, T) = 1 - tr(S+S . TTT)/IITII2,
(4)
where tr strands for trace, and S+ = ST(SST)-l. The optimal strategy, therefore,
is to choose the S that gives rise to the larger of E exp (_D2(S, Ti)/2a 2) , or the
smaller of ED2(S, Ti). (Since no probability is defined in this model, both measures
will be used and the results from the better one will be reported.)
2.4 The 2D affine ideal observer
We now calculate the Bayesian probability by assuming that the prior probability distribution of the 2D affine transformation, which is applied to the template
T i , AT
+ Tr
(~ ~) Ti + (~: ::: ~:),
=
obeys a Gaussian distribution
N(X o ,,,,/1 6 ), where Xo is the identity transformation
(1,0,0,1,0,0). We have
Ep(SIT i )
i:
xl'
= (a,b,c,d,tx,t y) =
dX exp (-IIATi + Tr - SII 2/2( 2)
=
E
=
EC(n, a, ",/)deC 1 (QD exp (tr (KfQi(QD-1QiKi) /2(12), (6)
where C(n, a, ",/) is a function of n, a, "'/; Q' = Q
Q _ ( XT . X T
Y T ?XT
(5)
+ ",/-212, and
X T ? Y T ) QK _ ( XT? Xs
YT ?YT
'
X T ?Ys
Y T . Xs)
-21
Y T .Ys
+"'/
2?
(7)
The free parameters are "'/ and the number of 2D rotated copies for each T i (since
a 2D affine transformation implicitly includes 2D rotations, and since a specific
prior probability distribution N(Xo, ",/1) is assumed, both free parameters should
be explored together to search for the optimal results).
2D Observers for Hwnan 3D Object Recognition?
833
?
?
?
?
?
?
Figure 1: Stimulus classes with increasing structural regularity: Balls, Irregular,
Symmetric, and V-Shaped. There were three objects in each class in the experiment.
2.5
The human observers
Three naive subjects were tested with four classes of objects: Balls, Irregular, Symmetric, and V-Shaped (Fig. 1). There were three objects in each class. For each
object, 11 template views were learned by rotating the object 60? /step, around
the X- and Y-axis, respectively. The 2D images were generated by orthographic
projection, and viewed monocularly. The viewing distance was 1.5 m. During the
test, the standard deviation of the Gaussian noise added to the target object was
(J"t = 0.254 cm. No feedback was provided.
Because the image information available to the humans was more than what was
available to the models (shading and occlusion in addition to the (x, y) positions of
the vertices), both learned and novel views were tested in a randomly interleaved
fashion. Therefore, the strategy that humans used in the task for the learned and
novel views should be the same. The number of self-occlusions, which in principle provided relative depth information, was counted and was about equal in both
learned and novel view conditions. The shading information was also likely to be
equal for the learned and novel views. Therefore, this additional information was
about equal for the learned and novel views, and should not affect the comparison
of the performance (humans relative to a model) between learned and novel views.
We predict that if the humans used a 2D affine strategy, then their performance
relative to the 2D affine ideal observer should not be higher for the novel views than
for the learned views. One reason to use the four classes of objects with increasing
structural regularity is that structural regularity is a 3D property (e.g., 3D Symmetric vs. Irregular), which the 2D models cannot capture. The exception is the
planar V-Shaped objects, for which the 2D affine models completely capture 3D rotations, and are therefore the "correct" models. The V-Shaped objects were used in
the 2D affine case as a benchmark. If human performance increases with increasing
structural regularity of the objects, this would lend support to the hypothesis that
humans have used 3D information in the task.
2.6
Measuring performance
A stair-case procedure [7] was used to track the observers' performance at 75%
correct level for the learned and novel views, respectively. There were 120 trials
for the humans, and 2000 trials for each of the models. For the GRBF model,
the standard deviation of the Gaussian function was also sampled to search for
the best result for the novel views for each of the 12 objects, and the result for
the learned views was obtained accordingly. This resulted in a conservative test
of the hypothesis of a GRBF model for human vision for the following reasons:
(1) Since no feedback was provided in the human experiment and the learned and
novel views were randomly intermixed, it is not straightforward for the model to
find the best standard deviation for the novel views, particularly because the best
standard deviation for the novel views was not the same as that for the learned
Z Liu and D. Kersten
834
ones. The performance for the novel views is therefore the upper limit of the
model's performance. (2) The subjects' performance relative to the model will be
defined as statistical efficiency (see below). The above method will yield the lowest
possible efficiency for the novel views, and a higher efficiency for the learned views,
since the best standard deviation for the novel views is different from that for the
learned views. Because our hypothesis depends on a higher statistical efficiency for
the novel views than for the learned views, this method will make such a putative
difference even smaller. Likewise, for the 2D affine ideal, the number of 2D rotated
copies of each template Ti and the value I were both extensively sampled, and the
best performance for the novel views was selected accordingly. The result for the
learned views corresponding to the same parameters was selected. This choice also
makes it a conservative hypothesis test.
3 Results
Learned Views
25
IJ
O
e-
O 20 Affine Nearest NtMghbor
:g
0
Human
20 Ideal
GRBF
rn 20 Affine kIoai
.?.
~
?
1.5
Novel Views
? Human
EJ 20 Ideal
o
o
e-
.?.
:!2
0
~
GRBF
20 Affine Nearesl N.tghbor
~ 2DAfllna~
1.5
~
81
l!
~
l-
I-
0.5
Object Type
Object Type
Figure 2: The threshold standard deviation of the Gaussian noise, added to the
distractor in the test pair, that keeps an observer's performance at the 75% correct
level, for the learned and novel views, respectively. The dotted line is the standard
deviation of the Gaussian noise added to the target in the test pair.
Fig. 2 shows the threshold performance. We use statistical efficiency E to compare human to model performance. E is defined as the information used by
humans relative to the ideal observer [3] : E = (d~uman/d~deal)2, where d'
is the discrimination index. We have shown in [4] that, in our task, E =
((a~1!f;actor)2 - (CTtarget)2) / ((CT~~~~~tor)2 - (CTtarget)2) , where CT is the threshold. Fig. 3 shows the statistical efficiency of the human observers relative to each
of the four models.
We note in Fig. 3 that the efficiency for the novel views is higher than those for the
learned views (several of them even exceeded 100%), except for the planar V-Shaped
objects. We are particularly interested in the Irregular and Symmetric objects in
the 2D affine ideal case, in which the pairwise comparison between the learned
and novel views across the six objects and three observers yielded a significant
difference (binomial, p < 0.05). This suggests that the 2D affine ideal observer
cannot account for the human performance, because if the humans used a 2D affine
template matching strategy, their relative performance for the novel views cannot
be better than for the learned views. We suggest therefore that 3D information was
used by the human observers (e.g., 3D symmetry). This is supported in addition
by the increasing efficiencies as the structural regularity increased from the Balls,
Irregular, to Symmetric objects (except for the V-Shaped objects with 2D affine
models).
2D Observers for Hwnan 3D Object Recognition?
300
..
l
.."
..!
"
$:
""
20 Ideal
250
o Learned
200
? Novel
250
l
f
'50
'"
~ ""
w
"-
II!
.'"
N
~
I .Noval
0 l&arnedl
~
--------------
" '"
Q
l>j
GRBF Modol
---
t
!
ObjoctTypo
300
20 Aftlne Nearest
250
200
o
Learned
? Novel
Q
N
Ighbor
l,..
j"
"
~
150
300
20 Affine Ideal
250
o Learned
200
? Novel
'50
j
i
I
~
ObJect Type
835
0
Object Type
ObjOGtType
Figure 3: Statistical efficiencies of human observers relative to the 2D ideal observer,
the GRBF model, the 2D affine nearest neighbor model, and the 2D affine ideal
observer_
4 Conclusions
Computational models of visual cognition are subject to information theoretic as
well as implementational constraints. When a model's performance mimics that of
human observers, it is difficult to interpret which aspects of the model characterize
the human visual system. For example, human object recognition could be simulated by both a GRBF model and a model with partial 3D information of the object.
The approach we advocate here is that, instead of trying to mimic human performance by a computational model, one designs an implementation-free model for a
specific recognition task that yields the best possible performance under explicitly
specified computational constraints. This model provides a well-defined benchmark
for performance, and if human observers outperform it, we can conclude firmly that
the humans must have used better computational strategies than the model. We
showed that models of independent 2D templates with 2D linear operations cannot
account for human performance. This suggests that our human observers may have
used the templates to reconstruct a representation of the object with some (possibly
crude) 3D structural information.
References
[1] Biederman I and Gerhardstein P C. Viewpoint dependent mechanisms in visual
object recognition: a critical analysis. J. Exp. Psych.: HPP, 21: 1506-1514, 1995.
[2] Biilthoff H H and Edelman S. Psychophysical support for a 2D view interpolation
theory of object recognition. Proc. Natl. Acad. Sci. , 89:60-64, 1992.
[3] Fisher R A. Statistical Methods for Research Workers. Oliver and Boyd, Edinburgh, 1925.
[4] Liu Z, Knill D C, and Kersten D. Object classification for human and ideal
observers. Vision Research, 35:549-568, 1995.
[5] Poggio T and Edelman S. A network that learns to recognize three-dimensional
objects. Nature, 343:263-266, 1990.
[6] Tarr M J and Biilthoff H H. Is human object recognition better described
by geon-structural-descriptions or by multiple-views? J. Exp. Psych.: HPP,
21:1494-1505,1995.
[7] Watson A B and Pelli D G. QUEST: A Bayesian adaptive psychometric method.
Perception and Psychophysics, 33:113-120, 1983.
[8] Werman M and Weinshall D. Similarity and affine invariant distances between
2D point sets. IEEE PAMI, 17:810-814,1995.
Toward a Single-Cell Account for
Binocular Disparity Tuning: An Energy
Model May be Hiding in Your Dendrites
Bartlett W. Mel
Department of Biomedical Engineering
University of Southern California, MC 1451
Los Angeles, CA 90089
mel@quake.usc.edu
Daniel L. Ruderman
The Salk Institute
10010 N. Torrey Pines Road
La Jolla, CA 92037
ruderman@salk.edu
Kevin A. Archie
Neuroscience Program
University of Southern California
Los Angeles, CA 90089
karchie@quake.usc.edu
Abstract
Hubel and Wiesel (1962) proposed that complex cells in visual cortex are driven by a pool of simple cells with the same preferred
orientation but different spatial phases. However, a wide variety of
experimental results over the past two decades have challenged the
pure hierarchical model, primarily by demonstrating that many
complex cells receive monosynaptic input from unoriented LGN
cells, or do not depend on simple cell input. We recently showed using a detailed biophysical model that nonlinear interactions among
synaptic inputs to an excitable dendritic tree could provide the nonlinear subunit computations that underlie complex cell responses
(Mel, Ruderman, & Archie, 1997). This work extends the result
to the case of complex cell binocular disparity tuning, by demonstrating in an isolated model pyramidal cell (1) disparity tuning
at a resolution much finer than the the overall dimensions of the
cell's receptive field, and (2) systematically shifted optimal disparity values for rivalrous pairs of light and dark bars-both in good
agreement with published reports (Ohzawa, DeAngelis, & Freeman, 1997). Our results reemphasize the potential importance of
intradendritic computation for binocular visual processing in particular, and for cortical neurophysiology in general.
A Single-Cell Accountfor Binocular Disparity Tuning
1
209
Introduction
Binocular disparity is a powerful cue for depth in vision. The neurophysiological
basis for binocular disparity processing has been of interest for decades, spawned
by the early studies of Rubel and Wiesel (1962) showing neurons in primary visual
cortex which could be driven by both eyes. Early qualitative models for disparity
tuning held that a binocularly driven neuron could represent a particular disparity
(zero, near, or far) via a relative shift of receptive field (RF) centers in the right
and left eyes. According to this model, a binocular cell fires maximally when an
optimal stimulus, e.g. an edge of a particular orientation, is simultaneously centered
in the left and right eye receptive fields, corresponding to a stimulus at a specific
depth relative to the fixation point. An account of this kind is most relevant to the
case of a cortical "simple" cell, whose phase-sensitivity enforces a preference for a
particular absolute location and contrast polarity of a stimulus within its monocular
receptive fields.
This global receptive field shift account leads to a conceptual puzzle, however, when
binocular complex cell receptive fields are considered instead, since a complex cell
can respond to an oriented feature nearly independent of position within its monocular receptive field. Since complex cell receptive field diameters in the cat lie in the
range of 1-3 degrees, the excessive "play" in their monocular receptive fields would
seem to render complex cells incapable of signaling disparity on the much finer scale
needed for depth perception (measured in minutes).
Intriguingly, various authors have reported that a substantial fraction of complex
cells in cat visual cortex are in fact tuned to left-right disparities much finer than
that suggested by the size of the monocular RF's. For such cells, a stimulus delivered at the proper disparity, regardless of absolute position in either eye, produces
a neural response in excess of that predicted by the sum of the monocular responses
(Pettigrew, Nikara, & Bishop, 1968; Ohzawa, DeAngelis, & Freeman, 1990; Ohzawa
et al., 1997). Binocular responses of this type suggest that for these cells, the left
and right RF's are combined via a correlation operation rather than a simple sum
(Nishihara & Poggio, 1984; Koch & Poggio, 1987). This computation has also been
formalized in terms of an "energy" model (Ohzawa et al., 1990, 1997), building
on the earlier use of energy models to account for complex cell orientation tuning
(Pollen & Ronner, 1983) and direction selectivity (Adelson & Bergen, 1985). In
an energy model for binocular disparity tuning, sums of linear Gabor filter outputs representing left and right receptive fields are squared to produce the crucial
multiplicative cross terms (Ohzawa et al., 1990, 1997).
Our previous biophysical modeling work has shown that the dendritic tree of a cortical pyramidal cells is well suited to support an approximative high-dimensional
quadratic input-output relation, where the second-order multiplicative cross terms
arise from local interactions among synaptic inputs carried out in quasi-isolated
dendritic "subunits" (Mel, 1992b, 1992a, 1993). We recently applied these ideas
to show that the position-invariant orientation tuning of a monocular complex cell
could be computed within the dendrites of a single cortical cell, based exclusively
upon excitatory inputs from a uniform, overlapping population of unoriented ON
and OFF cells (Mel et al., 1997). Given the similarity of the "energy" formulations
previously proposed to account for orientation tuning and binocu~ar disparity tuning, we hypothesized that a similar type of dendritic subunit computation could
underlie disparity tuning in a binocularly driven complex cell.
B. W. Mel, D. L Ruderman and K A. Archie
210
Parameter
Rm
Ra
Value
IOkO cm:l
2000cm
em
1.0ILF/cm~
Vrest
-70 mV
615
0.20,0.12 S/cm:l
0.05,0.03 S/cm:t.
0- 100 Hz
0.027 nS - 0.295 nS
0.5 ms, 3 ms
0.27 nS - 2.95 nS
0.5 ms, 50 ms
OmV
Compartments
Somatic !iNa, YnR
Dendritic !iNa, YnR
Input frequency
gAMPA
TAMPA (on,
of f)
gNMDA
7'NMDA (on,
off)
Esyn
Table 1: Biophysical simulation parameters. Details of HH channel implementation are given elsewhere (Mel, 1993); original HH channel implementation courtesy Ojvind Bernander and Rodney Douglas. In order that local EPSP size be
held approximately constant across the dendritic arbor, peak synaptic conductance at dendritic location x was approximately scaled to the local input resistance (inversely), given by 9syn(X) = C/Rin(X), where c was a constant, and
Rin(X) = max(Rin(X),200MO). Input resistance Rin(X) was measured for a passive cell. Thus 9syn was identical for all dendritic sites with input resistance below
200MO, and was given by the larger conductance value shown; roughly 50% of the
tree fell within a factor of 2 of this value. Peak conductances at the finest distal tips
were smaller by roughly a factor of 10 (smaller number shown). Somatic input resistance was near 24MO. The peak synaptic conductance values used were such that
the ratio of steady state current injection through NMDA vs. AMPA channels was
1.2?0.4. Both AMPA and NMDA-type synaptic conductances were modeled using
the kinetic scheme of Destexhe et al. (1994); synaptic activation and inactivation
time constants are shown for each.
2
Methods
Compartmental simulations of a pyramidal cell from cat visual cortex (morphology courtesy of Rodney Douglas and Kevan Martin) were carried out in NEURON
(Hines, 1989); simulation parameters are summarized in Table 1. The soma and dendritic membrane contained Hodgkin-Huxley-type (HH) voltage-dependent sodium
and potassium channels. Following evidence for higher spike thresholds and decremental propagation in dendrites (Stuart & Sakmann, 1994), HH channel density was
set to a uniform, 4-fold lower value in the dendritic membrane relative to that of the
cell body. Excitatory synapses from LGN cells included both NMDA and AMPAtype synaptic conductances. Since the cell was considered to be isolated from the
cortical network, inhibitory input was not modeled. Cortical cell responses were
reported as average spike rate recorded at the cell body over the 500 ms stimulus
period, excluding the 50 ms initial transient.
The binocular LGN consisted of two copies of the monocular LGN model used
previously (Mel et al., 1997), each consisting of a superimposed pair of 64x64 ON
and OFF subfields. LGN cells were modeled as linear, half-rectified center-surround
filters with centers 7 pixels in width. We randomly subsampled the left and right
LGN arrays by a factor of 16 to yield 1,024 total LGN inputs to the pyramidal cell.
A Single-Cell Account for Binocular Disparity Tuning
211
A developmental principle was used to determine the spatial arrangement of these
1,024 synaptic contacts onto the dendritic branches of the cortical cell, as follows.
A virtual stimulus ensemble was defined for the cell, consisting of the complete set
of single vertical light or dark bars presented binocularly at zero-disparity within
the cell's receptive field. Within this ensemble, strong pairwise correlations existed
among cells falling into vertically aligned groups of the same (ON or OFF) type,
and cells in the vertical column at zero horizontal disparity in the other eye. These
binocular cohorts of highly correlated LGN cells were labeled mutual "friends".
Progressing through the dendritic tree in depth first order, a randomly chosen LG N
cell was assigned to the first dendritic site. A randomly chosen "friend" of hers
was assigned to the second site, the third site was assigned to a friend of the site 2
input, etc., until all friends in the available subsample were assigned (4 from each
eye, on average). If the friends of the connection at site i were exhausted, a new
LGN cell was chosen at random for site i + 1. In earlier work, this type of synaptic
arrangement was shown to be the outcome of a Hebb-type correlational learning
rule, in which random, activity independent formation of synaptic contacts acted
to slowly randomize the axo-dendritic interface, shaped by Hebbian stabilization of
synaptic contacts based on their short-range correlations with other synapses.
3
Results
Model pyramidal cells configured in this way exhibited prominent phase-invariant
orientation tuning, the hallmark response property of the visual complex cell. Multiple orientation tuning curves are shown, for example, for a monocular complex cell,
giving rise to strong tuning for light and dark bars across the receptive field (fig. 1).
The bold curve shows the average of all tuning curves for this cell; the half-width at
half max is 25?, in the normal range for complex cells in cat visual cortex (Orban,
1984). When the spatial arrangement of LGN synaptic contacts onto the pyramidal cell dendrites was randomly scrambled, leaving all other model parameters
unchanged, orientation tuning was abolished in this cell (right frame), confirming
the crucial role of spatially-mediated nonlinear synaptic interactions (average curve
from left frame is reproduced for comparison).
Disparity-tuning in an orientation-tuned binocular model cell is shown in fig. 2, compared to data from a complex cell in cat visual cortex (adapted from Ohzawa et al.
(1997)). Responses to contrast matched (light-light) and contrast non-matched
(light-dark) bar pairs were subtracted to produce these plots. The strong diagonal
structure indicates that both the model and real cells responded most vigorously
when contrast-matched bars were presented at the same horizontal position in the
left and right-eye RF's (Le. at zero-disparity), whereas peak responses to contrastnon-matched bars occured at symmetric near and far, non-zero disparities.
4
Discussion
The response pattern illustrated in fig. 2A is highly similar to the response generated
by an analytical binocular energy model for a complex cell (Ohzawa et al., 1997):
{exp (-kXi) cos (271' f XL)
{exp (-kxiJ sin (271' f XL)
+ exp (-kX'kJ cos (271' f XR)}2 +
+ exp (-kXh) sin (271' f XR)}2,
(1)
where XL and X R are the horizontal bar positions to the two eyes, k is the factor
B. W. Mel, D_ L Ruderman and K. A. Archie
212
Ordered vs. Scrambled
Orientation Tuning
70
60
'0
Ql
~
Ql
"''5."
.e
Ql
(/)
50
40
+-
50
-?I-
45
-+-
U
,,*-
Ql
.... -
(/)
Us
-ll-
(/)
Ql
35
.e
30
"''5."
~
(/)
20
(/)
Ql
c:
8.
ex:
10
0
-90
25
20
Ql
ex:
ordered
scrambled -+-
40
Ql
30
c:
8.
-
55
average
lightO
dark 4
light 8
light16
dark 16
15
/'--- "'
+
10
-60
-30
0
30
Orientabon (degrees)
60
90
5
-90
'+---- /
-60
/
+ , ~,
I
' +_+- ~
-30
0
30
Orientation (degrees)
60
90
Figure 1: Orientation tuning curves are shown in the left frame for light and dark
bars at 3 arbitrary positions_ Essentially similar responses were seen at other receptive field positions, and for other complex cells_ Bold trace indicates average
of tuning curves at positions 0, 1, 2, 4, 8, and 16 for light and dark bars. Similar
form of 6 curves shown reflects the translation-invariance of the cell's response to
oriented stimuli, and symmetry with respect to ON and OFF input. Orientation
tuning is eliminated when the spatial arrangement of LGN synapses onto the model
cell dendrites is randomly scrambled (right frame).
Complex Cell Model
Complex Cell in Cat VI
Ohzawa, Deangelis, & Freeman, 1997
Right eye position
Right eye position
Figure 2: Comparison of disparity tuning in model complex cell to that of a binocular complex cell from cat visual cortex. Light or dark bars were presented simultaneously to the left and right eyes. Bars could be of same polarity in both eyes
(light, light) or different polarity (light, dark); cell responses for these two cases were
subtracted to produce plot shown in left frame. Right frame shows data similarly
displayed for a binocular complex cell in cat visual cortex (adapted from Ohzawa
et al. (1997)).
A Single-Cell Account for Binocular Disparity Tuning
that determines the width of the subunit RF's, and
213
f
is the spatial frequency.
In lieu of literal simple cell "subunits" , the present results indicate that the subunit
computations associated with the terms of an energy model could derive largely
from synaptic interactions within the dendrites of the individual cortical cell, driven
exclusively by excitatory inputs from unoriented, monocular ON and OFF cells
drawn from a uniform overlapping spatial distribution. While lateral inhibition
and excitation play numerous important roles in cortical computation, the present
results suggest they are not essential for the basic features of the nonlinear disparity
tuned responses of cortical complex cells. Further, these results address the paradox
as to how inputs from both unoriented LGN cells and oriented simple cells can
coexist without conflict within the dendrites of a single complex cell.
A number of controls from previous work suggest that this type of subunit processing is very robustly computed in the dendrites of an individual neuron, with little
sensitivity to biophysical parameters and modeling assumptions, including details of
the algorithm used to spatially organize the genicula-cortical projection, specifics of
cell morphology, synaptic activation density across the dendritic tree, passive membrane and cytoplasmic parameters, and details of the kinetics, voltage-dependence,
or spatial distribution of the voltage-dependent dendritic channels.
One important difference between a standard energy model and the intradendritic
responses generated in the present simulation experiments is that the energy model
has oriented RF structure at the linear (simple-cell-like) stage, giving rise to oriented, antagonistic ON-OFF subregions (Movshon, Thompson, & Tolhurst, 1978),
whereas the linear stage in our model gives rise to center-surround antagonism only
within individual LGN receptive fields. Put another way, the LGN-derived subunits
in the present model cannot provide all the negative cross-terms that appear in the
energy model equations, specifically for pairs of pixels that fall outside the range of
a single LG N receptive field.
While the present simulations involve numerous simplifications relative to the full
complexity of the cortical microcircuit, the results nonetheless emphasize the potential importance of intradendritic computation in visual cortex.
Acknowledgements
Thanks to Ken Miller, Allan Dobbins, and Christof Koch for many helpful comments
on this work. This work was funded by the National Science Foundation and the
Office of Naval Research, and by a Slo~n Foundation Fellowship (D.R.).
References
Adelson, E., & Bergen, J. (1985). Spatiotemporal energy models for the perception
of motion. J. Opt. Soc. Amer., A 2, 284-299.
Rines, M. (1989). A program for simulation of nerve equations with branching
geometries. Int. J. Biomed. Comput., 24, 55-68.
Rubel, D., & Wiesel, T . (1962) . Receptive fields, binocular interaction and functional architecture in the cat's visual cortex. J. Physiol., 160, 106- 154.
Koch, C., & Poggio, T . (1987) . Biophysics of computation: Neurons, synapses,
and membranes. In Edelman, G., Gall, W., & Cowan, W. (Eds.), Synaptic
junction, pp. 637-697. Wiley, New York.
Mel, B. (1992a). The clusteron: Toward a simple abstraction for a complex neuron. In Moody, J., Hanson, S., & Lippmann, R. (Eds.), Advances in Neural
214
B. W. Mel, D. L Ruderman and K. A Archie
Information Processing Systems, vol. 4, pp. 35-42. Morgan Kaufmann, San
Mateo, CA.
Mel, B. (1992b). NMDA-based pattern discrimination in a modeled cortical neuron.
Neural Computation, 4, 502-516.
Mel, B. (1993). Synaptic integration in an excitable dendritic tree. J. Neurophysiol.,
70(3), 1086-110l.
Mel, B., Ruderman, D., & Archie, K. (1997). Complex-cell responses derived from
center-surround inputs: the surprising power of intradendritic computation. In
Mozer, M., Jordan, M., & Petsche, T. (Eds.), Advances in Neural Information
Processing Systems, Vol. 9, pp. 83-89. MIT Press, Cambridge, MA.
Movshon, J., Thompson, I., & Tolhurst, D. (1978). Receptive field organization of
complex cells in the cat's striate cortex. J. Physiol., 283, 79-99.
Nishihara, H., & Poggio, T. (1984). Stereo vision for robotics. In Brady, & Paul
(Eds.), Proceedings of the First International Symposium of Robotics Research,
pp. 489-505. MIT Press, Cambridge, MA.
Ohzawa, I., DeAngelis, G., & Freeman, R. (1990). Stereoscopic depth discrimination
in the visual cortex: Neurons ideally suited as disparity detectors. Science,
249, 1037- 104l.
Ohzawa, I., DeAngelis, G. , & Freeman, R. (1997). Encoding of binocular disparity
by complex cells in the cat's visual cortex. J. Neurophysiol., June.
Orban, G. (1984). Neuronal operations in the visual cortex. Springer Verlag, New
York.
Pettigrew, J., Nikara, T., & Bishop, P. (1968). Responses to moving slits by single
units in cat striate cortex. Exp. Brain Res., 6, 373-390.
Pollen, D., & Ronner, S. (1983). Visual cortical neurons as localized spatial frequency filters. IEEE Trans. Sys. Man Cybero., 13, 907-916.
Stuart, G., & Sakmann, B. (1994). Active propagation of somatic action potentials
into neocortical pyramidal cell dendrites. Nature, 367, 69-72.
| 1365 |@word neurophysiology:1 trial:2 wiesel:3 stronger:1 d2:2 simulation:6 decomposition:1 tr:5 shading:2 vigorously:1 initial:1 liu:6 disparity:26 exclusively:2 daniel:2 tuned:3 past:1 current:1 surprising:1 hpp:2 activation:2 dx:1 must:3 finest:1 physiol:2 confirming:1 plot:2 v:3 discrimination:3 cue:1 selected:2 yr:2 half:3 accordingly:2 plane:2 sys:1 short:1 provides:3 tolhurst:2 location:2 preference:1 si1:1 sii:1 symposium:1 edelman:3 qualitative:1 fixation:1 advocate:1 pairwise:2 allan:1 ra:1 indeed:1 roughly:2 nor:1 distractor:1 morphology:2 brain:1 freeman:5 td:3 little:1 increasing:4 becomes:1 provided:3 hiding:1 monosynaptic:1 matched:4 lowest:1 what:3 weinshall:1 cm:6 kind:1 minimizes:1 psych:2 finding:1 transformation:15 brady:1 ti:13 rm:1 scaled:1 control:1 unit:1 underlie:2 yn:2 organize:1 appear:1 christof:1 engineering:1 local:3 vertically:1 limit:1 acad:1 omv:1 encoding:1 interpolation:2 approximately:2 pami:1 mateo:1 equivalence:1 suggests:2 co:2 limited:4 range:5 subfields:1 obeys:1 enforces:1 orthographic:1 xr:3 signaling:1 procedure:1 reject:1 gabor:1 matching:5 projection:4 word:2 radial:1 boyd:1 road:1 suggest:4 onto:4 cannot:6 coexist:1 twodimensional:1 put:1 kersten:6 ilf:1 equivalent:1 yt:2 center:5 courtesy:2 straightforward:1 regardless:1 thompson:2 focused:1 resolution:1 formalized:1 assigns:1 pure:1 factored:1 rule:3 estimator:2 ojvind:1 array:1 population:1 x64:1 variation:3 coordinate:1 modol:1 antagonistic:1 target:3 play:2 dobbin:1 approximative:1 gall:1 hypothesis:4 agreement:1 recognition:25 particularly:3 tributions:1 labeled:1 ep:2 role:2 solved:1 capture:2 ynr:2 calculate:1 substantial:1 mozer:1 developmental:1 complexity:1 ideally:1 depend:1 upon:1 efficiency:10 rin:4 basis:2 completely:1 neurophysiol:2 represented:1 tx:1 cat:12 various:1 derivation:1 deangelis:5 kevin:1 outcome:1 formation:1 cytoplasmic:1 outside:1 whose:3 larger:5 compartmental:1 otherwise:1 reconstruct:2 ability:1 torrey:1 delivered:1 reproduced:1 biophysical:4 analytical:1 interaction:5 epsp:1 j2:1 relevant:1 aligned:1 flexibility:1 achieve:1 description:1 normalize:1 los:2 potassium:1 impaired:1 regularity:5 produce:5 rotated:2 object:60 axo:1 depending:1 friend:5 derive:1 measured:3 ij:1 nearest:7 ex:3 strong:4 reco:1 implemented:2 ois:1 predicted:1 implies:2 indicate:1 qd:2 soc:1 direction:1 vrest:1 correct:4 filter:3 centered:1 human:54 stabilization:1 viewing:1 transient:1 gampa:1 virtual:1 sand:1 d_:1 opt:1 dendritic:17 adjusted:1 kinetics:1 around:1 considered:2 koch:3 normal:1 exp:13 gnmda:1 antagonism:1 cognition:1 predict:1 puzzle:1 mo:3 werman:1 pine:1 tor:1 early:2 smallest:1 proc:1 integrates:1 applicable:1 siti:2 tool:1 reflects:1 mit:2 kxh:1 gaussian:9 rather:1 inactivation:1 ej:1 varying:1 voltage:3 office:1 derived:4 june:1 rubel:2 naval:1 superimposed:1 indicates:2 contrast:5 sense:1 progressing:1 helpful:1 dependent:6 rigid:3 cri:1 bergen:2 abstraction:1 unlikely:2 typically:1 relation:1 quasi:1 interested:2 lgn:14 biomed:1 pixel:2 overall:1 among:4 orientation:13 classification:1 issue:1 provably:1 constrained:4 integration:2 psychophysics:1 spatial:8 mutual:1 equal:3 field:17 shaped:7 intriguingly:1 tarr:1 eliminated:1 identical:1 represents:2 stuart:2 look:1 adelson:2 nearly:1 excessive:1 mimic:3 report:1 stimulus:14 primarily:1 randomly:7 oriented:5 simultaneously:2 resulted:1 recognize:3 individual:3 national:1 familiar:2 usc:2 phase:3 occlusion:2 consisting:2 fire:1 subsampled:1 geometry:1 conductance:6 organization:1 interest:1 highly:2 extreme:2 light:13 stair:1 natl:1 held:2 noval:1 oliver:1 edge:1 closer:1 partial:1 necessary:1 worker:1 poggio:5 tree:6 euclidean:1 rotating:1 re:1 deformation:1 isolated:3 theoretical:1 increased:1 classify:1 earlier:2 modeling:2 column:1 ar:1 implementational:2 measuring:1 challenged:1 vertex:3 deviation:7 uniform:3 archie:6 too:1 optimally:1 stored:3 reported:3 characterize:1 answer:1 spatiotemporal:1 kxi:1 combined:1 st:1 density:2 peak:4 sensitivity:2 discriminating:1 thanks:1 international:1 yl:2 off:7 pool:1 tip:1 together:1 uiti:1 moody:1 squared:1 central:2 ambiguity:1 pettigrew:2 recorded:1 choose:5 possibly:1 slowly:1 literal:1 li:3 account:15 suggesting:1 potential:3 summarized:1 bold:2 includes:1 coefficient:1 int:1 configured:1 explicitly:2 mv:1 depends:4 vi:1 multiplicative:2 view:46 observer:44 closed:1 nishihara:2 bayes:1 parallel:2 ttt:1 rodney:2 contribution:1 compartment:1 responded:1 variance:3 qk:1 likewise:1 ensemble:2 yield:6 kxij:1 largely:1 miller:1 kaufmann:1 generalize:1 bayesian:5 intradendritic:4 overlook:1 mc:1 rectified:1 finer:3 published:1 detector:1 synapsis:4 synaptic:17 ed:4 energy:11 nonetheless:1 frequency:3 pp:4 associated:1 con:1 sampled:2 proved:1 occured:1 nmda:5 syn:2 nerve:1 exceeded:1 higher:5 planar:2 response:17 maximally:1 formulation:1 amer:1 microcircuit:1 ronner:2 binocular:20 biomedical:1 stage:2 correlation:3 until:1 horizontal:3 grbf:14 ruderman:7 nonlinear:4 overlapping:2 propagation:2 building:1 ohzawa:11 hypothesized:1 normalized:1 y2:2 consisted:1 assigned:4 spatially:2 symmetric:6 i2:1 illustrated:1 deal:1 distal:1 ll:1 during:1 branching:1 self:1 width:3 sin:2 mel:14 steady:1 excitation:1 criterion:1 generalized:1 trying:1 m:6 prominent:1 slo:1 neocortical:1 theoretic:1 complete:1 motion:1 reflection:3 passive:2 interface:1 image:23 hallmark:1 novel:31 recently:2 rotation:8 functional:1 ti11:1 unoriented:4 functionally:1 interpret:1 significant:2 surround:3 cambridge:2 tuning:24 similarly:1 funded:1 minnesota:1 moving:1 access:1 actor:1 similarity:4 cortex:15 inhibition:1 etc:1 showed:3 recent:1 jolla:1 driven:5 store:2 selectivity:1 verlag:1 incapable:1 watson:1 seen:1 morgan:1 greater:1 additional:2 determine:1 period:1 ii:3 branch:1 multiple:2 full:1 hebbian:1 match:1 cross:3 long:1 y:2 biophysics:1 converging:1 basic:1 vision:5 essentially:1 represent:2 robotics:2 dec:1 irregular:5 cell:77 addition:3 receive:1 whereas:2 fellowship:1 singular:1 pyramidal:7 leaving:1 crucial:3 appropriately:1 exhibited:1 fell:1 comment:1 subject:4 hz:1 cttarget:2 cowan:1 seem:1 jordan:1 iitii:1 structural:8 near:3 ideal:36 intermediate:1 cohort:1 destexhe:1 variety:1 affect:1 architecture:1 zili:1 idea:2 prototype:1 angeles:2 shift:2 whether:1 six:1 bartlett:1 movshon:2 stereo:1 render:1 resistance:4 york:2 action:1 kevan:1 detailed:2 sst:1 spawned:1 involve:1 dark:10 extensively:1 subregions:1 ken:1 diameter:1 outperform:2 inhibitory:1 shifted:1 dotted:1 stereoscopic:1 neuroscience:1 track:1 vol:2 group:1 key:1 four:3 soma:1 threshold:5 demonstrating:2 falling:1 drawn:1 neither:1 douglas:2 fraction:1 sum:3 angle:1 powerful:1 respond:1 distorted:2 hodgkin:1 extends:1 reasonable:1 putative:1 decision:2 interleaved:1 ct:2 simplification:1 correspondence:1 fold:1 quadratic:1 existed:1 yielded:1 activity:1 adapted:2 constraint:3 huxley:1 your:1 x2:2 aspect:3 orban:2 injection:1 martin:1 acted:1 department:1 geon:1 according:1 ball:3 membrane:4 smaller:5 across:4 em:1 iisii:1 conceivably:1 restricted:4 invariant:3 xo:2 binocularly:3 monocular:9 equation:2 previously:2 fail:1 mechanism:1 needed:1 hh:4 serf:1 lieu:1 available:5 operation:5 junction:1 hierarchical:1 petsche:1 robustly:1 subtracted:2 original:5 binomial:1 giving:2 monocularly:1 contact:4 unchanged:1 psychophysical:1 question:2 added:3 spike:2 arrangement:4 strategy:6 receptive:17 dependence:8 primary:1 randomize:1 diagonal:1 striate:2 southern:2 distance:3 separate:1 simulated:1 sci:1 lateral:1 extent:1 reason:2 toward:2 assuming:1 modeled:5 index:1 insufficient:1 polarity:3 ratio:1 difficult:1 esyn:1 intermixed:1 lg:2 ql:8 debate:2 trace:2 negative:1 rise:5 implementation:6 design:1 proper:1 sakmann:2 upper:1 vertical:2 observation:1 wire:1 neuron:9 benchmark:2 displayed:3 subunit:8 excluding:1 precise:1 paradox:1 frame:7 rn:1 somatic:3 gerhardstein:1 arbitrary:1 biederman:1 inferred:1 pair:7 specified:3 connection:1 pelli:1 conflict:1 hanson:1 california:2 pollen:2 learned:25 trans:1 address:1 bar:11 suggested:1 below:3 pattern:3 perception:3 program:2 rf:6 tampa:1 memory:3 lend:1 max:2 including:1 power:1 critical:1 sodium:1 representing:1 scheme:1 firmly:1 eye:12 inversely:1 numerous:2 axis:1 irrespective:1 carried:2 excitable:2 naive:1 bernander:1 embodied:1 mediated:1 kj:1 review:1 geometric:1 prior:3 acknowledgement:1 relative:13 ina:2 abolished:1 localized:1 foundation:2 degree:5 affine:37 sufficient:2 principle:2 viewpoint:17 systematically:1 quake:2 cd:1 translation:1 excitatory:3 elsewhere:1 supported:1 free:4 copy:4 allow:1 understand:2 institute:2 neighbor:5 template:24 taking:1 wide:1 fall:1 absolute:2 edinburgh:1 feedback:2 depth:6 dimension:1 evaluating:1 cortical:14 curve:7 computes:1 author:1 adaptive:1 ici:1 san:1 counted:1 ec:1 far:2 excess:1 approximate:1 lippmann:1 ed2:1 biilthoff:2 implicitly:1 preferred:1 keep:1 decremental:1 emphasize:1 global:1 hubel:1 active:1 conceptual:1 assumed:1 conclude:1 scrambled:4 search:2 decade:2 table:2 nature:3 channel:6 ca:4 symmetry:2 dendrite:9 distract:1 necessarily:1 complex:29 ampa:2 whole:1 noise:7 rivalrous:1 arise:1 knill:2 subsample:1 paul:1 allowed:1 hwnan:3 body:2 neuronal:1 fig:7 psychometric:1 site:7 ighbor:1 fashion:1 hebb:1 salk:2 wiley:1 n:4 position:10 xl:6 lie:1 crude:1 comput:1 third:1 learns:1 minute:1 familiarity:1 xt:3 specific:4 bishop:2 uman:1 showing:1 explored:1 x:2 evidence:2 sit:1 essential:2 clusteron:1 adding:2 importance:2 ci:3 mirror:1 nec:1 exhausted:1 kx:1 falsified:1 suited:2 led:1 eij:1 appearance:1 likely:1 neurophysiological:1 visual:22 positional:1 strand:1 contained:1 ordered:2 springer:1 determines:1 kinetic:1 hines:1 ma:2 identity:2 viewed:1 fisher:1 man:1 change:1 included:1 specifically:1 except:2 principal:1 conservative:2 called:2 total:1 correlational:1 iitj:1 experimental:3 la:1 arbor:1 invariance:1 slit:1 exception:1 sio:2 support:3 quest:1 tested:2 correlated:1 |
401 | 1,366 | Blind Separation of Radio Signals
Fading Channels
?
In
Kari Torkkola
Motorola, Phoenix Corporate Research Labs,
2100 E. Elliot Rd, MD EL508, Tempe, AZ 85284, USA
email: A540AA(Qemail.mot.com
Abstract
We apply information maximization / maximum likelihood blind
source separation [2, 6) to complex valued signals mixed with complex valued nonstationary matrices. This case arises in radio communications with baseband signals. We incorporate known source
signal distributions in the adaptation, thus making the algorithms
less "blind". This results in drastic reduction of the amount of data
needed for successful convergence. Adaptation to rapidly changing
signal mixing conditions, such as to fading in mobile communications, becomes now feasible as demonstrated by simulations.
1
Introduction
In SDMA (spatial division multiple access) the purpose is to separate radio signals
of interfering users (either intentional or accidental) from each others on the basis
of the spatial characteristics of the signals using smart antennas, array processing,
and beamforming [5, 8). Supervised methods typically use a variant of LMS (least
mean squares), either gradient based, or algebraic, to adapt the coefficients that
describe the channels or their inverses. This is usually a robust way of estimating
the channel but a part of the signal is wasted as predetermined training data, and
the methods might not be fast enough for rapidly varying fading channels.
Unsupervised methods either rely on information about the antenna array manifold,
or properties of the signals. Former approaches might require calibrated antenna
arrays or special array geometries. Less restrictive methods use signal properties
only, such as constant modulus, finite alphabet, spectral self-coherence, or cyclostationarity. Blind source separation (BSS) techniques typically rely only on source
signal independence and non-Gaussianity assumptions.
Our aim is to separate simultaneous radio signals occupying the same frequency
band, more specifically, radio signals that carry digital information. Since linear
mixtures of antenna signals end up being linear mixtures of (complex) baseband
signals due to the linearity of the downconversion process, we will apply BSS at
the baseband stage of the receiver. The main contribution of this paper is to
show that by making better use of the known signal properties, it is possible to
devise algorithms that adapt much faster than algorithms that rely only on weak
assumptions, such as source signal independence.
We will first discuss how the probability density functions (pdf) of baseband DPSK
signals could be modelled in' a way that can efficiently be used in blind separation
algorithms. We will incorporate those models into information maximization and
Blind Separation of Radio Signals in Fading Channels
757
into maximum likelihood approaches [2, 6). We will then continue with the maximum likelihood approach and other modulation techniques, such as QAM. Finally,
we will show in simulations, how this approach results in an adaptation process that
is fast enough for fading channels.
2
Models of baseband signal distributions
In digital communications the binary (or n-ary) information is transmitted as discrete combinations of the amplitude and/or the phase of the carrier signal. After
downconversion to baseband the instantaneous amplitude of the carrier can be observed as the length of a complex valued sample of the baseband signal, and the
phase of the carrier is discernible as the phase angle of the same sample. Possible
combinations that depend on the modulation method employed, are called symbol constellations. N-QAM (quadrature amplitude modulation) utilizes both the
amplitude and the phase, whereby the baseband signals can only take one of N
possible locations on a grid on the complex plane. In N-PSK (phase shift keying)
the amplitude of the baseband signal stays constant, but the phase can take any
of N discrete values. In DPSK (differential phase shift keying) the information is
encoded as the difference between phases of two consecutive transmitted symbols.
The phase can thus take any value, and since the amplitude remains constant, the
baseband signal distribution is a circle on the complex plane.
Information maximization BSS requires a nonlinear function that models the cumulative density function (cdf) of the data. This function and its derivative need
to be differentiable. In the case of a circular complex distribution with uniformly
distributed phase, there is only one important direction of deviation, the radial
direction. A smooth cdf G for a circular distribution at the unit circle can be
constructed using the hyperbolic tangent function as
G(z) = tanh(w(lzl - 1))
(1)
and the pdf, differentiated in the radial direction, that is, with respect to Izl is
8
g(z) = Bizi tanh(w(lzl - 1)) = w(l - tanh2(w(lzl - 1)))
(2)
where z = x + iy is a complex valued variable, and the parameter w controls the
steepness of the slope of the tanh function. Note that this is in contrast to more
commonly used coordinate axis directions to differentiate and to integrate to get
the pdf from the cdf and vice versa. These functions are plotted in Fig. 1.
a) CDF
b) PDF
Figure 1: Radial tanh with w=2.0 (equations 1 and 2) .
Note that we have not been worrying about the pdf integrating to unity. Thus we
could leave the first multiplicative constant w out of the definition of g. Scaling will
not be important for our purposes of using these functions as the nonlinearities in
the information maximization BSS. Note also that when the steepness w approaches
infinity, the densities approach the ideal density of a DPSK source, the unit circle.
Many other equally good choices are possible where the ideal density is reached as
a limit of a parameter value. For example, the radial section of the circular "ridge"
of the pdf could be a Gaussian.
758
3
K. Torkkola
The information maximization adaptation equation
The information maximization adaptation equation to learn the unmixing matrix
W using the natural gradient is [2]
AWex (guT
+ I)W
where
~. -
8
!!Jli..
(3)
YJ - 8Yi 8Ui
Vector U = W x denotes a time sample of the separated sources, x denotes the
corresponding time sample of the observed mixtures, and Yj is the nonlinear function
approximating the cdf of the data, which is applied to each component of the u.
Now we can insert (1) into Yj. Making use of {)lzI/{)z
~
Yj
= zllzl this yields for 'OJ:
() -()
() tanh(w (I Uj I-1
) ) = -2WYj-1
Uj-I
= -()
Yj Uj
Uj
(4)
When (4) is inserted into (3) we get
<lWex (I - 2 (Wjtanh(WI~~~jl-l?"j)
j
"H)
W
(5)
where (.)j denotes a vector with elements of varying j. Here, we have replaced
the transpose operator by the hermitian operator H, since we will be processing
complex data. We have also added a subscript to W as these parameters can be
learned, too. We will not show the adaptation equations due to lack of space.
4
Connection to the maximum likelihood approach
Pearlmutter and Parra have shown that (3) can be derived from the maximum
likelihood approach to density estimation [6]. The same fact has also been pointed
out by others, for example, by Cardoso [3]. We will not repeat their straightforward
derivation, but the final adaptation equation is of the following form:
AWex -
dO WTW = ((fj(Uj; Wj)) u T +
dW
Ii (Uj; Wj)
j
I) W.
(6)
where U = Wx are the sources separated from mixtures x, and fj(uj;wj) is the pdf
of source j parametrized by Wj. This is exactly the form of Bell and Sejnowski when
Ii is taken to be the derivative of the necessary nonlinearity gj, which was assumed
to be "close" to the true cdf of the source. Thus the information maximization
approach makes implicit assumptions about the cdf's of the sources in the form of
the nonlinear squashing function, and does implicit density estimation, whereas in
the ML approach the density assumptions are made explicit. This fact makes it
more intuitive and lucid to derive the adaptation for other forms of densities, and
also to extend it to complex valued variables.
Now, we can use the circular pdf's (2) depicted in Fig. 1 as the densities Ii (omitting
1- tanh 2 (wj(lujl- 1)). where the steepness Wj acts as the
scaling) fj(uj;wj)
single parameter of the density. Now we need to compute its derivative
=
fj(uj;wj) = (){) Ii(uj;wj) = -2tanh(wj(IUjl-l))Ii(Uj;Wj)WjIUjl
(7)
Uj
Uj
Inserting this into (6) and changing transpose operators into hermitians yields
<l
Wex (I _2 (Wjtanh(WI~~~jl- 1?,,; ) ; "H) W,
(8)
which is exactly the information maximization rule (5). Notice that at this time
we did not have to ponder what would be an appropriate way to construct the cdf
from the pdf for complex valued distributions.
759
Blind Separation of Radio Signals in Fading Channels
5
Modifications for QAM and other signal constellations
So far we have only looked at signals that lie on the unit circle, or that have a
constant modulus. Now we will take a look at other modulation techniques, in which
the alphabet is constructed as discrete points on the complex plane. An example
is the QAM (quadrature amplitude modulation), in which the signal alphabet is a
regular grid. For example, in 4-QAM, the alphabet could be A4 = {I+i, -I+i, -I-i,
l-i}, or any scaled version of A4.
In the ideal pdf of 4-QAM, each symbol is represented just as a point. Again, we
can construct a smoothed version of the ideal pdf as the sum of "bumps" over all
of the alphabet where the ideal pdf will be approached by increasing w.
g(U) =
2:)1 -
tanh2(wkl u - Uk!))
(9)
k
Now the density for each source j will be
!i(Uj; Wj) =
2:)1 -
tanh 2 (wkl u j
-
(10)
Uk!))
k
where Wj is now a vector of parameters Wk. In practice each
which case a single parameter W will suffice.
Wk
would be equal in
This density function could now be inserted into (6) resulting in the weight update
equation. However, since !i(Uj; Wj) is a sum of multiple components, f' / f will not
have a particularly simple form. In essence, for each sample to be processed, we
would need to evaluate all the components of the pdf model of the constellation.
This can be avoided by evaluating only the component of the pdf corresponding to
that symbol of the alphabet U c which is nearest to the current separated sample u.
This is a very good approximation when W is large. But the approximation does
not even have to be a good one when W is small, since the whole purpose of using
"wide" pdf components is to be able to evaluate the gradients on the whole complex
plane. Figure 2 depicts examples of this approximation with two different values of
w. The discontinuities are visible at the real and imaginary axes for the smaller w.
a) w = 1.0
b) w = 5.0
Figure 2: A piecewise continuous PDF for a 4-QAM source using the tanh function.
Thus for the 4-QAM, the complex plane will be divided into 4 quadrants, each
having its own adaptation rule corresponding to the single pdf component in that
quadrant. Evaluating (6) for each component of the sum gives
Ll..
W'" (I _2 (w. tanh(WI~~j
-
'"I) Uj \
"H )
W,
(11)
for each symbol k of the alphabet or for the corresponding location Uk on the
complex plane. This equation can be applied as such when the baseband signal is
sampled at the symbol rate. With oversampling, it may be necessary to include in
the pdf model the transition paths between the symbols, too.
6
Practical simplifications
To be able to better vectorize the algorithm, it is practical to accumulate ~ W
from a number of samples before updating the W. This amounts to computing an
expectation of ~W over a number, say, 10-500 samples of the mixtures. Looking
at the DPSK case, (5) or (8) the expectation of IUil in the denominator equals one
"near" convergence since we assume baseband signals that are distributed on the
unit circle.
Also, near the solution we can assume that the separated outputs Uj are close to
true distributions, the exact unit circle, which can be derived from h by increasing
its steepness. At the limit the tanh will equal the sign function, when the whole
adaptation, ignoring scaling, is
(12)
However, this simplification can only be used when the W is not too far off from the
correct solution. This is especially true when the number of available samples of the
mixtures is small. The smooth tanh is needed in the beginning of the adaptation
to give the correct direction to the gradient in the algorithm since the pdfs of the
outputs Uj are far from the ideal ones in the beginning.
7
Performance with static and fading signals
We have tested the performance of the proposed algorithm both with static and
dynamic (changing) mixing conditions. In the static case with four DPSK signals
(8 x oversampled) mixed with random matrices the algorithm needs only about
80 sample points (corresponding to 10 symbols) of the mixtures to converge to a
separating solution, whereas a more general algorithm, such as [4], needs about
800-1200 samples for convergence. We attribute this improvement to making much
better use of the baseband signal distributions.
In mobile communications the signals are subject to fading. If there is no direct line
of sight from the transmitter to the receiver, only multiple reflected and diffracted
signal components reach the receiver. When either the receiver or the transmitter
is moving, for example, in an urban environment, these components are changing
very rapidly. If the phases of the carrier signals in these components are aligned
the components add constructively at the receiver. If the phases of carriers are 180
degrees off the components add destructively.
Note that a half of a wavelength difference in the lengths of the paths of the received
components corresponds to a 180 degree phase shift. This is only about 0.17 m
at 900 MHz. Since this small a spatial difference can cause the signal to change
from constructive interference to a null received signal, the result is that both the
amplitude and the phase of the received signal vary seemingly randomly at a rate
that is proportional to relative speeds of the transmitter and the receiver. The
amplitude of the received signal follows a Rayleigh distribution, hence the name
Rayleigh fading. As an example, Figure 3 depicts a 0.1 second fragment of the
amplitude of a fading channel.
10
r
o
_lo F \\.
'-'\ /
I
V
-20
//--\" \1-\/\(\/'\\//-~-'V
i
Ii
-w
o
'V'
\f
\
,
om
0.02
0.03
0.04
0.05
0.06
0.07
0.08
0.()9
0.1
Figure 3: Amplitude (in dB) of a fading radio channel corresponding to a vehicle speed
of 60 mph, when the carrier is 900 Mhz. Horizontal axis is time in seconds.
761
Blind Separation of Radio Signals in Fading Channels
With fading sources, the problem is to be able to adapt to changing conditions,
keeping up with the fading rate. In the signal of Fig. 3 it takes less than 5 milliseconds to move from a peak of the amplitude into a deep fade. Assuming a symbol
rate of 20000 symbols/second, this corresponds to a mere 100 symbols during this
change.
We simulated again DPSK sources oversampling by 8 relative to the symbol rate.
The received sampled mixtures are
(13)
j
where 8j[n] are the source signals, fij[n] represents the fading channel from transmitter j to receiver i, and ni[n] represents the noise observed by receiver i.
In our experiments, we used a sliding window of 80 samples centered at the current
sample. The weight matrix update (the gradient) was calculated using all the
samples of the window, the weight matrix was updated, the window was slid one
sample forward, and the same was repeated. Using this technique we were able
to keep up with the fading rate corresponding to 60 mph relative speed of the
transmitter and the receiver. Figure 4 depicts how the algorithm tracks the fading
channels in the case of three simultaneous source signals.
4
l~ ..
'/,\"
.~. ~
.... ' .
.)~'
.... 1
:~
o
2000
4000
6000
8000
10000
12000
14000
16000
+'------'--~____'__'__~----'-----':
_:E+Ff!'-------'----.L..fc .?.. ?-'-------"'-'.t
o
2000
4000
6000
8000
10000
12000
14000
16000
5r-----~----~----~--~r-----~----~----~--__,
10
O~
____
~-L
_ _L -_ _ _ _L __ _
~L_
____
~
____
~~
__
~
__
~
o
2000
4000
6000
8000
10000
12000
14000
16000
Figure 4: Separation of three signals subject to fading channels. Top graph: The real
parts 16 independent fading channels. 2nd graph: The inverse of the instantaneous fading
conditions (only the real part is depicted) . This is one example of an ideal separation
solution. 3rd graph: The separation solution tracked by the algorithm. (only the real
part is depicted). Bottom graph: The resulting signal/interference (S/I) ratio in dB for
each of the four separated source signals. Horizontal axis is samples. 16000 samples (8 x
oversampled) corresponds to 0.1 seconds.
On the average, the S/I to start with is zero. The average output S/I is 20 dB
for the worst of the three separated signals. Since the mixing is now dynamic
the instantaneous mixing matrix, as determined by the instantaneous fades, can
occasionally be singular and cannot be inverted. Thus the signals at this instance
cannot be separated. In our 0.1 second test signal this occurred four times in the
three source signal case (9 independent fading paths), at which instances the output
762
K. Torkkola
SII bounced to or near zero momentarily for one or more of the separated signals.
Durations of these instances are short, lasting about 15 symbols, and covering about
3 per cent of the total signal time.
8
Related work and discussion
Although the whole field of blind source separation has started around 1985, rather
surprisingly, no application to radio communications has yet emerged. Most of
the source separation algorithms are based on higher-order statistics, and these
should be relatively straightforward to generalize for complex valued baseband data.
Perhaps the main reason is that all theoretical work has concentrated in the case of
static mixing, not in the dynamic case. Many communications channels are dynamic
in nature, and thus rapidly adapting methods are necessary.
Making use of all available knowledge of the sources, in this case the pdf's of the
source signals, allows successful adaptation based on a very small number of samples,
much smaller than by just incorporating the coarse shapes of the pdf's into the
algorithm. It is not unreasonable to presume this knowledge, on the contrary, the
modulation method of a communications system must certainly be known. To
our knowledge, no successful blind separation of signals subject to rapidly varying
mixing conditions, such as fading, has been reported in the literature.
Different techniques applied to separation of various simulated radio signals under
static mixing conditions have been described, for example, in [9, 4]. The maximum
likelihood method reported recently by Yellin and Friedlander [9] seems to be the
closest to our approach, but they only apply it to simulated baseband radio signals
with static mixing conditions.
It must also be noted that channel time dispersion is not taken into account in
our current simulations. This is valid only in cases where the delay spread is short
compared to the inverse of the signal bandwidths. If this is not a valid assumption,
separation techniques for convolutive mixtures, such as in [7] or [1], need to be
combined with the methods developed in this paper.
References
[1] S. Amari, S. Douglas, A. Cichocki, and H. H. Yang. Multichannel blind deconvolution
and equalization using the natural gradient. In Proc. 1st IEEE Signal Processing
Workshop on Signal Processing Advances in Wireless Communications, pages 101104, Paris, France, April 16-18 1997.
[2] A. Bell and T. Sejnowski. An information-maximisation approach to blind separation
and blind deconvolution. Neural Computation, 7(6):1129-1159, 1995.
[3] J.-F. Cardoso. Infomax and maximum likelihood for source separation. IEEE Letters
on Signal Processing, 4(4):112-114, April 1997.
[4] J.-F. Cardoso and B. Laheld. Equivariant adaptive source separation. IEEE 1'ransactions on Signal Processing, 44(12):3017-3030, December 1996.
[5] A. Paulraj and C. B. Papadias. Array processing in mobile communications. In
Handbook of Signal Processing. CRC Press, 1997.
[6] B. A. Pearlmutter and L. C. Parra. A context-sensitive generalization of ICA. In
International Conference on Neural Information Processing, Hong Kong, Sept. 24-27
1996. Springer.
[7] K. Torkkola. Blind separation of convolved sources based on information maximization.
In IEEE Workshop on Neural Networks for Signal Processing, pages 423-432, Kyoto,
Japan, September 4-6 1996.
[8] A.-J. van der Veen and A. Paulraj. An analytical constant modulus algorithm. IEEE
1hmsactions on Signal Processing, 44(5), May 1996.
[9] D. Yellin and B. Friedlander. A maximum likelihood approach to blind separation of
narrowband digital communication signals. In Proc. 30th Asilomar Conf. on Signals,
Systems, and Computers, 1996.
| 1366 |@word bounced:1 kong:1 version:2 seems:1 nd:1 simulation:3 carry:1 reduction:1 fragment:1 imaginary:1 current:3 com:1 yet:1 must:2 visible:1 wx:1 predetermined:1 shape:1 discernible:1 update:2 half:1 plane:6 beginning:2 short:2 awex:2 coarse:1 qam:8 location:2 constructed:2 direct:1 differential:1 sii:1 hermitian:1 ica:1 equivariant:1 motorola:1 window:3 increasing:2 becomes:1 estimating:1 linearity:1 suffice:1 null:1 what:1 developed:1 act:1 exactly:2 scaled:1 uk:3 control:1 unit:5 carrier:6 before:1 limit:2 tempe:1 subscript:1 path:3 modulation:6 might:2 practical:2 yj:5 practice:1 maximisation:1 veen:1 laheld:1 bell:2 hyperbolic:1 adapting:1 radial:4 integrating:1 regular:1 quadrant:2 get:2 cannot:2 close:2 operator:3 context:1 equalization:1 demonstrated:1 straightforward:2 duration:1 fade:2 rule:2 array:5 dw:1 jli:1 coordinate:1 updated:1 user:1 exact:1 element:1 particularly:1 updating:1 observed:3 inserted:2 bottom:1 worst:1 wj:14 momentarily:1 environment:1 ui:1 dynamic:4 depend:1 smart:1 division:1 basis:1 tanh2:2 represented:1 various:1 alphabet:7 derivation:1 separated:8 fast:2 describe:1 sejnowski:2 approached:1 encoded:1 emerged:1 valued:7 say:1 amari:1 statistic:1 antenna:4 final:1 seemingly:1 differentiate:1 differentiable:1 analytical:1 adaptation:12 inserting:1 aligned:1 rapidly:5 mixing:8 intuitive:1 az:1 convergence:3 unmixing:1 leave:1 derive:1 nearest:1 received:5 direction:5 fij:1 correct:2 attribute:1 centered:1 crc:1 require:1 generalization:1 parra:2 insert:1 around:1 intentional:1 lm:1 bump:1 vary:1 consecutive:1 purpose:3 estimation:2 proc:2 radio:12 tanh:12 sensitive:1 vice:1 occupying:1 gaussian:1 sight:1 aim:1 rather:1 mobile:3 varying:3 gut:1 derived:2 ax:1 pdfs:1 improvement:1 transmitter:5 likelihood:8 contrast:1 typically:2 baseband:15 france:1 spatial:3 special:1 equal:3 construct:2 field:1 having:1 paulraj:2 represents:2 look:1 unsupervised:1 others:2 piecewise:1 randomly:1 replaced:1 geometry:1 phase:14 circular:4 certainly:1 mixture:9 necessary:3 circle:6 plotted:1 theoretical:1 lucid:1 instance:3 mhz:2 maximization:9 deviation:1 delay:1 successful:3 too:3 mot:1 reported:2 calibrated:1 combined:1 st:1 density:13 peak:1 international:1 stay:1 destructively:1 off:2 infomax:1 iy:1 again:2 conf:1 derivative:3 japan:1 account:1 nonlinearities:1 wk:2 gaussianity:1 coefficient:1 lzi:1 blind:15 multiplicative:1 vehicle:1 lab:1 reached:1 start:1 slope:1 contribution:1 om:1 square:1 ni:1 characteristic:1 efficiently:1 yield:2 generalize:1 weak:1 modelled:1 mere:1 presume:1 ary:1 simultaneous:2 reach:1 email:1 definition:1 frequency:1 static:6 sampled:2 knowledge:3 amplitude:12 higher:1 supervised:1 reflected:1 april:2 just:2 stage:1 implicit:2 horizontal:2 nonlinear:3 lack:1 perhaps:1 modulus:3 name:1 omitting:1 usa:1 true:3 former:1 hence:1 elliot:1 ll:1 during:1 self:1 essence:1 whereby:1 covering:1 noted:1 hong:1 pdf:20 ridge:1 pearlmutter:2 fj:4 narrowband:1 instantaneous:4 recently:1 phoenix:1 tracked:1 keying:2 jl:2 extend:1 occurred:1 accumulate:1 versa:1 rd:2 grid:2 pointed:1 nonlinearity:1 moving:1 access:1 gj:1 add:2 closest:1 own:1 occasionally:1 binary:1 continue:1 yi:1 devise:1 der:1 inverted:1 transmitted:2 employed:1 converge:1 signal:68 ii:6 sliding:1 multiple:3 corporate:1 kyoto:1 smooth:2 faster:1 adapt:3 divided:1 equally:1 variant:1 denominator:1 expectation:2 izl:1 whereas:2 lzl:3 singular:1 source:26 subject:3 db:3 beamforming:1 contrary:1 december:1 nonstationary:1 near:3 yang:1 ideal:7 enough:2 independence:2 bandwidth:1 shift:3 algebraic:1 cause:1 deep:1 cardoso:3 amount:2 band:1 concentrated:1 processed:1 multichannel:1 oversampling:2 notice:1 millisecond:1 sign:1 track:1 per:1 discrete:3 steepness:4 four:3 urban:1 changing:5 douglas:1 wtw:1 wasted:1 worrying:1 graph:4 sum:3 yellin:2 inverse:3 angle:1 letter:1 separation:20 utilizes:1 coherence:1 scaling:3 simplification:2 accidental:1 fading:22 infinity:1 speed:3 relatively:1 combination:2 smaller:2 unity:1 wi:3 making:5 modification:1 lasting:1 interference:2 taken:2 asilomar:1 equation:7 remains:1 ponder:1 discus:1 needed:2 drastic:1 end:1 available:2 unreasonable:1 apply:3 spectral:1 differentiated:1 appropriate:1 convolved:1 cent:1 denotes:3 top:1 include:1 a4:2 restrictive:1 uj:18 especially:1 approximating:1 move:1 added:1 looked:1 md:1 september:1 gradient:6 separate:2 separating:1 mph:2 parametrized:1 simulated:3 manifold:1 vectorize:1 reason:1 assuming:1 length:2 ratio:1 constructively:1 dispersion:1 finite:1 communication:10 looking:1 incorporate:2 wkl:2 smoothed:1 paris:1 connection:1 oversampled:2 learned:1 discontinuity:1 able:4 usually:1 convolutive:1 oj:1 natural:2 rely:3 axis:3 started:1 cichocki:1 sept:1 literature:1 tangent:1 friedlander:2 relative:3 mixed:2 proportional:1 digital:3 integrate:1 degree:2 interfering:1 squashing:1 lo:1 repeat:1 surprisingly:1 transpose:2 keeping:1 wireless:1 l_:1 wide:1 distributed:2 van:1 bs:4 calculated:1 evaluating:2 cumulative:1 kari:1 transition:1 valid:2 forward:1 commonly:1 made:1 adaptive:1 avoided:1 far:3 keep:1 ml:1 handbook:1 receiver:9 assumed:1 continuous:1 channel:16 learn:1 robust:1 nature:1 ignoring:1 complex:16 did:1 main:2 spread:1 whole:4 noise:1 repeated:1 quadrature:2 fig:3 ff:1 depicts:3 explicit:1 lie:1 wyj:1 symbol:13 constellation:3 torkkola:4 deconvolution:2 incorporating:1 workshop:2 slid:1 depicted:3 rayleigh:2 fc:1 wavelength:1 springer:1 corresponds:3 cdf:8 psk:1 feasible:1 change:2 specifically:1 determined:1 uniformly:1 called:1 total:1 arises:1 constructive:1 evaluate:2 tested:1 |
402 | 1,367 | An Analog VLSI Model of the Fly
Elementary Motion Detector
Reid R. Harrison and Christof Koch
Computation and Neural Systems Program, 139-74
California Institute of Technology
Pasadena, CA 91125
[harrison,koch]@klab.caltech.edu
Abstract
Flies are capable of rapidly detecting and integrating visual motion information in behaviorly-relevant ways. The first stage of visual motion
processing in flies is a retinotopic array of functional units known as elementary motion detectors (EMDs). Several decades ago, Reichardt and
colleagues developed a correlation-based model of motion detection that
described the behavior of these neural circuits. We have implemented a
variant of this model in a 2.0-JLm analog CMOS VLSI process. The result is a low-power, continuous-time analog circuit with integrated photoreceptors that responds to motion in real time. The responses of the
circuit to drifting sinusoidal gratings qualitatively resemble the temporal
frequency response, spatial frequency response, and direction selectivity
of motion-sensitive neurons observed in insects. In addition to its possible engineering applications, the circuit could potentially be used as a
building block for constructing hardware models of higher-level insect
motion integration.
1 INTRODUCTION
Flies rely heavily on visual motion information to survive. In the fly, motion information is
known to underlie many important behaviors including stabilization during flight, orienting
towards small, rapidly-moving objects (Egelhaaf and Borst 1993), and estimating time-tocontact for safe landings (Borst and Bahde 1988). Some motion-related tasks like extending
the legs for landing can be excecuted less than 70 milliseconds after stimulus presentation.
The computational machinery performing this sensory processing is fast, small, low-power,
and robust.
There is good evidence that motion information is first extracted by local elementary motion detectors (see Egelhaaf et al. 1988 and references therein). These EMDs are ar-
An Analog VLSI Model of the Fly Elementary Motion Detector
a)
881
EMD Architecture
photoreceptors
high-pass filters
250
,
300
350
150
200
time[ms)
250
300
350
,
low-pass filters
,
150
200
time[msJ
!
40
<'
.?30
multipliers
E-20
8
subtraction
~
10 __
Ul
EMD output
o
50
100
Figure 1: Elementary Motion Detector. a) A simplified version of our EMD circuit architecture. In the actual circuit implementation, there are separate ON and OFF channels
that operate in parallel. These two channels are summed after the mulipIication. b) The
measured response of the EMD test circuit to a drifting sinusoidal grating. Notice that the
output is phase dependent, but has a positive mean response. If the grating was drifting in
the opposite direction, the circuit would give a negative mean response.
ranged retinotopically, and receive input from adjacent photoreceptors. The properties of
these motion-sensitive units have been studied extensively during the past 30 years. Direct
recording from individual EMDs is difficult due to the small size of the cells, but much
work has been done recording from large tangential cells that integrate the outputs of many
EMDs over large portions of the visual field. From these studies, the behavior of individual
EMDs has been inferred.
If we wish to study models of motion integration in the fly, we first need a model of the
EMD. Since many motion integration neurons in the fly are only a few synapses away
from muscles, it may be possible in the near future to contruct models that complete the
sensorimotor loop. If we wish to include the real world in the loop, we need a mobile
system that works in real time. In the pursuit of such a system, we follow the neuromorphic
engineering approach pioneered by Mead (Mead 1989) and implement a hardware model of
the fly EMD in a commercially available 2.0-J-tm CMOS VLSI process. All data presented
in this paper are from one such chip.
2 ALGORITHM AND ARCHITECTURE
Figure la shows a simplified version of the motion detector. This is an elaborated version of
the correlation-based motion detector first proposed by Reichardt and colleagues (see Reichardt 1987 and references therein). The Reichardt motion detector works by correlating
(by means of a multiplication) the response of one photoreceptor to the delayed response
of an adjacent photoreceptor. Our model uses the phase lag inherent in a low-pass filter
to supply the delay. The outputs from two mirror-symmetric correlators are subtracted to
remove any response to full-field flicker (ws = 0, Wt > 0).
Correlation-based EMDs are not pure velocity sensors. Their response is strongly affected
by the contrast and the spatial frequency components of the stimulating pattern. They
can best be described as direction-selective spatiotemporal filters. The mean steady-state
response R of the motion detector shown in Figure I a to a sinusoidal grating drifting in one
direction can be expressed as a separable function of stimulus amplitude (D.I), temporal
R. R. Harrison and C. Koch
882
c)
b)
a)
dV I
.. component 0fCTt
poSl\lve
t
v
+
negative component of C ~; +
dIout
t
dV
. cTt
1=
it
t
+ lout = lin
= CUr
It
Figure 2: EMD Subcircuits. a) Temporal derivative circuit. In combination with the firstorder low-pass filter inherent in the photoreceptor, this forms the high-pass filter with time
constant TH. The feedback amplifier enforces V = Yin, and the output is the current
needed for the nFET or pFET source follower to charge or discharge the capacitor C.
b) Current-mode low-pass filter. The time constant TL is determined by the bias current IT
(which is set by a bias voltage supplied from off-chip), the capacitance C, and the thermal
voltage UT = kT /q. c) Current-mode one-quadrant multiplier. The devices shown are
floating-gate nFETs. Two control gates capacitively couple to the floating node, forming a
capacitive divider.
frequency (Wt = 21r It), and spatial frequency (ws = 21r is):
(1)
(2)
where D.cp is the angular separation of the photoreceptors, TH is the time constant of the
high-pass filter, and TL is the time constant of the low-pass filter (see Figure 1a). (Note
that this holds only for motion in a particular direction. Motion detectors are not linearly
separable overall, but the single-direction analysis is useful for making comparisons.)
3 CIRCUIT DESCRIPTION
In addition to the basic Reichardt model described above, we include a high-pass filter in
series with the photoreceptor. This amplifies transient responses and removes the DC component of the photoreceptor signal. We primarily use the high-pass filter as a convenient
circuit to switch from a voltage-mode to a current-mode representation (see Figure 2a).
For the photoreceptor, we use an adaptive circuit developed by Delbruck (Delbruck and
Mead 1996) that produces an output voltage proportional to log intensity. We bias the photoreceptor very weakly to attenuate high temporal frequencies. This is directly followed by
a temporal derivative circuit (Mead 1989) (see Figure 2a), the result being a high-pass filter
with the dominant pole TH being set by the photoreceptor cutoff frequency. The outputs
of the temporal derivati?,re circuit are two unidirectional currents that represent the positive
and negative components of a high-pass filtered version of the photoreceptor output. This
resembles the ON and OFF channels found in many biological visual systems. Some studies suggest ON and OFF channels are present in the fly (Franceschini et al. 1989) but the
evidence is mixed (Egelhaaf and Borst 1992). This two-channel representation is useful for
current-mode circuits, since the following translinear circuits work only with unidirectional
An Analog VLSI Model of the Fly Elementary Motion Detector
883
0.8
0.2
0.1
1
10
Temporal Frequency J, [Hz]
100
Figure 3: Temporal Frequency Response. Circuit data was taken with is = 0.05 cyclesldeg
and 86% contrast. Theory trace is Rt(Wt) from Equation 2, where TH = 360 ms and TL
= 25 ms were directly measured in separate experiments - these terms were not fit to the
data. Insect data was taken from a wide-field motion neuron in the blowfly Calliphora
erythrocephala (O ' Carroll et al. 1996). All three curves were normalized by their peak
response.
currents. It should be noted that the use of ON and OFF channels introduces nonlinearities
into the circuit that are not accounted for in the simple model described by Equation 2.
The current-mode low-pass filter is shown in Figure 2b. The time constant TL is set by the
bias current 11'. This is a log-domain filter that takes advantage of the exponential behavior
of field-effect transistors (FETs) in the subthreshold region of operation (Minch, personal
communication).
The current-mode multiplier is shown in Figure 2c. This circuit is also translinear, using a
diode-connected FET to convert the input currents into log-encoded voltages. A weighted
sum of the voltages is computed with a capacitive divider, and the resulting voltage is
exponentiated by the output FET into the output current. The capacitive divider creates
a floating node, and the charge on all these nodes must be equalized to ensure matching
across independent multipliers. This is easily accomplished by exposing the chip to UV
light for several minutes. This circuit represents one of a family of floating-gate MaS
translinear circuits developed by Minch that are capable of computing arbitrary power laws
in current mode (Minch et al. 1996).
After the multiplication stage, the currents from the ON and OFF channels are summed,
and the final subtraction of the left and right channels is done off-chip. There is a gain
mismatch of approximately 2.5 between the left and right channels that is now compensated
for manually. This mismatch must be lowered before large on-chip arrays of EMDs are
practical. A new circuit designed to lessen this gain mismatch is currently being tested.
It is interesting to note that there is no significant offset error in the output currents from
each channel. This is a consequence of using translinear circuits which typically have gain
errors due to transistor mismatch, but no fixed offset errors.
4 EXPERIMENTS
As we showed in Equation 2, the motion detector's response to a drifting sinusoidal grating
of a particular direction should be a separable function of Ill, temporal frequency, and
R. R. Harrison and C. Koch
884
~con
08
..... .ID.!'ory
__ ._i~ect
,,
,,
~ 0.6
,
...,,
2-
!1
~
c
e
0,4
\
\
::.:
.~
...i}.
\
~
0.2
...
?.....
\
/'c'?????
~"'~/
~, 2
0,001
,
\
0,01
0,1
Spatial FrequencyIs [cycleoJdeg]
Figure 4: Spatial Frequency Response. Circuit data was taken with it = 4 Hz and 86%
contrast. Theory trace is Rs(w s ) from Equation 2 multiplied by exp( -w s 2/ K2) to account
for blurring in the optics, The photoreceptor spacing ~rp = 1.9 0 was directly measured in
an separate experiment. Only K and the overall magnitude were varied to fit the data.
Insect data was taken from a wide-field motion neuron in the hoverfly Volucella pelluscens
(O'Carroll et al. 1996). Circuit and insect data were nonnalized by their peak response.
spatial frequency, We tested the circuit along these axes using printed sinusoidal gratings
mounted on a rotating drum. A lens with an 8-mm focal length was mounted over the chip.
Each stimulus pattern had a fixed contrast ~I /21 and spatial frequency fs. The temporal
frequency was set by the pattern's angular velocity vas seen by the chip, where it = fsv.
The response of the circuit to a drifting sine wave grating is phase dependent (see Figure I b). In flies, this phase dependency is removed by integrating over large numbers of
EMDs (spatial integration). In order to evaluate the perfonnance of our circuit, we measured the mean response over time.
Figure 3 shows the temporal frequency response ofthe circuit as compared to theory, and to
a wide-field motion neuron in the fly. The circuit exhibits temporal frequency tuning. The
point of peak response is largely detennined by n, and can be changed by altering the lowpass filter bias current. The deviation of the circuit behavior from theory at low frequencies
is thought to be a consequence of crossover distortion in the temporal derivative circuit. At
high temporal frequencies, parasitic capacitances in current mirrors are a likely candidate
for the discrepancy. The temporal frequency response of the blowfly Calliphora is broader
than both the theory and circuit curves. This might be a result of time-constant adaptation
found in blowfly motion-sensitive neurons (de Ruyter van Steveninck et at. 1986).
Figure 4 shows the spatial frequency response of the circuit. The response goes toward zero
as Ws approaches zero, indicating that the circuit greatly attenuates full-field flicker. The
circuit begins aliasing at Ws = 1/2~rp, giving a response in the wrong direction. Spatial
aliasing has also been observed in flies (Gotz 1965). The optics used in the experiment act
as an antialiasing filter, so aliasing could be avoided by defocusing the lens slightly.
Figure 5 shows the directional tuning of the circuit. It can be shown that as long as the
spatial wavelength is large compared to ~rp, the directional sensitivity of a correlationbased motion detector should approximate a cosine function (Zanker 1990). The circuit's
perfonnance matches this quite well. Motion sensitive neurons in the fly show cosine-like
direction selectivity.
Figure 6 shows the contrast response of the circuit. Insect EMDs show a saturating contrast
An Analog VLSI Model of the Fly Elementary Motion Detector
~con
0.8
~
.... ..t!J~ory
__ .J!!~ect
0.6
\
,\
''.
~\, '.
9. 0.4
.
?
,, ''.
'
~ 02
, '.
co::
~
i
a
~ri"')I.'
-0.2
~-o.4
885
lO-.
.t"
_ ...
-0.6
"
-0.8
-1
-150
-100
-50
a
50
100
150
Direction of MolIOn [deg)
Figure 5: Directional Response. Circuit data was taken with it = 6 Hz, is = 0.05 cycles/deg
and 86% contrast. Theory trace is cos 0:, where 0: is the direction of motion relative to
the axis along the two photoreceptors. Insect data was taken from the HI neuron in the
blowfly Calliphora erythrocephala (van Hateren 1990). HI is a spiking neuron with a low
spontaneous firing rate. The flattened negative responses visible in the graph are a result of
the cell's limited dynamic range in this region. All three curves were normalized by their
peak response.
response curve, which can be accounted for by introducing saturating nonlinearities before
the multiplication stage (Egelhaaf and Borst 1989). We did not attempt to model contrast
saturation in our circuit, though it could be added in future versions.
5
CONCLUSIONS
We implemented and tested an analog VLSI model of the fly elementary motion detector.
The circuit's spatiotemporal frequency response and directional selectivity is qualitatively
similar to the responses of motion-sensitive neurons in the fly. This circuit could be a useful
building block for constructing analog VLSI models of motion integration in flies. As an
integrated, low-power, real-time sensory processor, the circuit may also have engineering
applications.
Acknowledgements
This work was supported by the Center for Neuromorphic Systems Engineering as a part
of NSF's Engineering Research Center program, and by ONR. Reid Harrison is supported
by an NDSEG fellowship from ONR. We thank Bradley Minch, Ho1ger Krapp, and Rainer
Deutschmann for invaluable discussions.
References
A. Borst and S. Bahde (1988) Visual information processing in the fly's landing system. 1.
Compo Physiol. A 163: 167-173.
T. Delbruck and C. Mead (1996) Analog VLSI phototransduction by continuous-time,
adaptive, logarithmic photoreceptor circuits. CNS Memo No . 30, Caltech.
M. Egelhaaf, K. Hausen, W. Reichardt, and C. Wehrhahn (1988) Visual course control in
886
R. R. Harrison and C. Koch
/ ,'JI.
~con
.... --
....?.t!'!'ory
08
,'
__.j~~ct
"i'!
8-
;l
~06
>"
I
?
I
I
<.>
::;
.~O.4
I
,-
/'
I
U
I
'" 02
.,/
I
I
I
I
..... i ........
I
-....-
10
20
30
40
50
60
70
80
90
100
Contrast &112[[%]
Figure 6: Contrast Response. Circuit data was taken with It = 6 Hz and is = 0.1 cycles/deg.
Theory trace is Rj(D.I) from Equation 2 with its magnitude scaled to fit the circuit data.
Insect data was taken from the HS neuron in the blowfly Calliphora erythrocephala (Egelhaaf and Borst 1989). Circuit and insect data were normalized by their peak response.
flies relies on neuronal computation of object and background motion. TINS 11: 351-3 5 8.
M. Egelbaaf and A. Borst (1989) Transient and steady-state response properties of movement detectors. J. Opt. Soc. Am. A 6: 116-127.
M. Egelbaaf and A. Borst (1992) Are there separate ON and OFF channels in fly motion
vision? Visual Neuroscience 8: 151-164.
M. Egelbaaf and A. Borst (1993) A look into the cockpit of the fly: Visual orientation,
algorithms, and identified neurons. J. Neurosci. 13: 4563-4574.
N. Franceschini, A. Riehle, and A. Ie Nestour (1989) Directionally selective motion detection by insect neurons. In StavengaIHardie (eds.), Facets o/Vision, Berlin: Springer-Verlag.
K.G. Gotz (1965) Die optischen Ubertragungseigenschaften der Komplexaugen von
Drosophila. Kybemetik2: 215-221.
J.H. van Hateren (1990) Directional tuning curves, elementary movement detectors, and
the estimation of the direction of visual movement. Vision Res. 30: 603-614.
C. Mead (1989) Analog VLSI and Neural Systems. Reading, Mass.: Addison-Wesley.
B.A. Minch, C. Diorio, P. Hasler, and C. Mead (1996) Translinear circuits using subthreshold floating-gate MaS transistors. Analog Int. Circuits and Signal Processing 9: 167-179.
D.C. O'Carroll, N.J. Bidwell, S.B. Laughlin, and EJ. Warrant (1996) Insect motion detectors matched to visual ecology. Nature 382: 63-66.
W. Reichardt (1987) Evaluation of optical motion information by movement detectors. J.
CompoPhys. A 161: 533-547.
R.R. de Ruyter van Steveninck, W.H. Zaagman, and H.A.K. Mastebroek (1986) Adaptation
of transient responses of a movement-sensitive neuron in the visual system of the blowfly
Calliphora erythrocephala. Bioi. Cybern. 54: 223-236.
J.M. Zanker (1990) On the directional sensitivity of motion detectors. Bioi. Cybern. 62:
177-183.
| 1367 |@word h:1 version:5 r:1 series:1 past:1 bradley:1 current:18 follower:1 must:2 exposing:1 physiol:1 visible:1 remove:2 designed:1 device:1 compo:1 filtered:1 detecting:1 node:3 along:2 direct:1 supply:1 ect:2 behavior:5 aliasing:3 borst:9 actual:1 begin:1 retinotopic:1 estimating:1 matched:1 circuit:50 mass:1 developed:3 jlm:1 temporal:15 firstorder:1 charge:2 act:1 k2:1 wrong:1 scaled:1 control:2 unit:2 underlie:1 zaagman:1 christof:1 reid:2 positive:2 before:2 engineering:5 local:1 consequence:2 id:1 mead:7 firing:1 approximately:1 might:1 therein:2 studied:1 resembles:1 co:2 limited:1 range:1 steveninck:2 practical:1 enforces:1 block:2 implement:1 crossover:1 thought:1 printed:1 convenient:1 matching:1 integrating:2 quadrant:1 suggest:1 deutschmann:1 cybern:2 landing:3 compensated:1 center:2 go:1 pure:1 array:2 discharge:1 spontaneous:1 heavily:1 pioneered:1 us:1 velocity:2 observed:2 fly:23 region:2 connected:1 cycle:2 diorio:1 movement:5 removed:1 calliphora:5 dynamic:1 personal:1 weakly:1 creates:1 blurring:1 fsv:1 easily:1 lowpass:1 chip:7 fast:1 equalized:1 quite:1 lag:1 encoded:1 distortion:1 final:1 directionally:1 advantage:1 transistor:3 adaptation:2 relevant:1 loop:2 rapidly:2 detennined:1 description:1 amplifies:1 extending:1 produce:1 cmos:2 object:2 measured:4 grating:7 soc:1 implemented:2 diode:1 resemble:1 direction:12 safe:1 filter:16 stabilization:1 translinear:5 transient:3 drosophila:1 opt:1 biological:1 elementary:9 hold:1 mm:1 koch:5 klab:1 exp:1 mastebroek:1 estimation:1 currently:1 sensitive:6 weighted:1 sensor:1 contruct:1 ej:1 mobile:1 voltage:7 broader:1 rainer:1 ax:1 greatly:1 contrast:10 am:1 dependent:2 integrated:2 typically:1 pasadena:1 vlsi:10 w:4 selective:2 overall:2 ill:1 orientation:1 insect:11 spatial:11 integration:5 summed:2 field:7 emd:7 manually:1 represents:1 look:1 survive:1 warrant:1 discrepancy:1 future:2 commercially:1 stimulus:3 inherent:2 primarily:1 zanker:2 few:1 tangential:1 individual:2 delayed:1 lessen:1 floating:5 phase:4 cns:1 attempt:1 amplifier:1 detection:2 ecology:1 evaluation:1 introduces:1 light:1 kt:1 capable:2 machinery:1 perfonnance:2 capacitively:1 rotating:1 re:2 facet:1 ar:1 altering:1 delbruck:3 neuromorphic:2 pole:1 deviation:1 introducing:1 ory:3 delay:1 dependency:1 spatiotemporal:2 minch:5 peak:5 sensitivity:2 ie:1 off:8 von:1 ndseg:1 bahde:2 derivative:3 account:1 sinusoidal:5 nonlinearities:2 de:2 int:1 sine:1 franceschini:2 portion:1 wave:1 parallel:1 unidirectional:2 elaborated:1 largely:1 subthreshold:2 ofthe:1 directional:6 antialiasing:1 ago:1 processor:1 detector:20 synapsis:1 ed:1 colleague:2 frequency:21 sensorimotor:1 con:3 cur:1 couple:1 gain:3 ut:1 amplitude:1 wesley:1 higher:1 follow:1 response:36 done:2 though:1 strongly:1 angular:2 stage:3 correlation:3 flight:1 mode:8 orienting:1 building:2 effect:1 ranged:1 multiplier:4 divider:3 normalized:3 symmetric:1 riehle:1 adjacent:2 during:2 noted:1 steady:2 die:1 fet:3 cosine:2 m:3 complete:1 motion:43 cp:1 invaluable:1 functional:1 spiking:1 ji:1 retinotopically:1 analog:11 significant:1 attenuate:1 tuning:3 uv:1 focal:1 phototransduction:1 had:1 moving:1 lowered:1 carroll:3 dominant:1 showed:1 ctt:1 selectivity:3 verlag:1 onr:2 accomplished:1 der:1 caltech:2 muscle:1 seen:1 erythrocephala:4 subtraction:2 signal:2 full:2 rj:1 match:1 long:1 lin:1 va:1 variant:1 basic:1 vision:3 represent:1 lve:1 cell:3 receive:1 addition:2 fellowship:1 background:1 spacing:1 harrison:6 source:1 operate:1 recording:2 hz:4 capacitor:1 correlators:1 near:1 switch:1 fit:3 architecture:3 identified:1 opposite:1 drum:1 tm:1 ul:1 f:1 useful:3 extensively:1 hardware:2 supplied:1 flicker:2 nsf:1 millisecond:1 notice:1 neuroscience:1 affected:1 cutoff:1 hasler:1 graph:1 year:1 convert:1 sum:1 family:1 separation:1 hi:2 ct:1 followed:1 optic:2 ri:1 subcircuits:1 performing:1 separable:3 optical:1 combination:1 pfet:1 across:1 slightly:1 making:1 leg:1 dv:2 gotz:2 defocusing:1 emds:9 taken:8 equation:5 needed:1 addison:1 pursuit:1 available:1 operation:1 multiplied:1 away:1 blowfly:6 subtracted:1 gate:4 drifting:6 rp:3 capacitive:3 include:2 ensure:1 giving:1 capacitance:2 added:1 rt:1 responds:1 exhibit:1 separate:4 thank:1 berlin:1 toward:1 length:1 nfet:1 difficult:1 potentially:1 trace:4 negative:4 memo:1 implementation:1 attenuates:1 neuron:14 thermal:1 communication:1 nonnalized:1 dc:1 varied:1 arbitrary:1 intensity:1 inferred:1 california:1 pattern:3 mismatch:4 lout:1 reading:1 program:2 saturation:1 including:1 power:4 rely:1 technology:1 axis:1 reichardt:7 acknowledgement:1 multiplication:3 relative:1 law:1 mixed:1 interesting:1 proportional:1 mounted:2 integrate:1 correlationbased:1 lo:1 course:1 changed:1 accounted:2 supported:2 bias:5 exponentiated:1 laughlin:1 institute:1 wide:3 van:4 feedback:1 curve:5 world:1 sensory:2 hausen:1 qualitatively:2 adaptive:2 simplified:2 avoided:1 approximate:1 deg:3 correlating:1 photoreceptors:5 continuous:2 decade:1 channel:11 nature:1 robust:1 ca:1 ruyter:2 krapp:1 constructing:2 domain:1 did:1 linearly:1 neurosci:1 egelhaaf:6 neuronal:1 tl:4 wish:2 exponential:1 candidate:1 tin:1 minute:1 msj:1 offset:2 evidence:2 flattened:1 mirror:2 magnitude:2 yin:1 logarithmic:1 wavelength:1 likely:1 forming:1 visual:12 expressed:1 saturating:2 springer:1 relies:1 extracted:1 ma:2 stimulating:1 bioi:2 presentation:1 towards:1 determined:1 wt:3 lens:2 pas:13 la:1 photoreceptor:11 indicating:1 parasitic:1 hateren:2 evaluate:1 tested:3 |
403 | 1,368 | Computing with Action Potentials
John J. Hopfield*
Carlos D. Brody t
Sam Roweis t
Abstract
Most computational engineering based loosely on biology uses continuous variables to represent neural activity. Yet most neurons communicate with action potentials. The engineering view is equivalent to using
a rate-code for representing information and for computing. An increasing number of examples are being discovered in which biology may not
be using rate codes. Information can be represented using the timing of
action potentials, and efficiently computed with in this representation.
The "analog match" problem of odour identification is a simple problem
which can be efficiently solved using action potential timing and an underlying rhythm. By using adapting units to effect a fundamental change
of representation of a problem, we map the recognition of words (having uniform time-warp) in connected speech into the same analog match
problem. We describe the architecture and preliminary results of such a
recognition system. Using the fast events of biology in conjunction with
an underlying rhythm is one way to overcome the limits of an eventdriven view of computation. When the intrinsic hardware is much faster
than the time scale of change of inputs, this approach can greatly increase
the effective computation per unit time on a given quantity of hardware.
1 Spike timing
Most neurons communicate using action potentials - stereotyped pulses of activity that are
propagated along axons without change of shape over long distances by active regenerative
processes. They provide a pulse-coded way of sending information. Individual action
potentials last about 2 ros. Typical active nerve cells generate 5-100 action potentials/sec.
Most biologically inspired engineering of neural networks represent the activity of a nerve
cell by a continuous variable which can be interpreted as the short-time average rate of
generating action potentials. Most traditional discussions by neurobiologists concerning
how information is represented and processed in the brain have similarly relied on using
"short term mean firing rate" as the carrier of information and the basis for computation.
But this is often an ineffective way to compute and represent information in neurobiology.
t
*Dept. of Molecular Biology, Princeton University. jhopfield@watson.princeton. edu
Computation & Neural Systems, California Institute of Technology.
Computing with Action Potentials
167
To define "short term mean firing rate" with reasonable accuracy, it is necessary to either
wait for several action potentials to arrive from a single neuron, or to average over many
roughly equivalent cells. One of these necessitates slow processing; the other requires
redundant "wetware".
Since action potentials are short events with sharp rise times, action potential timing is
another way that information can be represented and computed with ([Hopfield, 1995]).
Action potential timing seems to be the basis for some neural computations, such as the
determination of a sharp response time to an ultrasonic pulse generated by the moustache
bat. In this system, the bat generates a 10 ms pulse during which the frequency changes
monotonically with time (a "chirp"). In the cochlea and cochlear nucleus, cells which
are responsive to different frequencies will be sequentially driven, each producing zero
or one action potentials during the time when the frequency is in their responsive band.
These action potentials converge onto a target cell. However, while the times of initiation
of the action potentials from the different frequency bands are different, the length and
propagation speed of the various axons have been coordinated to result in all the action
potentials arriving at the target cell at the same time, thus recognizing the "chirped" pulse
as a whole, while discriminating against random sounds of the same overall duration.
Taking this hint from biology, we next investigate the use of action potential timing to represent information and compute with in one of the fundamental computational problems
relevant to olfaction, noting why the elementary "neural net" engineering solution is poor,
and showing why computing with action potentials lacks the deficiencies of the conventional elementary solution.
2 Analog match
The simplest computational problem of odors is merely to identify a known odor when a
single odor dominates the olfactory scene. Most natural odors consist of mixtures of several molecular species. At some particular strength a complex odor b can be described by
the concentrations Nt of its constitutive molecular of species i. If the stimulus intensity
changes, each component increases (or decreases) by the same multiplicative factor. It is
convenient to describe the stimulus as a product of two factors, an intensity .A and normalized components n~ as:
.A
= ~jNJ
=>
b _ Nbj
\
i A
ni -
or
Nbi -Ani
_ \ b
(1)
The n~ are normalized, or relative concentrations of different molecules, and .A describes
the overall odor intensity. Ideally, a given odor quality is described by the pattern of n~ ,
which does not change when the odor intensity .A changes. When a stimulus s described
by a set {NJ} is presented, an ideal odor quality detector answers "yes" to the question "is
odor b present?" if and only if for some value of .A:
N~~.An~
J
3
't/J'
(2)
This general computation has been called analog match. 1
The elementary "new?at net" way to solve analog match and recognize a single odor independent of intensity would bt: to use a single "grandmother unit" of the following type.
I The analog match problem of olfaction is actually viewed through olfactory receptor cells. Studies of vertebrate sensory cells have shown that each molecular species stimulates many different
sensory cells, and each cell is excited by many different molecular species. The pattern of relative
excitation across the population of sensory cell classes determines the odor quality in the generalist
olfactory system. There are about 1000 broadly responsive cell types; thus, the olfactory systems of
higher animals apparently solve an analog match problem of the type described by (2), except that
the indices refer to cell types, and the actual dimension is no more than 1000.
168
1. J. Hopfield, C. D. Brody and S. Roweis
Call the unknown odor vector I, and the weight vector W. The input to the unit will then
be I . W. If W = n / lin II and I is pre-normalized by dividing by the Euclidean magnitude
11111, recognition can be identified by I . W > .95 , or whatever threshold describes the
degree of precision in identification which the task requires.
This solution has four major weaknesses.
1. Euclidean normalization is used; not a trivial calculation for real neural hardware.
2. The size of input components Ik and their importance is confounded. If a weak
component has particular importance, or a strong one is not reliable, there is no
way to represent this. W describes only the size of the target odor components.
3. There is no natural composition if the problem is to be broken into a hierarchy
by breaking the inputs into several parts, solving independently, and feeding these
results on to a higher level unit for a final recognition. This is best seen by analogy
to vision. If I recognize in a picture grandmother's nose at one scale, her mouth
at another, and her right eye at a third scale, then it is assuredly not grandmother.
Separate normalization is a disaster for creating hierarchies.
4. A substantial number of inputs may be missing or giving grossly wrong information. The "dot-product-and-threshold" solution cannot contend with this problem.
For example, in olfaction, two of the common sources of noise are the adaptation
of a subset of sensors due to previous strong odors, and receptors stuck "on" due
to the retention of strongly bound molecules from previous odors.
All four problems are removed when the information is encoded and computed with in an
action potential representation, as illustrated below.. The three channels of analog input
Ia'/b, Ie are illustrated on the left. They are converted to a spike timing representation by
the position of action potentials with respect to a fiducial time T. The interval between
T and the time of an action potential in a channel j is equal to log Ij. Each channel is
connected to an output unit through a delay line of length t,.j = logn~, where n b is the
target vector to be identified. When the analog match criterion is satisAed, the pulses on
all three channels will arrive at the target unit at the same time, driving it strongly. If all
inputs are scaled by a, then the times of the action potentials will all be changed by log a.
The three action potentials will arrive at the recognition unit simultaneously, but a a time
shifted by loga. Thus a pattern can be recognized (or not) on the basis of its relative
components. Scale information is retained in the time at which the recognition unit is
driven. The system clearly "composes", and difficulty (3) is sunnounted. No normalization
is required, eliminating difficulty (1). Each pathway has two parameters describing it, a
delay (which contains the information about the pattern to be recognized) and a synaptic
strength (which describes the weight of the action potential at the recognition unit). Scale
and importance are separately represented. The central computational motif is very similar
to that used in bat sonar, using relative timing to represent information and time delays to
represent target patterns.
1\
~--I
a t ilOg
b
~ log
C
?? ) log
Imin
Imax
1
1
o
:j
m~a
~
~
---t>@recognitionunitsums
EPSPs then thresholds
t - -.....-+~
._
T';
Delays set
prototype pattern
0
T
~
Weights set relath'c
feature importance
Computing with Action Potentials
169
This system also tolerates errors due to missing or grossly inaccurate information. The
figure below illustrates this fact for the case of three inputs, and contrasts the receptive fields
of a system computing with action potentials with those of a conventional grandmother
cell. (The only relevant variables are the projections of the input vector on the surface of
the unit sphere, as illustrated.) When the thresholds are set high, both schemes recognize a
small, roughly circular region around the target pattern (here chosen as 111). Lowering the
recognition threshold in the action-potential based scheme results in a star-shaped region
being recognized; this region can be characterized as "recognize if any two components
are in the correct ratio, independent of the size of the third component." Pattern 110 is thus
recognized as being similar to 111 while still rejecting most of the space as not resembling
the target. In contrast, to recognize 110 with the conventional unit requires such threshold
lowering that almost any vector would be recognized.
Spike Timing
thresh =0.7
0.6
Normalize, Dot Product
0.4
thresh =0.99 0.95
0.90
This method of representation and computation using action potential timing requires a
fiducial time available to all neurons participating in stimulus encoding. Fiducial times
might be externally generated by salient events, as they are in the case of moustache bat
sonar. Or they could be internally generated, sporadically or periodically. In the case of
the olfactory system, the first processing area of all animals has an oscillatory behavior. A
large piece of the biophysics of neurons can be represented by the idea that neurons are
leaky integrators, and that when their internal potential is pushed above a critical value,
they produce an action potential, and their internal potential is reset a fixed distance below
threshold. When a sub-threshold input having a frequency j is combined with a steady
analog current I, the system generates action potentials at frequency j, but whose phase
with respect to the underlying oscillation is a monotone function of I. Thus the system
encodes I into a phase (or time) of an action potential with respect to the underlying rhythm.
Interestingly, in mammals, the second stage of the olfactory system, the prepiriform cortex,
has slow axons propagating signals across it. The propagation time delays are comparable
to 1/ j. The system has the capability of encoding and analyzing information in action
potential timing.
3 Time warp and speech
Recognizing syllables or words independent of a uniform stretch (''uniform time warp")
can in principle be cast as an analog match problem and transformed into neural variables [Hopfield, 1996]. We next describe this approach in relationship to a previous "neural network" way of recognizing words in connected speech [Hopfield and Tank, 1987,
Unnikrishnan et aI., 1991, Unnikrishnan et aI., 1992] (URT for short).
A block diagram below shows the UHT neural network for recognizing a small vocabulary
of words in connected speech. The speech signal is passed through a bank of band-pass
filters, and an elementary neural feature detector then examines whether each frequency is
a local maximum of the short-term power spectrum. If so, it propagates a "1" down a delay
line from that feature detector, thus converting the pattern of features in time into a pattern
in space. The recognition unit for a particular word is then connected to these delay lines
by a pattern of weights which are trained on a large data base.
1. 1. Hopfield, C. D. Brody and S. Roweis
170
Time De\ays
UHT
diagram
Time De\ays
Time-warp
diagram
The conceptual strength of this circuit is that it requires no indication of the boundaries
between words. Indeed, there is no such concept in the circuit. The conceptual weakness
of this "neural network" is that the recognition process for a particular word is equivalent to
sliding a rigid template across the feature pattern. Unfortunately, even a single speaker has
great variation in the duration of a given word under different circumstances, as illustrated
in the two spectrograms below. Clearly no single template will fit these both of these
utterances of "one" very well. This general problem is known as time-warp. A time-warp
invariant recognizer would have considerable advantage.
Two instances of "one"
[~~Ci
00
0.3
Smoothed spectrogram Thresholding
00
0.3
Time since feature
~~t~
00
0.3
00
0.3
The UHT approach represents a sequence by the presence of a signal on feature signal lines
A, B, C, as shown on the left of the figure below. Suppose the end of the word occurs at
some particular time as indicated. Then the feature starts and stops can be described as an
analog vector of times, whose components are shown by the arrows as indicated. In this
representation, a word which is spoken more slowly simply has all its vector components
multiplied by a common factor. The problem of recognizing words within a uniform time
warp is thus isomorphic with the analog match problem, and can be readily solved by
using action potential timing and an underlying rhythm, as described above. In our present
modeling, the rhythm has a frequency of 50 Hz, significantly faster than the rate at which
new features appear in speech. This frequency corresponds to the clock rate at which
speech features are effectively "sampled". In the UHT circuit this rate was set by the
response timescale of the recognition units. But where each template in the UHT circuit
attempted only a single match with the feature vector per sample, this circuit allows the
attempted match of many possible time-warps with the feature vector per sample. (The
range of time-warps allowed is determined by the oscillation frequency and the temporal
resolution of the spike timing system.)
A: 000111111000001
B: 00111100000000
C: 00000001111000
A:
new.,
B:
rep
C:
start feature
stop feature
The block diagram of the neural circuit necessary to recognize words in connected speech
with uniform time warp is sketched above. It looks superficially similar to the UHT circuit
beside it. except for the insertion of a ramp generator and a phase encoder between the
Computing with Action Potentials
171
feature detectors and the delay system. Recognizing a feature activates a ramp generator
whose output decays. This becomes the input to a "neuron" which has an additional oscillatory input at frequency f. If the ramp decay and oscillation shapes are properly matched,
the logarithm of the time since the occurrence of a feature is encoded in action potential
timing as above. Following this encoding system there is a set of tapped delay lines of the
same style which would have been necessary to solve the olfactory decoding problem. The
total the amount of hardware is similar to the UHT approach because the connections and
delay lines dominate the resource requirements.
The operation of the present circuit is, however, entirely different. What the present circuit
does is to "remember" recent features by using ramp generators, encode the logarithms
of times since features into action potential timing, and recognize the pattern with a timedelay circuit. The time delays in the present circuit have an entirely different meaning from
those of the UHT circuit, since they are dimensionally not physical time, but instead are
a representation of the logarithm of feature times. The time delays are only on the scale
of 1/f rather than the duration of a word. There are simple biological implementations of
these ideas. For example, when a neuron responds, as many do, to a step in its input by
generating a train of action potentials with gradually falling firing frequency (adaptation),
the temporal spacing between the action potentials is an implicit representation of the time
since the "step" occurred (see [Hopfield, 1996]).
For our initial engineering investigations, we used very simple features. The power within
each frequency band is merely thresholded. An upward crossing of that threshold represents a "start" feature for that band, and a downward crossing a "end" feature. A pattern of
such features is identified above beside the spectrograms. Although the pattern of feature
vectors for the two examples of "one" do not match well because of time warp, when the
logarithms of the patterns are taken, the difference between the two patterns is chiefly a
shift, i.e. the dominant difference between the patterns is merely uniform time warp.
To recognize the spoken digit "one", for example, the appropriate delay for each channel
was chosen so as to minimize the variance of the post-delay spike times (thus aligning the
spikes produced by all features), averaged over the different exemplars which contained
that feature. All channels with a feature present were given a unity weight connection at
that delay value; inactive channels were given weight zero. The figure below shows, on the
left, the spike input to the recognition unit (top) and the sum of the EPSPs caused by these
inputs (bottom). The examples of "one" produced maximum outputs in different cycles
of the oscillation, corresponding to the actual "end times" at which the words should be
viewed as recognized. Only the maximum cycle for each utterance is shown here. Within
their maximum cycle, different examples of the utterances produced maximal outputs at
different phases of the cycle, corresponding to the fact that the different utterances were
recognized as having different time warp factors. The panels on the right show the result
of playing spoken "four"s into the same recognition unit.
Three "ones"s
III
Q)
Several "four"s
.....;... ...
.<
,?
~.::~~ :,
~
?
,,_ ......P": " ?. - ... . .... . .
::J
iii
_
Q)
f~.{
~----~'~"'~
" ---------
.~
>
~
:1::
c:::
threshold
::J
"5
a.
S~---'-"~~------
o
time within cycle
time within cycle
172
J 1. Hopfield, C. D. Brody and S. Roweis
There is no difficulty in distinguishing "ones" from other digits. When, however, the possibility of adjusting the time-warp is turned off, resulting in a "rigid" template it was not
possible to discriminate between "one" and other digits. (Disabling time-warp effectively
forces recognition to take place at the same "time" during each oscillation. Imagine drawing a vertical line in the figure and notice that it cannot pass through all the peaks of output
unit activities.)
We have described the beginning of a research project to use action potentials and timing to
solve a real speech problem in a "neural" fashion. Very unsophisticated features were used,
and no competitive learning was employed in setting the connection weights. Even so, the
system appears to function in a word-spotting mode, and displays a facility of matching
patterns with time warp. Its intrinsic design makes it insensitive to burst noise and to
frequency-band noise.
How is computation being done? After features are detected, rates of change are slow, and
little additional information is accumulated during say a 50 ms. interval. If we let "time
be its own representation", as Carver Mead used to say, we let the information source be
the effective clock, and the effective clock rate is only about 20 Hz. Instead, by adding a
rhythm, we can interleave many calculations (in this particular case about the possibility of
different time warps) while the basic inputs are changing very little. Using an oscillation
frequency of 50 Hz and a resolving time of I ms in the speech example we describe increases the effective clock rate by more than a factor of 10 compared to the effective clock
rate of the UHT computation.
We believe that "time as its own representation" is a loser for processing information when
the computation desired is complex but the data is slowly changing. No computer scientist
would use a computer with a 24 Hz clock to analyze a movie because the movie is viewed
at 24 frames a second. Biology will surely have found its way out of this "paced by the
environment" dilemma. Finally, because problems are easy or hard according to how algorithms fit on hardware and according to the representation of information, the differences in
operation between the system we have described and conventional ANN suggest the utility
of thinking about other problems in a timing representation.
Acknowledgements
The authors thank Sanjoy Mahajan and Erik Winfree for comments and help with preparation of the manuscript. This work was supported in part by the Center for Neuromorphic
Systems Engineering as a part of the National Science Foundation Engineering Research
Center Program under grant EEC-9402726. Roweis is supported by the Natural Sciences
and Engineering Research Council of Canada under an NSERC 1967 Award.
References
[Hopfield, 1995] Hopfield, J. (1995). Pattern recognition computation using action potential timing
for stimulus representation. Nature, 376:3J---36.
[Hopfield, 1996] Hopfield, J. (1996). Transforming neural computations and representing time. Proceedings o/the National Academy o/Sciences, 93:15440-15444.
[Hopfield and Tank, 1987] Hopfield, J. and Tank, D. (1987). Neural computation by concentrating
information in time. Proceedings 0/ the National Academy 0/ Sciences, 84: 1896-1900.
[Unnikrishnan et aI., 1991] Unnikrishnan, K., Hopfield, J., and Tank, D. (1991). Connected digit
speaker-dependent speech recognition using a neural network with time-delayed connections.
IEEE Transactions on Signal ProceSSing, 39:698-713.
[Unnikrishnan et aI., 1992] Unnikrishnan, K., Hopfield, J., and Tank, D. (1992).
Speakerindependent digit recognition using a neural network with time-delayed connections. Neural
Computation, 4:108-119.
| 1368 |@word eliminating:1 interleave:1 seems:1 pulse:6 excited:1 mammal:1 initial:1 contains:1 interestingly:1 odour:1 current:1 nt:1 yet:1 readily:1 john:1 periodically:1 speakerindependent:1 shape:2 beginning:1 short:6 along:1 burst:1 ik:1 pathway:1 olfactory:7 indeed:1 behavior:1 roughly:2 brain:1 integrator:1 inspired:1 actual:2 little:2 increasing:1 vertebrate:1 becomes:1 project:1 underlying:5 matched:1 circuit:12 panel:1 what:1 interpreted:1 spoken:3 nj:1 temporal:2 remember:1 ro:1 wrong:1 scaled:1 whatever:1 unit:17 internally:1 grant:1 appear:1 producing:1 carrier:1 retention:1 engineering:8 timing:18 local:1 scientist:1 limit:1 receptor:2 encoding:3 analyzing:1 mead:1 firing:3 might:1 chirp:1 range:1 averaged:1 unsophisticated:1 bat:4 block:2 digit:5 area:1 adapting:1 significantly:1 convenient:1 projection:1 word:15 matching:1 pre:1 wait:1 suggest:1 onto:1 cannot:2 equivalent:3 map:1 conventional:4 missing:2 center:2 resembling:1 duration:3 independently:1 resolution:1 examines:1 imax:1 dominate:1 population:1 variation:1 target:8 hierarchy:2 suppose:1 imagine:1 wetware:1 us:1 distinguishing:1 tapped:1 crossing:2 recognition:17 bottom:1 solved:2 nbi:1 region:3 connected:7 cycle:6 decrease:1 removed:1 substantial:1 environment:1 broken:1 insertion:1 transforming:1 ideally:1 trained:1 solving:1 dilemma:1 basis:3 necessitates:1 hopfield:16 represented:5 various:1 train:1 fast:1 describe:4 effective:5 detected:1 whose:3 encoded:2 solve:4 constitutive:1 ramp:4 drawing:1 say:2 encoder:1 timescale:1 final:1 advantage:1 indication:1 sequence:1 net:2 product:3 reset:1 adaptation:2 maximal:1 relevant:2 turned:1 loser:1 roweis:5 academy:2 participating:1 normalize:1 requirement:1 produce:1 generating:2 help:1 propagating:1 exemplar:1 ij:1 tolerates:1 disabling:1 strong:2 dividing:1 epsps:2 correct:1 filter:1 feeding:1 preliminary:1 investigation:1 biological:1 elementary:4 stretch:1 around:1 ays:2 great:1 driving:1 major:1 recognizer:1 council:1 clearly:2 sensor:1 activates:1 rather:1 conjunction:1 encode:1 unnikrishnan:6 properly:1 moustache:2 greatly:1 contrast:2 motif:1 rigid:2 dependent:1 accumulated:1 inaccurate:1 bt:1 her:2 transformed:1 tank:5 overall:2 sketched:1 upward:1 logn:1 animal:2 equal:1 field:1 having:3 shaped:1 biology:6 represents:2 look:1 thinking:1 stimulus:5 hint:1 simultaneously:1 recognize:8 national:3 individual:1 delayed:2 phase:4 olfaction:3 investigate:1 circular:1 possibility:2 weakness:2 mixture:1 necessary:3 carver:1 loosely:1 euclidean:2 logarithm:4 desired:1 instance:1 modeling:1 neuromorphic:1 subset:1 uniform:6 recognizing:6 delay:15 answer:1 stimulates:1 eec:1 combined:1 fundamental:2 peak:1 discriminating:1 ie:1 off:1 decoding:1 central:1 slowly:2 creating:1 style:1 potential:43 converted:1 de:2 star:1 sec:1 coordinated:1 caused:1 piece:1 multiplicative:1 view:2 apparently:1 analyze:1 sporadically:1 start:3 relied:1 carlos:1 capability:1 competitive:1 minimize:1 ni:1 accuracy:1 variance:1 efficiently:2 identify:1 yes:1 weak:1 identification:2 rejecting:1 produced:3 composes:1 detector:4 oscillatory:2 synaptic:1 against:1 grossly:2 frequency:15 propagated:1 stop:2 sampled:1 adjusting:1 concentrating:1 actually:1 nerve:2 appears:1 manuscript:1 higher:2 response:2 done:1 strongly:2 stage:1 implicit:1 clock:6 propagation:2 lack:1 mode:1 quality:3 indicated:2 believe:1 effect:1 normalized:3 concept:1 facility:1 mahajan:1 illustrated:4 during:4 speaker:2 rhythm:6 excitation:1 steady:1 timedelay:1 m:3 criterion:1 meaning:1 common:2 physical:1 insensitive:1 analog:13 occurred:1 refer:1 composition:1 ai:4 similarly:1 dot:2 cortex:1 surface:1 base:1 aligning:1 dominant:1 own:2 recent:1 thresh:2 driven:2 loga:1 dimensionally:1 initiation:1 rep:1 watson:1 seen:1 additional:2 spectrogram:3 employed:1 converting:1 recognized:7 converge:1 surely:1 redundant:1 monotonically:1 signal:5 ii:1 sliding:1 resolving:1 sound:1 match:13 faster:2 determination:1 calculation:2 long:1 lin:1 sphere:1 characterized:1 concerning:1 post:1 molecular:5 award:1 regenerative:1 coded:1 biophysics:1 basic:1 vision:1 circumstance:1 winfree:1 cochlea:1 represent:7 normalization:3 disaster:1 cell:14 separately:1 spacing:1 interval:2 diagram:4 source:2 ineffective:1 comment:1 hz:4 call:1 noting:1 ideal:1 presence:1 iii:2 easy:1 uht:9 fit:2 architecture:1 identified:3 idea:2 prototype:1 shift:1 inactive:1 whether:1 utility:1 passed:1 speech:11 action:41 amount:1 band:6 hardware:5 processed:1 simplest:1 generate:1 shifted:1 notice:1 per:3 broadly:1 four:4 salient:1 threshold:10 falling:1 changing:2 ultrasonic:1 ani:1 thresholded:1 lowering:2 merely:3 monotone:1 sum:1 communicate:2 arrive:3 almost:1 reasonable:1 place:1 neurobiologist:1 oscillation:6 comparable:1 pushed:1 entirely:2 brody:4 bound:1 paced:1 syllable:1 display:1 activity:4 strength:3 deficiency:1 scene:1 encodes:1 generates:2 speed:1 generalist:1 imin:1 according:2 poor:1 describes:4 across:3 sam:1 unity:1 biologically:1 invariant:1 gradually:1 taken:1 resource:1 describing:1 nose:1 end:3 sending:1 confounded:1 available:1 operation:2 multiplied:1 appropriate:1 occurrence:1 responsive:3 odor:16 jnj:1 top:1 giving:1 question:1 quantity:1 spike:7 occurs:1 receptive:1 concentration:2 traditional:1 fiducial:3 responds:1 distance:2 separate:1 thank:1 cochlear:1 trivial:1 erik:1 code:2 length:2 index:1 retained:1 relationship:1 ratio:1 unfortunately:1 rise:1 implementation:1 design:1 unknown:1 contend:1 vertical:1 neuron:8 ilog:1 neurobiology:1 frame:1 discovered:1 smoothed:1 sharp:2 intensity:5 canada:1 cast:1 required:1 connection:5 california:1 spotting:1 below:7 pattern:20 grandmother:4 program:1 reliable:1 mouth:1 ia:1 event:3 critical:1 natural:3 difficulty:3 power:2 force:1 representing:2 scheme:2 movie:2 technology:1 eye:1 picture:1 utterance:4 acknowledgement:1 relative:4 beside:2 analogy:1 generator:3 foundation:1 nucleus:1 degree:1 propagates:1 principle:1 thresholding:1 bank:1 playing:1 changed:1 supported:2 last:1 arriving:1 warp:17 institute:1 template:4 taking:1 leaky:1 overcome:1 dimension:1 vocabulary:1 boundary:1 superficially:1 sensory:3 stuck:1 author:1 transaction:1 active:2 sequentially:1 conceptual:2 spectrum:1 continuous:2 urt:1 sonar:2 why:2 channel:7 nature:1 molecule:2 complex:2 stereotyped:1 arrow:1 whole:1 noise:3 allowed:1 fashion:1 slow:3 axon:3 precision:1 sub:1 position:1 breaking:1 third:2 externally:1 down:1 showing:1 decay:2 dominates:1 intrinsic:2 consist:1 adding:1 effectively:2 importance:4 ci:1 magnitude:1 illustrates:1 downward:1 simply:1 contained:1 nserc:1 corresponds:1 determines:1 chiefly:1 viewed:3 ann:1 considerable:1 change:8 hard:1 typical:1 except:2 determined:1 called:1 specie:4 pas:2 isomorphic:1 total:1 discriminate:1 attempted:2 sanjoy:1 internal:2 preparation:1 dept:1 princeton:2 |
404 | 1,369 | Learning Continuous Attractors in
Recurrent Networks
H. Sebastian Seung
Bell Labs, Lucent Technologies
Murray Hill, NJ 07974
seung~bell-labs.com
Abstract
One approach to invariant object recognition employs a recurrent neural network as an associative memory. In the standard depiction of the
network's state space, memories of objects are stored as attractive fixed
points of the dynamics. I argue for a modification of this picture: if an
object has a continuous family of instantiations, it should be represented
by a continuous attractor. This idea is illustrated with a network that
learns to complete patterns. To perform the task of filling in missing information, the network develops a continuous attractor that models the
manifold from which the patterns are drawn. From a statistical viewpoint, the pattern completion task allows a formulation of unsupervised
learning in terms of regression rather than density estimation .
A classic approach to invariant object recognition is to use a recurrent neural network as an associative memory[l]. In spite of the intuitive appeal and biological
plausibility of this approach, it has largely been abandoned in practical applications.
This paper introduces two new concepts that could help resurrect it: object representation by continuous attractors, and learning attractors by pattern completion.
In most models of associative memory, memories are stored as attractive fixed points
at discrete locations in state space[l]. Discrete attractors may not be appropriate for
patterns with continuous variability, like the images of a three-dimensional object
from different viewpoints. When the instantiations of an object lie on a continuous
pattern manifold, it is more appropriate to represent objects by attractive manifolds
of fixed points, or continuous attractors.
To make this idea practical, it is important to find methods for learning attractors
from examples. A naive method is to train the network to retain examples in shortterm memory. This method is deficient because it does not prevent the network
from storing spurious fixed points that are unrelated to the examples. A superior
method is to train the network to restore examples that have been corrupted, so
that it learns to complete patterns by filling in missing information.
Learning Continuous Attractors in Recurrent Networks
(a)
655
(b)
Figure 1: Representing objects by dynamical attractors. (a) Discrete attractors.
(b) Continuous attractors.
Learning by pattern completion can be understood from both dynamical and statistical perspectives. Since the completion task requires a large basin of attraction
around each memory, spurious fixed points are suppressed. The completion task
also leads to a formulation of unsupervised learning as the regression problem of
estimating functional dependences between variables in the sensory input.
Density estimation, rather than regression, is the dominant formulation of unsupervised learning in stochastic neural networks like the Boltzmann machine[2] . Density
estimation has the virtue of suppressing spurious fixed points automatically, but it
also has the serious drawback of being intractable for many network architectures.
Regression is a more tractable, but nonetheless powerful, alternative to density
estimation.
In a number of recent neurobiological models, continuous attractors have been used
to represent continuous quantities like eye position-[3], direction of reaching[4], head
direction[5], and orientation of a visual stimulus[6]. Along with these models, the
present work is part of a new paradigm for neural computation based on continuous
attractors.
1
DISCRETE VERSUS CONTINUOUS ATTRACTORS
Figure 1 depicts two ways of representing objects as attractors of a recurrent neural
network dynamics. The standard way is to represent each object by an attractive
fixed point[l], as in Figure 1a. Recall of a memory is triggered by a sensory input,
which sets the initial conditions. The network dynamics converges to a fixed point,
thus retrieving a memory. If different instantiations of one object lie in the same
basin of attraction, they all trigger retrieval of the same memory, resulting in the
many-to-one map required for invariant recognition.
In Figure 1b, each object is represented by a continuous manifold of fixed points.
A one-dimensional manifold is shown, but generally the attractor should be multidimensional, and is parametrized by the instantiation or pose parameters of the
object . For example, in visual object recognition, the coordinates would include the
viewpoint from which the object is seen.
The reader should be cautioned that the term "continuous attractor" is an idealization and should not be taken too literally. In real networks, a continuous attractor
is only approximated by a manifold in state space along which drift is very slow.
This is illustrated by a simple example, a descent dynamics on a trough-shaped
energy landscape[3]. If the bottom of the trough is perfectly level, it is a line of
fixed points and an ideal continuous attract or of the dynamics. However, any slight
imperfections cause slow drift along the line. This sort of approximate continuous
attract or is what is found in real networks, including those trained by the learning
656
H S. Seung
(a)
hidden layer
(b)
~
visible layer
Figure 2: (a) Recurrent network. (b) Feedforward autoencoder.
algorithms to be discussed below.
2
DYNAMICS OF MEMORY RETRIEVAL
The preceding discussion has motivated the idea of representing pattern manifolds
by continuous attractors. This idea will be further developed with the simple network shown in Figure 2a, which consists of a visible layer Xl E Rnl and a hidden
layer X2 E Rn2. The architecture is recurrent, containing both bottom-up connections (the n2 x nl matrix W2d and top-down connections (the nl x n2 matrix
WI2). The vectors bl and b2 represent the biases ofthe neurons. The neurons have
a rectification nonlinearity [x]+ = max{x, O}, which acts on vectors component by
component.
There are many variants of recurrent network dynamics: a convenient choice is the
following discrete-time version, in which updates of the hidden and visible layers
alternate in time. After the visible layer is initialized with the input vector Xl (0),
the dynamics evolves as
X2(t)
Xl (t)
=
=
[b 2 + W2IXI(t -1)]+ ,
[b l + W12X2(t)]+ .
(1)
If memories are stored as attractors, iteration of this dynamics can be regarded as
memory retrieval.
Activity circulates around the feedback loop between the two layers. One iteration
of this loop is the map Xl(t - 1) ~ X2(t) ~ Xl(t). This single iteration is equivalent to the feedforward architecture of Figure 2b. In the case where the hidden
layer is smaller than the visible layers, this architecture is known as an auto encoder network[7]. Therefore the recurrent network dynamics (1) is equivalent to
repeated iterations of the feedforward autoencoder. This is just the standard trick
of unfolding the dynamics of a recurrent network in time, to yield an equivalent
feedforward network with many layers[7]. Because of the close relationship between
the recurrent network of Figure 2a and the autoencoder of Figure 2b, it should not
be surprising that learning algorithms for these two networks are also related, as
will be explained below.
3
LEARNING TO RETAIN PATTERNS
Little trace of an arbitrary input vector Xl (0) remains after a few time steps of the
dynamics (1). However, the network can retain some input vectors in short-term
memory as "reverberating" patterns of activity. These correspond to fixed points of
the dynamics (1); they are patterns that do not change as activity circulates around
the feedback loop.
Learning Continuous Attraclors in Recurrent Networlcs
657
This suggests a formulation of learning as the optimization of the network's ability to
retain examples in short-term memory. Then a suitable cost function is the squared
difference IXI (T) - Xl (0)12 between the example pattern Xl (0) and the network's
short-term memory Xl (T) of it after T time steps. Gradient descent on this cost
function can be done via backpropagation through time[7].
If the network is trained with patterns drawn from a continuous family, then it can
learn to perform the short-term memory task oy developing a continuous attractor
that lies near the examples it is trained on. When the hidden layer is smaller than
the visible layer, the dimensionality of the attractor is limited by the size of the
hidden layer.
For the case of a single time step (T = 1), training the recurrent network of Figure
2a to retain patterns is equivalent to training the autoencoder of Figure 2b by
minimizing the squared difference between its input and output layers, averaged
over the examples[8]. From the information theoretic perspective, the small hidden
layer in Figure 2b acts as a bottleneck between the input and output layers, forcing
the autoencoder to learn an efficient encoding of the input.
For the special case of a linear network, the nature of the learned encoding is
understood completely. Then the input and output vectors are related by a simple
matrix multiplication. The rank of the matrix is equal to the number of hidden
units. The average distortion is minimized when this matrix becomes a projection
operator onto the subspace spanned by the principal components of the examples[9].
From the dynamical perspective, the principal subspace is a continuous attractor
of the dynamics (1). The linear network dynamics converges to this attractor in
a single iteration, starting from any initial condition. Therefore we can interpret
principal component analysis and its variants as methods of learning continuous
attractors[lO].
4
LEARNING TO COMPLETE PATTERNS
Learning to retain patterns in short-term memory only works properly for architectures with a small hidden layer. The problem with a large hidden layer is evident
when the hidden and visible layers are the same size, and the neurons are linear.
Then the cost function for learning can be minimized by setting the weight matrices
equal to the identity, W 21 = W l2 = I. For this trivial minimum, every input vector
is a fixed point of the recurrent network (Figure 2a), and the equivalent feedforward
network (Figure 2b) exactly realizes the identity map. Clearly these networks have
not learned anything.
Therefore in the case of a large hidden layer, learning to retain patterns is inadequate. Without the bottleneck in the architecture, there is no pressure on the
feedforward network to learn an efficient encoding. Without constraints on the dimension of the attractor, the recurrent network develops spurious fixed points that
have nothing to do with the examples.
These problems can be solved by a different formulation of learning based on the
task of pattern completion. In the completion task of Figure 3a, the network is
initialized with a corrupted version of an example. Learning is done by minimizing
the completion error, which is the squared difference IXI (T) - dl 2 between the uncorrupted pattern d and the final visible vector Xl (T). Gradient descent on completion
error can be done with backpropagation through time[ll].
This new formulation of learning eliminates the trivial identity map solution men-
H. S. Seung
658
(a)
~1
L
_
retention.
~1
..
_
1
~ completio~ ~
It
~
It
___
(b)
topographic feature map
9x9 patch
missing
sensory
Input
retrieved
memory
Figure 3: (a) Pattern retention versus completion. (b) Dynamics of pattern completion.
(b)
5x5 receptive fields
Figure 4: (a) Locally connected architecture. (b) Receptive fields of hidden neurons.
tioned above: while the identity network can retain any example, it cannot restore
corrupted examples to their pristine form. The completion task forces the network
to enlarge the basins of attraction of the stored memories, which suppresses spurious fixed points. It also forces the network to learn associations between variables
in the sensory input.
5
LOCALLY CONNECTED ARCHITECTURE
Experiments were conducted with images of handwritten digits from the USPS
database described in [12]. The example images were 16 x 16, with a gray scale
ranging from a to 1. The network was trained on a specific digit class, with the
goal of learning a single pattern manifold. Both the network architecture and the
nature of the completion task were chosen to suit the topographic structure present
in visual images.
The network architecture was given a topographic organization by constraining the
synaptic connectivity to be local, as shown in Figure 4a. Both the visible and hidden
layers of the network were 16 x 16. The visible layer represented an image, while
the hidden layer was a topographic feature map. Each neuron had 5 x 5 receptive
and projective fields, except for neurons near the edges, which had more restricted
connectivity.
In the pattern completion task, example images were corrupted by zeroing the
pixels inside a 9 x 9 patch chosen at a random location, as shown in Figure 3a.
The location of the patch was randomized for each presentation of an example.
The size of the patch was a substantial fraction of the 16 x 16 image, and much
larger than the 5 x 5 receptive field size. This method of corrupting the examples
gave the completion task a topographic nature, because it involved a set of spatially
contiguous pixels. This topographic nature would have been lacking if the examples
had been corrupted by, for example, the addition of spatially uncorrelated noise.
Figure 3b illustrates the dynamics of pattern completion performed by a network
Learning Continuous Attractors in Recurrent Networks
659
trained on examples of the digit class "two." The network is initialized with a
corrupted example of a "two." After the first itex:ation of the dynamics, the image
is partially restored. The second iteration leads to superior restoration, with further
sharpening of the image. The "filling in" phenomenon is also evident in the hidden
layer.
The network was first trained on a retrieval dynamics of one iteration. The resulting
biases and synaptic weights were then used as initial conditions for training on a
retrieval dynamics of two iterations. The hidden layer developed into a topographic
feature map suitable for representing images of the digit "two." Figure 4b depicts
the bottom-up receptive fields of the 256 hidden neurons. The top-down projective
fields of these neurons were similar, but are not shown.
This feature map is distinct from others[13) because of its use of top-down and
bottom-up connections in a feedback loop. The bottom-up connections analyze
images into their constituent features, while the top-down connections synthesize
images by composing features. The features in the top-down connections can be
regarded as a "vocabulary" for synthesis of images. Since not all combinations of
features are proper patterns, there must be some "grammatical" constraints on their
combination. The network's ability to complete patterns suggests that some of these
constraints are embedded in the dynamical equations of the network. Therefore the
relaxation dynamics (1) can be regarded as a process of massively parallel constraint
satisfaction.
6
CONCLUSION
I have argued that continuous attractors are a natural representation for pattern
manifolds. One method of learning attractors is to train the network to retain
examples in short-term memory. This method is equivalent to autoencoder learning,
and does not work if the number of hidden units is large. A better method is to train
the network to complete patterns. For a locally connected network, this method
was demonstrated to learn a topographic feature map. The trained network is able
to complete patterns, indicating that syntactic constraints on the combination of
features are embedded in the network dynamics.
Empirical evidence that the network has indeed learned a continuous attractor is
obtained by local linearization of the network (1). The linearized dynamics has
many eigenvalues close to unity, indicating the existence of an approximate continuous attractor. Learning with an increased number of iterations in the retrieval
dynamics should improve the quality of the approximation.
There is only one aspect of the learning algorithm that is specifically tailored for
continuous attractors. This aspect is the limitation of the retrieval dynamics (1)
to a few iterations, rather than iterating it all the way to a true fixed point. As
mentioned earlier, a continuous attractor is only an idealization; in a real network
it does not consist of true fixed points, but is just a manifold to which relaxation is
fast and along which drift is slow. Adjusting the shape of this manifold is the goal
of learning; the exact locations of the true fixed points are not relevant.
The use of a fast retrieval dynamics removes one long-standing objection to attractor
neural networks, which is that true convergence to a fixed point takes too long. If all
that is desired is fast relaxation to an approximate continuous attractor, attractor
neural networks are not much slower than feedforward networks.
In the experiments discussed here, learning was done with backpropagation through
time. Contrastive Hebbian learning[14] is a simpler alternative. Part of the image
660
H S. Seung
is held clamped, the missing values are filled in by convergence to a fixed point,
and an anti-Hebbian update is made. Then the missing values are clamped at their
correct values, the network converges to a new fixed point, and a Hebbian update
is made. This procedure has the disadvantage of requiring true convergence to a
fixed point, which can take many iterations. It also requires symmetric connections,
which may be a representational handicap.
This paper addressed only the learning of a single attractor to represent a single
pattern manifold. The problem of learning multiple attractors to represent mUltiple
pattern classes will be discussed elsewhere, along with the extension to network
architectures with many layers.
Acknowledgments This work was supported by Bell Laboratories. I thank J. J.
Hopfield, D. D. Lee, L. K. Saul, N. D. Socci, H. Sompolinsky, and D. W. Tank for
helpful discussions.
References
[1] J. J. Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proc. Nat. Acad. Sci. USA, 79:2554-2558, 1982.
[2] D. H. Ackley, G. E. Hinton, and T. J. Sejnowski. A learning algorithm for Boltzmann
machines. Cognitive Science, 9:147-169, 1985.
[3] H. S. Seung. How the brain keeps the eyes still. Proc . Natl. Acad. Sci. USA,93:1333913344, 1996.
[4] A. P. Georgopoulos, M. Taira, and A. Lukashin. Cognitive neurophysiology of the
motor cortex. Science, 260:47-52, 1993.
[5] K Zhang. Representation of spatial orientation by the intrinsic dynamics of the
head-direction cell ensemble: a theory. J. Neurosci., 16:2112-2126, 1996.
[6] R . Ben-Yishai, R. L. Bar-Or, and H. Sompolinsky. Theory of orientation tuning in
visual cortex. Proc. Nat. Acad. Sci. USA, 92:3844-3848, 1995 .
[7] D.E. Rumelhart, G.E. Hinton, and R.J . Williams. Learning internal representations
by error propagation. In D.E. Rumelhart and J.L. McClelland, editors, Parallel Distributed Processing, volume 1, chapter 8, pages 318-362. MIT Press, Cambridge, 1986.
[8] G. W . Cottrell, P. Munro, and D. Zipser. Image compression by back propagation: an
example of extensional programming. In N. E. Sharkey, editor, Models of cognition:
a review of cognitive science. Ablex, Norwood, NJ, 1989.
[9] P. Baldi and K Hornik. Neural networks and principal component analysis: Learning
from examples without local minima. Neural Networks, 2:53-58, 1989.
[10] H. S. Seung. Pattern analysis and synthesis in attractor neural networks. In K-Y. M.
Wong, 1. King, and D.-y' Yeung, editors, Theoretical Aspects of Neural Computation :
A Multidisciplinary Perspective, Singapore, 1997. Springer-Verlag.
[11] F.-S. Tsung and G. W . Cottrell. Phase-space learning. Adv. Neural Info . Proc. Syst.,
7:481-488, 1995.
[12] Y. LeCun et al. Learning algorithms for classification: a comparison on handwritten
digit recognition. In J.-H. Oh, C. Kwon, and S. Cho, editors, Neural networks: the
statistical mechanics perspective, pages 261-276, Singapore, 1995. World Scientific.
[13] T . Kohonen. The self-organizing map. Proc. IEEE, 78:1464-1480, 1990.
[14] J. J . Hopfield, D. I. Feinstein, and R. G. Palmer. "Unlearning" has a stabilizing effect
in collective memories. Nature, 304:158-159, 1983.
| 1369 |@word neurophysiology:1 version:2 compression:1 linearized:1 contrastive:1 pressure:1 initial:3 suppressing:1 com:1 surprising:1 must:1 cottrell:2 visible:10 extensional:1 shape:1 motor:1 remove:1 update:3 short:6 location:4 simpler:1 zhang:1 along:5 rnl:1 retrieving:1 consists:1 unlearning:1 inside:1 baldi:1 indeed:1 mechanic:1 brain:1 automatically:1 little:1 becomes:1 estimating:1 unrelated:1 what:1 suppresses:1 developed:2 sharpening:1 nj:2 every:1 multidimensional:1 act:2 exactly:1 unit:2 retention:2 understood:2 local:3 acad:3 encoding:3 suggests:2 limited:1 projective:2 palmer:1 averaged:1 practical:2 acknowledgment:1 lecun:1 backpropagation:3 digit:5 procedure:1 empirical:1 bell:3 convenient:1 projection:1 spite:1 onto:1 close:2 cannot:1 operator:1 wong:1 equivalent:6 map:10 demonstrated:1 missing:5 williams:1 starting:1 stabilizing:1 attraction:3 regarded:3 spanned:1 oh:1 classic:1 coordinate:1 trigger:1 exact:1 programming:1 ixi:2 trick:1 synthesize:1 rumelhart:2 recognition:5 approximated:1 database:1 bottom:5 ackley:1 solved:1 connected:3 adv:1 sompolinsky:2 substantial:1 mentioned:1 seung:7 dynamic:27 trained:7 ablex:1 completely:1 usps:1 hopfield:3 emergent:1 represented:3 chapter:1 train:4 distinct:1 fast:3 sejnowski:1 larger:1 distortion:1 encoder:1 ability:3 topographic:8 syntactic:1 final:1 associative:3 triggered:1 eigenvalue:1 kohonen:1 relevant:1 loop:4 organizing:1 representational:1 intuitive:1 constituent:1 convergence:3 converges:3 ben:1 object:16 help:1 recurrent:16 completion:16 pose:1 cautioned:1 direction:3 drawback:1 correct:1 stochastic:1 argued:1 biological:1 extension:1 around:3 cognition:1 estimation:4 proc:5 realizes:1 unfolding:1 mit:1 clearly:1 imperfection:1 rather:3 reaching:1 properly:1 rank:1 helpful:1 attract:2 spurious:5 hidden:19 pixel:2 tank:1 classification:1 orientation:3 spatial:1 special:1 equal:2 field:6 shaped:1 enlarge:1 unsupervised:3 filling:3 minimized:2 others:1 stimulus:1 develops:2 serious:1 employ:1 few:2 kwon:1 taira:1 phase:1 attractor:40 suit:1 organization:1 introduces:1 nl:2 natl:1 yishai:1 held:1 edge:1 literally:1 filled:1 initialized:3 desired:1 theoretical:1 increased:1 earlier:1 contiguous:1 disadvantage:1 restoration:1 cost:3 inadequate:1 conducted:1 too:2 stored:4 corrupted:6 cho:1 density:4 randomized:1 retain:9 standing:1 lee:1 synthesis:2 connectivity:2 squared:3 x9:1 containing:1 cognitive:3 syst:1 rn2:1 b2:1 trough:2 tsung:1 performed:1 lab:2 analyze:1 sort:1 parallel:2 largely:1 ensemble:1 yield:1 ofthe:1 landscape:1 correspond:1 handwritten:2 sebastian:1 synaptic:2 nonetheless:1 energy:1 involved:1 sharkey:1 adjusting:1 recall:1 dimensionality:1 back:1 formulation:6 done:4 just:2 propagation:2 multidisciplinary:1 quality:1 gray:1 scientific:1 usa:3 effect:1 concept:1 true:5 requiring:1 spatially:2 symmetric:1 laboratory:1 illustrated:2 attractive:4 ll:1 x5:1 self:1 anything:1 hill:1 evident:2 complete:6 theoretic:1 image:15 ranging:1 superior:2 functional:1 physical:1 volume:1 discussed:3 slight:1 association:1 interpret:1 cambridge:1 tuning:1 zeroing:1 nonlinearity:1 had:3 cortex:2 depiction:1 dominant:1 recent:1 perspective:5 retrieved:1 forcing:1 massively:1 verlag:1 uncorrupted:1 seen:1 minimum:2 preceding:1 paradigm:1 multiple:2 hebbian:3 plausibility:1 long:2 retrieval:8 variant:2 regression:4 yeung:1 iteration:11 represent:6 tailored:1 cell:1 addition:1 objection:1 addressed:1 eliminates:1 deficient:1 zipser:1 near:2 ideal:1 feedforward:7 constraining:1 gave:1 architecture:11 perfectly:1 idea:4 bottleneck:2 motivated:1 munro:1 cause:1 generally:1 iterating:1 locally:3 mcclelland:1 singapore:2 discrete:5 drawn:2 prevent:1 relaxation:3 fraction:1 idealization:2 powerful:1 family:2 reader:1 patch:4 layer:26 handicap:1 activity:3 constraint:5 georgopoulos:1 x2:3 networlcs:1 aspect:3 developing:1 alternate:1 combination:3 smaller:2 suppressed:1 unity:1 evolves:1 modification:1 explained:1 invariant:3 restricted:1 taken:1 rectification:1 equation:1 remains:1 lukashin:1 feinstein:1 tractable:1 appropriate:2 alternative:2 slower:1 existence:1 abandoned:1 top:5 include:1 murray:1 bl:1 quantity:1 restored:1 receptive:5 dependence:1 gradient:2 subspace:2 thank:1 sci:3 parametrized:1 manifold:12 argue:1 trivial:2 relationship:1 minimizing:2 info:1 trace:1 proper:1 boltzmann:2 collective:2 perform:2 neuron:8 descent:3 anti:1 hinton:2 variability:1 head:2 arbitrary:1 drift:3 required:1 connection:7 learned:3 able:1 bar:1 dynamical:4 pattern:33 below:2 wi2:1 including:1 memory:22 max:1 suitable:2 ation:1 satisfaction:1 force:2 restore:2 natural:1 representing:4 improve:1 technology:1 eye:2 picture:1 shortterm:1 naive:1 autoencoder:6 auto:1 review:1 l2:1 multiplication:1 lacking:1 embedded:2 oy:1 men:1 limitation:1 versus:2 norwood:1 basin:3 viewpoint:3 corrupting:1 editor:4 storing:1 uncorrelated:1 lo:1 elsewhere:1 supported:1 bias:2 saul:1 distributed:1 grammatical:1 feedback:3 dimension:1 vocabulary:1 world:1 sensory:4 made:2 approximate:3 neurobiological:1 keep:1 instantiation:4 continuous:32 socci:1 learn:5 nature:5 composing:1 hornik:1 neurosci:1 noise:1 n2:2 nothing:1 repeated:1 depicts:2 slow:3 position:1 xl:10 lie:3 clamped:2 learns:2 down:5 lucent:1 specific:1 reverberating:1 appeal:1 virtue:1 evidence:1 dl:1 intractable:1 consist:1 intrinsic:1 linearization:1 nat:2 illustrates:1 visual:4 partially:1 springer:1 identity:4 goal:2 presentation:1 tioned:1 king:1 change:1 specifically:1 except:1 principal:4 indicating:2 internal:1 phenomenon:1 |
405 | 137 | 720
AN ELECTRONIC PHOTORECEPTOR
SENSITIVE TO SMALL CHANGES IN INTENSITY
T. Delbriick and C. A. Mead
256-80 Computer Science
California Institute of Technology
Pasadena, CA 91125
ABSTRACT
We describe an electronic photoreceptor circuit that is sensitive to
small changes in incident light intensity. The sensitivity to change8
in the intensity is achieved by feeding back to the input a filtered
version of the output. The feedback loop includes a hysteretic element. The circuit behaves in a manner reminiscent of the gain
control properties and temporal responses of a variety of retinal
cells, particularly retinal bipolar cells. We compare the thresholds
for detection of intensity increments by a human and by the circuit. Both obey Weber's law and for both the temporal contrast
sensitivities are nearly identical.
We previously described an electronic photoreceptor that outputs a voltage that is
logarithmic in the light intensity (Mead, 1985). This report describes an extension
of this circuit which was based on a suggestion by Frank Werblin that biological
retinas may achieve greater sensitivity to change8 in the illumination by feeding
back a filtered version of the output.
OPERATION OF THE CIRCUIT
The circuit (Figure 1) consists of a phototransistor (P), exponential feedback to
P (Ql, Q2, and Q3), a transconductance amplifier (A), and the hysteretic element
(Q4 and Qs). In general terms the operation of the circuit consists of two stages of
amplification with hysteresis in the feedback loop. The light falls on the parasitic
bipolar transistor P. (The rest of the circuit is shielded by metal.) P's collector
is the substrate and the base is an isolated well. P and Ql form the first stage of
amplification. The light produces a base current Is for P. The emitter current IE
is PIs, neglecting collector resistance for now. P is typically a few hundred. The
feedback current IQl is set by the gate voltage on QdQ2' which is set by the current
through Q3, which is set by the feedback voltage Vjb. In equilibrium Vjb will be
such that IQl = IE and some voltage Vp will be the output of the first stage. The
An Electronic Photoreceptor Sensitive to Small Changes
negative feedback through the transconductance amplifier A will make Vp ~ V/ b ?
This voltage is logarithmic in the light intensity, since in subthreshold operation
the currents through Q2 and Q3 are exponential in their gate to source voltages.
The DC output of the circuit will be Vout ~ V/b = Vdd - (2kT /q) log IE, neglecting
the back-gate effect for Q2. Figure Sa (DC output) shows that the assumption of
subthreshold operation is valid over about 4 orders of magnitude.
G
Vout
T
O
Figure 1. The photoreceptor circuit.
Now what happens when the intensity increases a bit? Figure 2a shows the current
through P and Ql as a function of the voltage Vp. Both P and Ql act like current
sources in parallel with a resistance, where the value of the current is set, respectively, by the light intensity and by the feedback voltage V/b. When the intensity
increases a bit the immediate result is that the curve labeled IE in Figure 2a will
shift upwards a little to the curve labeled I~. But IQl won't change right away it is set by the delayed feedback. The effect on Vp will be that Vp will drop by the
amount of the shift in the intersection of the curves in Figure 2a, to Vj,. Because
interesting gain control properties arise here we will analyze this before going on
with the rest of the circuit.
In Figure 2b, we model P and Ql as current sources with associated drain/collector
resistances. Now,
721
722
Delbro.ck and Mead
rp and rQl physically arise from the familiar Early effect, a variation of depletion
region thickness causing a variation in the channel length or base thickness. It is
a reasonable approximation to model the drain or collector resistance due to the
Early effect as r = Vel I, where Ve is the Early voltage and is typically tens of volts,
and I is the value of the current source.
I
~1
rQl
~1
12:
Vp
rp
12:
~
Vp Vp
a
V
1
...
Figure 2. The first stage of amplification. a: The curves show the current through
the phototransistor P and feedback transistor Ql as a function of the voltage Vp .
Since Ip = IQl the intersection gives the voltage Vp. b: An equivalent circuit model
for these transistors in the linear region. The Early effect leads to drain/collector
resistances inversely proportional to the value of the current source.
Substituting this approximation for rp and rQl into the above expression for 6Vp
and letting 6IE = P6Is, we obtain
6V
= 6Is
Is (Ve,p IIVe ,Ql)
where Ve,p and Ve,Ql are the Early voltages associated with the phototransistor and
the feedback transistor QlI respectively. In other words, the change in Vp is just
proportional to the "contrast" 6 Is / Is. Figure 4a shows test results which support
this model.
A detector which encodes the intensity logarithmically (so that the output V in
response to an input I is V = log I) would also give 6V = 61/1. Our in our circuit
the gain control properties for transients arise from an unrelated property of the
conductances of the sensor and the feedback element. Comparing the gains for DC
and for transients in our circuit and using the expression for the DC output given
earlier, we find that the ratio of the gains is
transient gain ~ Ve,p IIVe ,Ql ~ 200
2kT/q
DC gain
723
An Electronic Photoreceptor Sensitive to Small Changes
assuming Ve ,pllVe ,Ql
= lOV and kT/q = 25mV.
Finally, let us consider the operation of the rest of the circuit. The second stage of
amplification is done by the transconductance amplifier A. A produces a current
which is pr~ortional to the tanh of the difference between the two inputs, I
G tanh( v;k;7~b). When the output of the amplifier is taken as a voltage, the voltage
gain is typically a few hundred. Following the transconductance amplifier there is
a pair of diode connected transistors, Q4 and Q5 (Figure 3a), which we call the
hysteretic element. This pair of transistors has an I-V characteristic which is similar
to that of Figure 3. The hysteretic element conducts very little until the voltage
across it becomes substantial. Thus the transconductance amplifier works in a dual
voltage-output/current-output mode. Small changes in the output voltage result in
little change in the feedback voltage. Larger changes in the output voltage cause
current to flow through the hysteretic element, completing the feedback loop. This
represents a form of memory, or hysteresis, for the past state of the output and a
sensitivity to small changes in the input around the past history of the input.
=
I
v
I
a
b
Figure 3. a: The hysteretic element. b: 1- l': characteristic.
COMPARISON OF CmCUIT AND RETINAL CELLS
We felt that since the circuit was motivated by biology it might be interesting to
compare the operational characteristics of the circuit and of retinal cells. Since the
circuit has no spatial extent it cannot model any of the spatially mediated effects
(such as center4Jurround) seen in retinas. Nonetheless we had hoped to capture
some of the temporal effects seen in retinal cells. In Figure 4 we compare the
responses of the circuit and responses of a retinal bipolar cell to diffuse flashes of
light. The circuit has response characteristics closest to those of retinal bipolar
cells.
The circuit's gain control properties are very similar to those of bipolar cells. Figure
5 shows that both the circuit and bipolar cells tend to control their gain so that
they maintain a constant output amplitude for a given change in the log of the
intensity.
724
DelbIilck and Mead
The response characteristics of the circuit differ from those of bipolar cells in the
following ways. First, the gain of the circuit for transients is much larger than that
of the bipolar cell, as can be seen in Figure 5, with concomitantly much smaller
dynamic range. The dynamic range of the bipolar cell is about 1.5 - 2 log units
around the steady intensity, while for the circuit the dynamic range is only about
0.1 log unit.
2.5
Input
BACKGROUND
-1.0 background
IV~
0.2ms
~%
\to% ch~ge
1%
-3.0 background
(2 decades attenuated)
Noise level
20ms
a
Figure 4. The responses of the circuit compared with responses of a retinal bipolar
cell. a: The output of the circuit in response to changes in the intensity. The
background levels refer to the same scale as shown in Figure 5. The bottom curve
shows the noise level; from this one can see that a detection criterion of signal/ noise
ratio equals 2 is satisfied for increments of 1- 2%, in agreement with Figure 6. Note
that a 2 decade attenuation hardly changes the response amplitude but the time
constant increases by a factor of a hundred. b: The response of a bipolar cell (from
Werblin, 1974). The numbers next to the responses are the log of the intensity of
the flash substituted for the intial value of the intensity. Note the bidirectionality
of the response compared to the circuit.
An Electronic Photoreceptor Sensitive to Small Changes
1
~V)
(mV)
'0
~
6
}
4
2
6V~tOr-~-+~~~~~
-2
-5 -4 -3 -2 -1
log (Intensity)
a
-4
log (Intensity)
b
Figure 5. The operating curves of the circuit (a) compared with retinal bipolar
cells (b) (adapted from Werblin, 1974). The curves show the height of the peak of
the response to flashes substituted for the initial intensity. The initial intensity is
given by the intersection of the curves with the abscissa. Note the difference of the
gain and the dynamic range. The squares show the DC responses. The slope of
the DC response for the circuit is less than expected, probably because there is a
leakage current through the hysteretic element.
Second, the response of bipolar cells is symmetrical for increases and decreases in
the intensity. This can probably be traced to the symmetrical responses of the cones
from which they receive direct input. The circuit, on the other hand, only responds
strongly to increases in the light intensity. The response in our circuit only becomes
symmetrical for output voltage swings comparable to leT / q, probably because the
limiting process is recombination in the base of the phototransistor.
Third, the control of time constants is dramatically different. In Figure 4a the top
set of responses is on a time scale 100 times expanded relative to the bottom scale.
The circuit's time constant, in other words, is roughly inversely proportional to the
light intensity. This is not the case for bipolar cells. Although we do not show
it here, the time constant of the responses of bipolar cells hardly varies with light
intensity over at least 4 orders of magnitude (Werblin, 1974).
The circuit's action differs much more from that of photoreceptors, amacrine, or
ganglion cells. Cones show a much larger sustained response relative to their transient response. Amacrine and ganglion cells spike; our circuit does not. And the
circuit differs from on/off amacrine and ganglion cells in the asymmetry of its response to increases and decreases in light intensity.
725
726
Delbruck and Mead
EYE vs. CHIP
We compared the sensitivity to small changes in the light intensity for one of us and
for the circuit in order to get an idea of the performance of the circuit relative to a
subjective scale. The thresholds for detection of intensity increments are shown in
Figure 6.
10
? Human
? Photoreceptor
1.6%
9
8
log(Increment) 7
6
5
?
6
7
8
9
10 11
log( Absorbable Quanta/ sec)
Figure 6. Thresholds for detection of flicker.
The subject (TD) sat in a darkened room foveating a flickering yellow (583nm) LED
at a distance of 75cm. The LED subtended 22' of arc and the frequency of flicker
was 5 Hz. Threshold determination was made by a series of trials, indicated by a
buzzer, in which the computer either caused the LED to flicker or not to flicker.
The percentage of flicker was started at some large amount for which determination
was unambiguous. For each trial the subject pressed a button if he thought he saw
the LED flickering. A incorrect response would cause the percentage of flicker for
the next trial to be increased. A correct response would cause the percentage of
flicker to be decreased with a probability of 0.44. The result after a hundred trials
would be a curve of percentage flicker vs. trial number which started high and then
leveled off around the 75% correct level, taken to be the threshold (Levitt, 1971).
The threshold for the circuit was determined by shining the same LED onto the chip
directly. The threshold was defined as the smallest amount of flicker that would
result in a signal-to-noise ratio of 2 for the output. The two sets of thresholds were
scaled relative to each other using a Tektronics photometer and were both scaled to
read in terms of absorbable quanta, defined for the human as the number of photons
hitting the cornea and for the circuit as the number of quanta that hit the area of
the phototransistor. There is a bias here favoring the circuit, since the other parts
An Electronic Photoreceptor Sensitive to Small Changes
of the circuit have not been included in this area. Including the rest of the circuit
would raise the thresholds for the circuit by a factor of about 3.
The results show that both the circuit and the human approximately obey Weber's
law (61/1 at threshold is a constant), and the sensitivities are nearly the same. The
highest sensitivity for the circuit, measured at an incident intensity of 660#-, W / cm2 ,
was 1.2%. This is about half the intensity in a brightly lit office.
APPLICATIONS
We are trying to develop these sensors for use in neurophysiological optical dye
recording. These experiments require a sensor capable of recording changes in the
incident intensity of about 1 part in 103 - 4 (Grinvald, 1985). The current technique
is to use integrated arrays of photodiodes each with a dedicated rack-mounted low
noise amplifier. We will try to replace this arrangement with arrays of receptors
of the type discussed in this paper. Currently we are 1 to 2 orders of magnitude
short of the required sensitivity, but we hope to improve the performance by using
hybrid bipolar/FET technology.
CONCLUSION
This circuit represents an example of an idea from biology directly and simply
synthesised in silicon. The resulting circuit incorporated not only the intended
idea, sensitivity to changes in illumination, but also gave rise to an unexpected gain
control mechanism unrelated to exponential feedback. The circuit differs in several
ways from its possible biological analogy but remains an interesting and potentially
useful device.
Aeknowledgements
This work was supported by the System Development Foundation and the Office of
Naval Research. Chips were fabricated through the MOSIS foundry.
We thank Frank Werblin for helpful comments and Mary Ann Maher for editorial
assistance.
Refereneea
A. Grinvald. (1985) Real time optical ma~ping of neuronal activity: from single
growth cones to the intact mammalian braili. Ann. Rev. Neurolci. 8:263-305.
H. Levitt. (1971) 'Iransformed up-down methods in psychoacoustics. J. Acoud. Soc.
Am. 49:467-477.
C. Mead. (1985) A Sensitive Electronic Photoreceptor. In 1985 Chapel Hill Conference on VLSI. 463-471.
F. Werblin. (1974) Control of Retinal Sensitivity II: Lateral Interactions at the
Outer Plexiform Layer. J. PhYliology. 68:62-87.
727
| 137 |@word trial:5 version:2 cm2:1 pressed:1 initial:2 series:1 past:2 subjective:1 current:17 comparing:1 reminiscent:1 drop:1 v:2 half:1 device:1 short:1 filtered:2 height:1 direct:1 incorrect:1 consists:2 sustained:1 manner:1 lov:1 expected:1 roughly:1 abscissa:1 td:1 little:3 becomes:2 brightly:1 unrelated:2 circuit:49 what:1 cm:1 q2:3 fabricated:1 temporal:3 act:1 attenuation:1 growth:1 bipolar:16 scaled:2 hit:1 control:8 unit:2 werblin:6 before:1 iql:4 cornea:1 receptor:1 mead:6 approximately:1 might:1 emitter:1 range:4 differs:3 area:2 thought:1 word:2 shielded:1 cannot:1 get:1 onto:1 phototransistor:5 equivalent:1 cmcuit:1 chapel:1 q:1 array:2 variation:2 increment:4 limiting:1 substrate:1 agreement:1 element:8 logarithmically:1 particularly:1 mammalian:1 labeled:2 bottom:2 capture:1 region:2 connected:1 decrease:2 highest:1 substantial:1 dynamic:4 vdd:1 q5:1 raise:1 chip:3 describe:1 larger:3 ip:1 transistor:6 interaction:1 causing:1 loop:3 achieve:1 amplification:4 qli:1 asymmetry:1 produce:2 develop:1 measured:1 sa:1 soc:1 diode:1 differ:1 correct:2 human:4 transient:5 require:1 feeding:2 biological:2 extension:1 around:3 equilibrium:1 substituting:1 tor:1 early:5 smallest:1 tanh:2 currently:1 sensitive:7 saw:1 hope:1 sensor:3 amacrine:3 ck:1 voltage:20 office:2 q3:3 foveating:1 naval:1 contrast:2 am:1 helpful:1 typically:3 integrated:1 pasadena:1 vlsi:1 favoring:1 going:1 dual:1 development:1 spatial:1 equal:1 identical:1 represents:2 biology:2 lit:1 nearly:2 buzzer:1 report:1 few:2 retina:2 ve:6 delayed:1 familiar:1 intended:1 maintain:1 amplifier:7 detection:4 conductance:1 light:12 kt:3 synthesised:1 capable:1 neglecting:2 conduct:1 iv:1 concomitantly:1 isolated:1 increased:1 earlier:1 delbruck:1 intial:1 hundred:4 thickness:2 varies:1 peak:1 sensitivity:10 ie:5 off:2 satisfied:1 nm:1 photon:1 retinal:10 sec:1 includes:1 hysteresis:2 caused:1 mv:2 leveled:1 try:1 analyze:1 parallel:1 slope:1 square:1 characteristic:5 subthreshold:2 yellow:1 vp:12 vout:2 history:1 detector:1 ping:1 nonetheless:1 frequency:1 associated:2 gain:13 amplitude:2 back:3 response:26 vel:1 done:1 strongly:1 just:1 stage:5 subtended:1 until:1 hand:1 rack:1 mode:1 indicated:1 mary:1 effect:7 swing:1 spatially:1 volt:1 read:1 assistance:1 unambiguous:1 steady:1 won:1 fet:1 m:2 criterion:1 trying:1 hill:1 dedicated:1 upwards:1 photometer:1 weber:2 behaves:1 discussed:1 he:2 refer:1 silicon:1 had:1 operating:1 base:4 closest:1 dye:1 seen:3 greater:1 signal:2 ii:1 photodiodes:1 determination:2 editorial:1 physically:1 achieved:1 cell:21 receive:1 background:4 decreased:1 source:5 rest:4 plexiform:1 probably:3 comment:1 subject:2 tend:1 hz:1 recording:2 flow:1 call:1 variety:1 gave:1 idea:3 attenuated:1 shift:2 expression:2 motivated:1 resistance:5 cause:3 hardly:2 action:1 dramatically:1 useful:1 amount:3 ten:1 percentage:4 flicker:9 hysteretic:7 threshold:10 traced:1 button:1 mosis:1 cone:3 reasonable:1 electronic:8 comparable:1 bit:2 layer:1 completing:1 activity:1 adapted:1 encodes:1 diffuse:1 maher:1 felt:1 transconductance:5 expanded:1 optical:2 describes:1 across:1 smaller:1 rev:1 happens:1 pr:1 depletion:1 taken:2 previously:1 remains:1 mechanism:1 letting:1 ge:1 operation:5 obey:2 away:1 gate:3 rp:3 top:1 recombination:1 leakage:1 arrangement:1 spike:1 responds:1 darkened:1 distance:1 thank:1 lateral:1 outer:1 extent:1 assuming:1 length:1 ratio:3 ql:10 potentially:1 frank:2 negative:1 rise:1 arc:1 immediate:1 incorporated:1 dc:7 delbriick:1 intensity:29 pair:2 required:1 neurolci:1 california:1 including:1 memory:1 hybrid:1 improve:1 technology:2 inversely:2 eye:1 started:2 mediated:1 drain:3 relative:4 law:2 suggestion:1 interesting:3 proportional:3 mounted:1 analogy:1 foundation:1 incident:3 metal:1 pi:1 supported:1 bias:1 institute:1 fall:1 feedback:14 curve:9 valid:1 foundry:1 quantum:3 made:1 q4:2 photoreceptors:1 symmetrical:3 sat:1 decade:2 channel:1 ca:1 operational:1 vj:1 substituted:2 vjb:2 noise:5 arise:3 collector:5 levitt:2 neuronal:1 grinvald:2 exponential:3 third:1 shining:1 down:1 magnitude:3 hoped:1 illumination:2 psychoacoustics:1 intersection:3 logarithmic:2 led:5 simply:1 ganglion:3 neurophysiological:1 hitting:1 unexpected:1 ch:1 iive:2 ma:1 ann:2 flash:3 room:1 flickering:2 replace:1 change:18 included:1 determined:1 photoreceptor:10 intact:1 parasitic:1 support:1 |
406 | 1,370 | MELONET I: Neural Nets for Inventing
Baroque-Style Chorale Variations
Dominik Hornel
dominik@ira.uka.de
Institut fur Logik, Komplexitat und Deduktionssysteme
Universitat Fridericiana Karlsruhe (TH)
Am Fasanengarten 5
D-76128 Karlsruhe, Germany
Abstract
MELONET I is a multi-scale neural network system producing
baroque-style melodic variations. Given a melody, the system invents a four-part chorale harmonization and a variation of any
chorale voice, after being trained on music pieces of composers like
J. S. Bach and J . Pachelbel. Unlike earlier approaches to the learning of melodic structure, the system is able to learn and reproduce
high-order structure like harmonic, motif and phrase structure in
melodic sequences. This is achieved by using mutually interacting
feedforward networks operating at different time scales, in combination with Kohonen networks to classify and recognize musical
structure. The results are chorale partitas in the style of J. Pachelbel. Their quality has been judged by experts to be comparable to
improvisations invented by an experienced human organist.
1
INTRODUCTION
The investigation of neural information structures in music is a rather new, exciting research area bringing together different disciplines such as computer science,
mathematics, musicology and cognitive science. One of its aims is to find out what
determines the personal style of a composer. It has been shown that neural network
models - better than other AI approaches - are able to learn and reproduce styledependent features from given examples, e.g., chorale harmonizations in the style
of Johann Sebastian Bach (Hild et al., 1992) . However when dealing with melodic
sequences, e.g., folk-song style melodies, all of these models have considerable difficulties to learn even simple structures. The reason is that they are unable to
capture high-order structure such as harmonies , motifs and phrases simultaneously
occurring at multiple time scales. To overcome this problem, Mozer (Mozer, 1994)
D. Hamel
888
proposes context units that learn reduced descriptions of a sequence of individual
notes. A similar approach in MELONET (Feulner et Hornel, 1994) uses delayed
update units that do not fire each time their input changes but rather at discrete
time intervals. Although these models perform well on artificial sequences , they
produce melodies that suffer from a lack of global coherence.
The art of melodic variation has a long tradition in Western music . Almost every
great composer has written music pieces inventing variations of a given melody, e.g.,
Mozart's famous variations KV 265 on the melody "Ah! Vous dirai-je, Maman",
also known as "Twinkle twinkle little star". At the beginning of this tradition there
is the baroque type of chorale variations. These are organ or harpsichord variations
of a chorale melody composed for use in the Protestant church. A prominent representative of this kind of composition is J. Pachelbel (1653 - 1706) who wrote about
50 chorale variations or partitas on various chorale melodies.
2
TASK DESCRIPTION
Given a chorale melody, the learning task is achieved in two steps:
1. A chorale harmonization of the melody is invented.
2. One of the voices of the resulting chorale is chosen and provided with
melodic variations.
Both subtasks are directly learned from music examples composed by J. Pachelbel
and performed in an interactive composition process which results in a chorale
variation of the given melody. The first task is performed by HARMONET, a
neural network system which is able to harmonize melodies in the style of various
composers like J. S. Bach. The second task is performed by the neural network
system MELONET I, presented in the following . For simplicity we have considered
melodic variations consisting of 4 sixteenth notes for each melody quarter note .
This is the most common variation type used by baroque composers and presents a
good starting point for even more complex variation types, since there are enough
music examples for training and testing the networks, and because it allows the
representation of higher-scale elements in a rather straightforward way.
HARMONET is a system producing four-part chorales in various harmonization
styles, given a one-part melody. It solves a musical real-world problem on a performance level appropriate for musical practice. Its power is based on a coding scheme
capturing musically relevant information. and on the integration of neural networks
and symbolic algorithms in a hierarchical system, combining the advantages of both.
The details are not discussed in this paper. See (Hild et aI., 1992) or (Hornel et
Ragg, 1996a) for a detailed account .
3
A MULTI-SCALE NEURAL NETWORK MODEL
The learning goal is twofold. On the one hand, the results produced by the system
should conform to musical rules. These are melodic and harmonic constraints such
as the correct resolving of dissonances or the appropriate use of successive interval
leaps. On the other hand, the system should be able to capture stilistic features
from the learning examples, e.g., melodic shapes preferred by J. Pachelbel. The
observation of musical rules and the aesthetic conformance to the learning set can
be achieved by a multi-scale neural network model. The complexity of the learning
task is reduced by decomposition in three subtasks (see Figure 1):
MELONEI' I: Neural Netsfor Inventing Baroque-Style Chorale Variations
Harmony
T
889
T D, T S, T, D T
k=tmod4
: MelodIc Vanallon
Figure 1: Structure of the system and process of composing a new melodic variation.
A melody (previously harmonized by HARMONET) is passed to the supernet which
predicts the current motif class MGT from a local window given by melody notes MT
to MT +2 and preceding motif class MGT-I. A similar procedure is performed at a
lower time scale by the su bnet which predicts the next motif note Nt based on M CT,
current harmony HT and preceding motif note Nt-I. The result is then returned
to the supernet through the motif classifier to be considered when computing the
next motif class MCT +1 .
1. A melody variation is considered at a higher time scale as a sequence of
melodic groups, so-called motifs. Each quarter note of the given melody
is varied by one motif. Before training the networks, motifs are classified
according to their similarity.
2. One neural network is used to learn the abstract sequence of motif classes.
Motif classes are represented in a l-of-n coding form where n is a fixed
number of classes. The question it solves is: What kind of motif 'fits' a
melody note depending on melodic context and the motif that has occurred
before? No concrete notes are fixed by this network. It works at a higher
scale and will therefore be called stlpernet in the following.
3. Another neural network learns the implementation of abstract motif classes
into concrete notes depending on a given harmonic context. It produces
a sequence of sixteenth notes - four notes per motif - that result in a
melodic variation of the given melody. Because it works one scale below
the supernet, it is called stlbnet.
4. The subnet sometimes invents a sequence of notes that does not coincide
890
D. Homel
with the motif class determined by the supernet. This motif will be considered when computing the next motif class , however. and should therefore
match the notes previously formed by the subnet. It is therefore reclassified
by the motif classifier before the supernet determines the next motif class.
The motivation of this separation into supernet and subnet arised from the following
consideration : Having a neural network that learns sequences of sixteenth notes, it.
would be easier for this network to predict notes given a contour of each motif. i.e.
a sequence of interval directions to be produced for each quarter note. Consider
a human organist who improvises a melodic variation of a given melody in real
time. Because he has to take his decisions in a fraction of a second, he must at
least have some rough idea in mind about what kind of melodic variation should
be applied to the next melody note to obtain a meaningful continuation of the
variation. Therefore, a neural network was introduced at a higher time scale , the
training of which really improved the overall behavior of the system and not just
shifted the learning problem to another time scale.
4
MOTIF CLASSIFICATION AND RECOGNITION
In order to realize learning at different time scales as described above , we need
a recognition component to find a suitable classification of motifs . This can be
achieved using unsupervised learning, e.g. , agglomerative hierarchical clustering or
Kohonen's topological feature maps (Kohonen, 1990). The former has the disadvantage however that an appropriate distance measure is needed which determines the
similarity between small sequences of notes respectively intervals, whereas the latter
allows to obtain appropriate motif classes through self-organization within a twodimensional surface. Figure 2 displays the motif representation and distribution of
motif contours over a 10xlO Kohonen feature map. In MELONET I, the Kohonen
algorithm is applied to all motifs contained in the training set. Afterwards a corresponding motif classification tree is recursively built from the Kohonen map. While
cutting this classification tree at lower levels we can get more and more classes. One
important problem remains to find an appropriate number of classes for the given
learning task. This will be discussed in section 6.
~
... -"
....... .... .. .... ... ...... ....... .. ... .... ... .
, jJ 3jl
',
Winner
1st interval
2nd interval
3rd interval
-1
Figure 2: Motifrepresentation example (left) and motif contour distribution (right)
over a 10xlO Kohonen feature map developed from one Pachelbel chorale variation
(initial update area 6x6, initial adaptation height 0.95, decrease factor 0.995). Each
cell corresponds to one unit in the KFM. One can see the arrangement of regions
responding to motifs having different motif contours.
891
MELONEr I: Neural Nets for Inventing Baroque-Style Chorale Variations
5
REPRESENTATION
In general one can distinguish two groups of motifs: Melodic motifs prefer small intervals, mainly seconds, harmonic motifs prefer leaps and harmonizing notes (chord
notes) . Both motif groups heavily rely on harmonic information. In melodic motifs dissonances should be correctly resolved, in harmonic motifs notes must fit the
given harmony. Small deviations may have a significant effect on the quality of
musical results. Thus our idea was to integrate musical knowledge about interval
and harmonic relationships into an appropriate interval representation. Each note
is represented by its interval to the first motif note, the so-called reference note.
This is an important element contributing to the success of MELONET I. A similar
idea for Jazz improvisation was followed in (Baggi, 1992) .
The interval coding shown in Table 1 considers several important relationships:
neighboring intervals are realized by overlapping bits, octave invariance is represented using a special octave bit. The activation of the overlapping bit was reduced
from 1 to 0.5 in order to allow a better distinction of the intervals. 3 bits are
used to distinguish the direction of the interval , 7 bits represent interval size. Complementary intervals such as ascending thirds and descending sixths have similar
representations because they lead to the same note and can therefore be regarded as
harmonically equivalent. A simple rhythmic element was then added using a tenuto
bit (not shown -in Table 1) which is set when a note is tied to its predecessor. This
final 3+1+7+1=12 bit coding gave the best results in our simulations.
Table 1: Complementary Interval Coding
direction
ninth
octave
seventh
sixth
fifth
fourth
third
second
pnme
second
third
fourth
fifth
sixth
seventh
octave
ninth
\.
\.
\.
\.
\.
\.
\.
\.
-+
/'
/'
/'
/'
/'
/'
/'
/'
o0
o0
o0
1
1
1
100
100
1 0 0
1 o 0
1 o 0
010
o0 1
0 0 1
0 0 1
o0 1
o0 1
o0 1
0 0 1
o0 1
octave
1
1
0
0
0
0
0
0
0
0
0
0
0
0
0
1
1
interval size
0
1
0.5
0
0
0
0
0
1
0.5
0
0
0
0
0
1
0.5
0
0
1
0.5
0
0
0
0
0
1
0.5
0
0
0
0
0
1
0
0
0
1
0.5
0
0
0
0
0
1
0.5
0
0
0
0
0
0
0
0
0
1
0.5
0
0
0
0
0
1
0.5
0
0
0
0
0
0
0
0
0
1
0.5
0
0
0
0
0
1
0.5
0
0
0
0.5
0
0
0
0
0
1
0.5
0
0
0
0
0
1
0.5
0
0
1
0.5
0
0
0
0
0
1
0.5
0
0
0
0
0
1
0.5
0
Now we still need a representation for harmony. It can be encoded as a harmonic
field which is a vector of chord notes of the diatonic scale. The tonic T in C major
for example contains 3 chord notes - C, E and G - which correspond to the first,
third and fifth degree of the C major scale (1010100). This representation may be
further improved. We have already mentioned that each note is represented by the
interval to the first motif note (reference note). We can now encode the harmonic
field starting with the first motif note instead of the first degree of the scale . This is
equivalent to rotating the bits of the harmonic field vector. An example is displayed
in Figure 3. The harmony of the motif is the dominant D, the first motif note is
B which corresponds to the seventh degree of the C major scale. Therefore the
D.Homel
892
harmonic field for D (0100101) is rotated by one position to the right resulting in
(1010010). Starting with the first note B. the harmonic field indicates the intervals
that lead to harmonizing notes B, D and G. In the right part of Figure 3 one can
see a correspondance between bits activated in the harmonic field and bits set to 1
in the three interval codings. This kind of representation helps the neural network
to directly establish a relationship between intervals and given harmony.
,J
J 3d I
D
third up
sixth up
pnme
o0
o0
1
1
010
harmonic field
0
0
0
0
0
1
0.5
0
0
1
0
0
0
0
0
0
0.5
0
0
1
0
0
0
0.5
1
0
1
0
0
1
0
Figure 3: Example illustrating the relationship between interval coding and rotat.ed
harmonic field. Each note is represented by its interval to the first note.
6
PERFORMANCE
We carried out several simulations to evaluate the performance of the system. Many
improvements could be found however by just listening to the improvisations produced by the neural organist. One important problem was to find an appropriate
number of classes for the given learning task . The following table lists the classification rate on the learning and validation set of the supernet and the subnet using 5,
12 and 20 motif classes. The learning set was automatically built from 12 Pachelbel
chorale variations corresponding to 2220 patterns for the subnet and 555 for the
supernet. The validation set includes 6 Pachelbel variations corresponding to 1396
patterns for the subnet and 349 for the supernet. Supernet and subnet were then
trained independently with the RPROP learning algorithm .
learning set
validation set
5 classes
91.17%
49.85%
s'Upernet
12 classes
86.85%
40.69%
20 classes
87.57%
37.54%
5 classes
86.31%
79.15%
s'Ubnet
12 classes
93.92%
83.38%
20 classes
95 .68%
86.96%
The classification rate of both networks strongly depends on the number of classes,
esp. on the validation set of the supernet. The smaller the number of classes,
the better is the classification of the supernet because there are less alternatives
to choose from. We can also notice an opposite development of the classification
behavior for the subnet. The bigger the number of classes. the easier the subnet will
be able to determine concrete motif notes for a given motif class. One can imagine
that the optimal number of classes lies somewhere in the middle. Another idea is
to form a committee of networks each of which is trained with different number of
classes.
We have also tested MELONET I on melodies that do not belong to the baroque
era. Figure 4 shows a harmonization and variation of the melody "Twinkle twinkle
little star" used by Mozart in his famous piano variations. It was produced by a
network committee formed by 3*2=6 networks trained with 5, 12 and 20 classes.
7
CONCLUSION
We have presented a neural network system inventing baroque-style variations on
given melodies whose qualities are similar to those of an experienced human organ-
MELONEI'L? Neural Nets for Inventing Baroque-Style Chorale Variations
893
!.
Figure 4: Melodic variation on "Twinkle twinkle little star"
ist. The complex musical task could be learned introducing a multi-scale network
model with two neural networks cooperating at different time scales , together with
an unsupervised learning mechanism able to classify and recognize relevant musical
structure.
We are about to test this multi-scale approach on learning examples of other epochs,
e.g., on compositions of classical composers like Haydn and Mozart or on Jazz
improvisations. First results confirm that the system is able to reproduce stylespecific elements of other kinds of melodic variation as well. Another interesting
question is whether the global coherence of the musical results may be further
improved adding another network working at a higher level of abstraction, e.g., at.
a phrase level. In summary, we believe that this approach presents an important
step towards the learning of complete melodies.
References
Denis L. Baggi. NeurSwing: An Intelligent Workbench for the Investigation of
Swing in Jazz. In: Readings in Computer-Generated Music, IEEE Computer Society Press, pp. 79-94, 1992.
Johannes Feulner, Dominik Hornel. MELONET: Neural networks that learn
harmony-based melodic variations. In: Proceedings of the 1994 International Computer Music Conference. ICMA Arhus , pp. 121-124, 1994.
Hermann Hild, Johannes Feulner, Wolfram Menzel. HARMONET: A Neural Net
for Harmonizing Chorales in the Style of J. S. Bach. In: Advances in Neural
Information Processing 4 (NIPS 4), pp. 267-274. 1992.
Dominik Hornel, Thomas Ragg . Learning Musical Structure and Style by Recognition, Prediction and Evolution. In: Proceedings of the 1996 International Computer Music Conference. ICMA Hong Kong , pp. 59-62, 1996.
Dominik Hornel , Thomas Ragg. A Connectionist Model for the Evolution of Styles
of Harmonization. In: Proceedings of the 1996 International Conference on Music
Perception and Cognition. Montreal, 1996.
Teuvo Kohonen. The Self-Organizing Map. In: Proceedings of the IEEE, Vol. 78,
no. 9, pp. 1464-1480 , 1990.
Michael C. Mozer . Neural Network music composition by prediction . In: Connection Science 6(2,3), pp. 247-280 , 1994.
| 1370 |@word kong:1 illustrating:1 middle:1 nd:1 simulation:2 decomposition:1 recursively:1 initial:2 contains:1 feulner:3 current:2 nt:2 activation:1 written:1 must:2 realize:1 shape:1 update:2 beginning:1 wolfram:1 denis:1 successive:1 harmonize:1 height:1 harmonically:1 mgt:2 predecessor:1 behavior:2 multi:5 automatically:1 little:3 window:1 provided:1 what:3 kind:5 developed:1 every:1 interactive:1 classifier:2 unit:3 producing:2 before:3 local:1 esp:1 era:1 improvisation:4 testing:1 practice:1 procedure:1 area:2 melodic:21 symbolic:1 get:1 judged:1 twodimensional:1 context:3 descending:1 equivalent:2 map:5 straightforward:1 starting:3 independently:1 simplicity:1 rule:2 regarded:1 his:2 variation:32 imagine:1 heavily:1 us:1 element:4 recognition:3 predicts:2 invented:2 capture:2 kfm:1 region:1 decrease:1 chord:3 mentioned:1 mozer:3 und:1 complexity:1 personal:1 trained:4 resolved:1 various:3 represented:5 artificial:1 whose:1 encoded:1 final:1 sequence:11 advantage:1 net:4 adaptation:1 kohonen:8 relevant:2 combining:1 neighboring:1 organizing:1 sixteenth:3 description:2 kv:1 produce:2 organist:3 rotated:1 help:1 depending:2 montreal:1 solves:2 direction:3 hermann:1 correct:1 human:3 melody:25 subnet:9 musically:1 inventing:6 really:1 investigation:2 baroque:9 harmonization:6 hild:3 considered:4 great:1 cognition:1 predict:1 major:3 jazz:3 harmony:8 leap:2 organ:2 rough:1 aim:1 rather:3 harmonizing:3 ira:1 encode:1 improvement:1 fur:1 indicates:1 mainly:1 tradition:2 am:1 motif:47 abstraction:1 reproduce:3 germany:1 overall:1 classification:8 deduktionssysteme:1 proposes:1 development:1 art:1 integration:1 special:1 harmonet:4 field:8 having:2 dissonance:2 unsupervised:2 connectionist:1 intelligent:1 composed:2 simultaneously:1 recognize:2 individual:1 delayed:1 consisting:1 fire:1 organization:1 johann:1 activated:1 folk:1 institut:1 tree:2 rotating:1 classify:2 earlier:1 disadvantage:1 phrase:3 introducing:1 deviation:1 menzel:1 seventh:3 teuvo:1 universitat:1 musicology:1 st:1 international:3 discipline:1 michael:1 together:2 concrete:3 choose:1 cognitive:1 expert:1 style:15 account:1 de:1 star:3 coding:7 includes:1 depends:1 piece:2 performed:4 correspondance:1 formed:2 musical:11 who:2 correspond:1 famous:2 produced:4 ah:1 classified:1 sebastian:1 ed:1 sixth:4 rprop:1 pp:6 knowledge:1 higher:5 ubnet:1 x6:1 improved:3 strongly:1 just:2 hand:2 working:1 su:1 overlapping:2 lack:1 western:1 quality:3 believe:1 karlsruhe:2 effect:1 former:1 swing:1 evolution:2 hamel:1 mozart:3 self:2 hong:1 prominent:1 octave:5 complete:1 harmonic:15 consideration:1 common:1 quarter:3 mt:2 winner:1 jl:1 discussed:2 occurred:1 he:2 belong:1 significant:1 composition:4 ai:2 rd:1 mathematics:1 similarity:2 operating:1 surface:1 dominant:1 success:1 preceding:2 determine:1 resolving:1 multiple:1 afterwards:1 match:1 bach:4 long:1 bigger:1 prediction:2 sometimes:1 represent:1 achieved:4 cell:1 whereas:1 interval:25 uka:1 unlike:1 bringing:1 feedforward:1 aesthetic:1 enough:1 fit:2 gave:1 opposite:1 idea:4 listening:1 whether:1 o0:10 passed:1 song:1 suffer:1 returned:1 jj:1 detailed:1 johannes:2 reduced:3 continuation:1 shifted:1 notice:1 per:1 correctly:1 conform:1 logik:1 discrete:1 vol:1 diatonic:1 group:3 ist:1 four:3 ht:1 cooperating:1 fraction:1 fourth:2 almost:1 separation:1 coherence:2 decision:1 prefer:2 comparable:1 bit:10 capturing:1 ct:1 followed:1 distinguish:2 display:1 topological:1 constraint:1 according:1 combination:1 smaller:1 mutually:1 previously:2 remains:1 committee:2 mechanism:1 needed:1 mind:1 ascending:1 hierarchical:2 appropriate:7 alternative:1 voice:2 thomas:2 responding:1 clustering:1 supernet:12 music:11 somewhere:1 establish:1 workbench:1 classical:1 society:1 question:2 arrangement:1 realized:1 added:1 already:1 distance:1 unable:1 agglomerative:1 considers:1 reason:1 relationship:4 implementation:1 perform:1 observation:1 displayed:1 tonic:1 interacting:1 varied:1 ninth:2 subtasks:2 introduced:1 connection:1 learned:2 distinction:1 nip:1 able:7 below:1 pattern:2 perception:1 reading:1 built:2 komplexitat:1 power:1 suitable:1 difficulty:1 rely:1 chorale:20 scheme:1 church:1 carried:1 epoch:1 piano:1 contributing:1 interesting:1 validation:4 integrate:1 degree:3 exciting:1 summary:1 allow:1 rhythmic:1 fifth:3 overcome:1 world:1 contour:4 invents:2 coincide:1 preferred:1 cutting:1 wrote:1 dealing:1 confirm:1 global:2 reclassified:1 table:4 learn:6 composing:1 composer:6 complex:2 ragg:3 motivation:1 complementary:2 je:1 representative:1 experienced:2 position:1 lie:1 tied:1 dominik:5 third:5 learns:2 list:1 adding:1 occurring:1 easier:2 mct:1 contained:1 xlo:2 corresponds:2 determines:3 goal:1 towards:1 harmonized:1 twofold:1 considerable:1 change:1 determined:1 called:4 invariance:1 meaningful:1 latter:1 bnet:1 evaluate:1 tested:1 |
407 | 1,371 | Detection of first and second order motion
Alexander Grunewald
Division of Biology
California Institute of Technology
Mail Code 216-76
Pasadena, CA 91125
alex@vis.caltech.edu
Heiko Neumann
Abteilung Neuroinformatik
Vniversitat VIm
89069 VIm
Germany
hneumann@neuro.informatik.uni-ulm.de
Abstract
A model of motion detection is presented. The model contains
three stages. The first stage is unoriented and is selective for contrast polarities. The next two stages work in parallel. A phase
insensitive stage pools across different contrast polarities through
a spatiotemporal filter and thus can detect first and second order
motion. A phase sensitive stage keeps contrast polarities separate,
each of which is filtered through a spatiotemporal filter, and thus
only first order motion can be detected. Differential phase sensitivity can therefore account for the detection of first and second order
motion. Phase insensitive detectors correspond to cortical complex
cells, and phase sensitive detectors to simple cells.
1
INTRODUCTION
In our environment objects are constantly in motion, and the visual system faces
the task of identifying the motion of objects. This task can be subdivided into two
components: motion detection and motion integration. In this study we will look at
motion detection. Recent psychophysics has made a useful distinction between first
and second order motion. In first order motion an absolute image feature is moving.
For example, a bright bar moving on a dark background is an absolute feature
because luminance is moving. In second order motion a relative image feature is
moving, for example a contrast reversing bar. No longer is it possible to identify
the moving object through its luminance, but only that it has different luminance
with respect to the background. Humans are very sensitive to first order motion,
but can they detect second order motion? Chubb & Sperling (1988) showed that
subjects are in fact able to detect second order motion. These findings have since
been confirmed in many psychophysical experiments, and it has become clear that
the parameters that yield detection of first and second order motion are different,
suggesting that separate motion detection systems exist.
A. Grunewald and H. Neumann
802
1.1
Detection of first and second order motion
First order motion, which is what we encounter in our daily lives, can be easily
detected by finding the peak in the Fourier energy distribution. The motion energy
detector developed by Adelson & Bergen (1985) does this explicitly, and it turns
out that it is also equivalent to a Reichardt detector (van Santen & Sperling, 1985).
However, these detectors cannot adequately detect second order motion, because
second order motion stimuli often contain the maximum Fourier energy in the opposite direction (possibly at a different velocity) as the actual motion. In other
words, purely linear filters, should have opposite directional tuning for first and
second order motion. This is further illustrated in Figure 1.
FIRST ORDER MOTION
Stimulus
Energy
Reconstructed
80
G)
E
60
=40
20
0
20 4060 80100
20 40 60 80 100
20 40 60 80 100
space
SECOND ORDER MOTION
Stimulus
Energy
80
80
~60
60
;:;40
40
20
20
0
20 40 60 80 100
space
0
20 40 60 80 100
Reconstructed
20 40 60 80 100
Figure 1: Schematic of first and second order motion, their peak Fourier energy,
and the reconstruction. The peak Fourier energy is along the direction of motion
for first order motion, and in the opposite direction for second order motion. For
this reason a linear filter cannot detect second order motion.
One way to account for second order motion detection is to transform the second
order motion signal into a first order signal. H second order motion is defined by
contrast reversals, then detecting contrast edges and then rectifying the resulting
signal of contrast will yield a first order motion signal. Thus this approach includes
three steps: orientation detection, rectification and finally motion detection (Wilson
et al., 1992) .
803
Detection of First and Second Order Motion
1.2
Visual physiology
Cells in the retina and the lateral geniculate nucleus (LGN) have concentric (and
hence unoriented) receptive fields which are organized in an opponent manner.
While the center of such an ON cell is excited by a light increment, the surround is
excited by a light decrement, and vice versa for OFF cells. It is only at the cortex
that direction and orientation selectivity arise. Cortical simple cells are sensitive to
the phase of the stimulus, while complex cells are not (Hubel & Wiesel, 1962).
Most motion models take at least partial inspiration from known physiology and
anatomy, by relating the kernels of the motion detectors to the physiology of cortical cells. The motion energy model in particular detects orientation and first order
motion at the same time. Curiously, all motion models essentially ignore the concentric opponency of receptive fields in the LG N. This is usually justified by pointing
to the linearity of simple cells with respect to stimulus parameters. However, it
has been shown that simple cells in fact exhibit strong nonlinearities (Hammond
& MacKay, 1983). Moreover, motion detection does require at least one stage of
nonlinearity (Poggio & Reichardt, 1973). The present study develops a model of
first and second order motion detection which explicitly includes an unoriented processing stage, and phase sensitive and phase insensitive motion detectors are built
from these unoriented signals. The former set of detectors only responds to first
order motion, while the second set of detectors responds to both types of motion.
We further show the analogies that can be drawn between these detector types and
simple and complex cells in cat visual cortex.
2
MODEL DESCRIPTION
The model is two-dimensional, one dimension is space, which means that space has
been collapsed onto a line, and the other dimension is time. The input image to
the model is a space-time matrix of luminances, as shown in figure 1. At each
processing stage essentially the same operations are performed. First the input
signal is convolved with the appropriate kernel. At each stage there are multiple
kernels, to generate the different signal types at that stage. For example, there
are ON and OFF signals at the unoriented stage. Next the convolved responses
are subtracted from each other. At the unoriented stage this means ON-OFF and
OFF-ON. In the final step these results are half-wave rectified to only yield positive
signals.
Unoriented
Phase insensitive
Phase sensitive
space
Figure 2: The kernels in the model. For the unoriented (left plot) and phase sensitive
(right plot) kernel plots black indicates OFF regions, white ON regions, and grey
zero input. For the phase insensitive plot (middle) grey denotes ON and OFF input,
and black denotes zero input.
804
A. Grunewald and H. Neumann
At the unoriented stage the input pattern is convolved with a difference of Gaussians kernel. This kernel has only a spatial dimension, no temporal dimension (see
figure 2). As described earlier, competition is between ON and OFF signals, followed by half-wave rectification. This ensures that at each location only one set of
unoriented signals is present. A simulation of the signals at the unoriented stage is
shown in figure 3. For first order motion, ON signals are at locations corresponding
to the inside of the moving bar. With each shift of the bar the signals also move.
Similarly, the OFF signals correspond to the outside of the bar, and also move with
the bar. For second order motion the contrast polarity reverses. Thus ON signals
correspond to the inside when the bar is bright, and to the outside when the bar is
dark, and vice versa for OFF signals. Thus any ON or OFF signals to the leading
edge of the bar will remain active after the bar moves.
Stimulus
80
.::
20
o
."
...-
Unoriented
ON
OFF
.rl'
20 40 60 80 100
space
Stimulus
20 40 60 80 100
20 40 60 80 100
OFF
ON
80
60
40
20
20 40 60 80 100
20 40 60 80 100
0
20 40 60 80 100
space
Figure 3: Unoriented signals to first and second order motion. ON signals are at
the bright side of any contrast transition, while OFF signals are at the dark side.
In first order motion ON and OFF move synchronously to the moving stimulus. In
second order motion ON and OFF signals persist, since the leading edge becomes
the trailing edge, and at the same time the contrast reverses, which means that at
a particular spatial location the contrast remains constant.
At the phase insensitive stage the unoriented ON and OFF signals are added, and
then the result is convolved with an energy detection filter.- The pooling of ON
and OFF signals means that the contrast transitions in the image are essentially
full-wave rectified. This causes phase insensitivity. These pooled signals are then
convolved with a space-time oriented filter (see figure 2). Competition between opposite directions of motion ensures that only one direction is active. A consequence
of the pooling of unoriented ON and OFF signals at this stage is that the resulting signals are invariant to first or second order motion. Thus phase insensitivity
Detection of First and Second Order Motion
805
makes this stage able to detect both first and second order motion. These signals
are shown in figure 4. In a two-dimensional extension of this model these detectors
would also be orientation selective. The simplest way to obtain this would be via
elongation along the preferred orientation.
Phase insensitive
Stimulus
left
20 40 60 80 100
right
20 40 60 80 100
space
Stimulus
20 40 60 80 100
left
right
80
::??
60
40
20
20 40 60 80 100
space
o
20 40 60 80 100
?
?
/
20 40 60 80 100
Figure 4: Phase insensitive signals to first and second order motion. For both
stimuli there are no leftwards signals, and robust rightwards signals.
At the phase sensitive stage unoriented ON and OFF signals are separately convolved with space-time oriented kernels which are offset with respect to each other
(see figure 2). The separate treatment of ON and OFF signals yields phase sensitivity. At each location there are four kernels: two for the two directions of motion,
and two for the two phases. Competition occurs between signals of opposite direction tuning, and opposite phase preference. To avoid activation in the opposite
direction of motion slightly removed from the location of the edge spatially broadly
tuned inhibition is necessary. This is provided by the phase insensitive signals, thus
avoiding feedback loops among phase sensitive detectors. First order signals from
the unoriented stage match the spatiotemporal filters in the preferred direction, and
thus phase sensitive signals arise. However, due to their phase reversal, second order motion input, provides poor motion signals, which are quenched through phase
insensitive inhibition. These signals are shown in figure 5.
These simulations show that first and second order motion are detected differently.
First order motion is detected by phase sensitive and phase insensitive motion detectors, while second order motion is only detected by the latter. From this we
conclude that first order motion is a more potent stimulus, and that the detection
of second order is more restricted, since it depends on a single type of detector. In
particular, the size of the stimulus and its velocity have to be matched to the energy
A. Grunewald and H. Neumann
806
Phase sensitive
Stimulus
20 40 60 80 100
space
Stimulus
20 40 60 80 100
DL left
DL right
20 40 60 80 100
DL left
20 40 60 80 100
DL right
20 40 60 80 100
space
Figure 5: Phase sensitive signals to first and second order motion. Only the darklight signals are shown. First order motion causes a consistent rightward motion
signal, while second order motion does not.
filters for motion signals to arise.
3
RELATION TO PHYSIOLOGY
The relationship between the model and physiology is straightforward. Unoriented
signals correspond to LGN responses, phase insensitive signals to complex cell responses, and phase sensitive signals to simple cell responses. Thus the model suggests that both simple and some complex cells receive direct LGN input. Moreover
these complex cells inhibit simple cells. With an additional threshold in simple
cells this inhibition could also be obtained via complex to simple cell excitation.
We stress that we are not ruling out that many complex cells receive only simple
cell input. Rather, the present research shows that if all complex cells receive only
simple cell input, second order motion cannot be detected. Hence at least some
complex cell responses need to be built up directly from LGN responses. Several
lines of evidence from cat physiology support this suggestion. First, the mean latencies of simple and complex cells are about equal (Bullier & Henry, 1979), suggesting
that at least some complex cells receive direct LGN input. Second, noise stimuli
can selectively activate complex cells, without activation of simple cells (Hammond,
1991). Third, cross-correlation analyses show that complex cells do receive simple
cell input (Ghose et ai., 1994).
The present model predicts that some cortical complex cells should respond to
Detection of First and Second Order Motion
807
second order motion. Zhou & Baker (1993) investigated this, and found that some
complex cells in area 17 respond to second order motion. Moreover, they found
that simple cells of a particular first order motion preference did not reverse their
motion preference when stimulated with second order motion, which would occur
if simple cells were just linear filters. We interpret this as further evidence that
complex cells provide inhibitory input to simple cells. If complex cells are built up
from LGN input, then orientation selectivity in two-dimensional space cannot be
obtained based on simple cell input, but rather requires complex cells with elongated
receptive fields. Thus we predict that there ought to be a correlation in complex
cells between elongated receptive fields and dependence on direct LGN input.
In conclusion we have shown how the phase sensitivity of motion detectors can be
mapped onto the ability to detect only first order motion, or both first and second
order motion. This suggests that it is not necessary to introduce a orientation
detection stage before motion detection can take place, thus simplifying the model
of motion detection. Furthermore we have shown that the proposed model is in
accord with known physiology.
Acknow ledgments
This work was supported by the McDonnell-Pew program in Cognitive Neuroscience.
References
Adelson, E. & Bergen (1985). Spatiotemporal energy models for the perception of
motion. J. Opt. Soc. Am. A, 2, 284-299.
Bullier, J. & Henry, G. H. (1979). Ordinal position of neurons in cat striate cortex.
J. Neurophys., 42, 1251-1263.
Chubb, C. & Sperling, G. (1988). Drift-balanced random stimuli: a general basis
for studying non-Fourier motion perception. J. Opt. Soc . Am. A, 5, 1986-2007.
Ghose, G. M., Freeman, R. D. & Ohzawa, I. {1994}. Local intracortical connections
in the cat's visual cortex: postnatal development and plasticity. J. Neurophys., 12,
1290-1303.
Hammond, P. (1991). On the response of simple and complex cells to random dot
patterns. Vis. Res., 31,47-50.
Hammond, P. & MacKay, D. (1983). Influence of luminance gradient reversal On
simple cells in feline striate cortex. J. Physiol., 331, 69-87.
Hubel, D. H. & Wiesel, T. N. (1962). Receptive fields, binocular interaction and
functional architecture in the cat's visual cortex. J. Physiol., 160, 106-154.
Poggio, T. & Reichardt, W. (1973). Considerations on models of movement detection. Kybernetik, 12, 223-227.
van Santen & Sperling, G. (1985). Elaborated Reichardt detectors. J. Opt. Soc.
Am. A, 2, 300-321.
Wilson, H. R., Ferrera, V. P. & Yo, C. (1992). A psychophysically motivated model
for two-dimensional motion perception. Vis . Neurosci., 9, 79-97.
Zhou, Y.X. & Baker, C. L. (1993). A processing stream in mammalian visual cortex
neurons for non-Fourier responses. Science, 261, 98-101.
| 1371 |@word middle:1 wiesel:2 grey:2 simulation:2 simplifying:1 excited:2 contains:1 tuned:1 neurophys:2 activation:2 physiol:2 plasticity:1 plot:4 half:2 postnatal:1 filtered:1 provides:1 detecting:1 location:5 preference:3 along:2 direct:3 differential:1 become:1 grunewald:4 inside:2 introduce:1 manner:1 freeman:1 detects:1 actual:1 becomes:1 provided:1 linearity:1 moreover:3 matched:1 baker:2 what:1 developed:1 finding:2 ought:1 temporal:1 positive:1 before:1 local:1 consequence:1 kybernetik:1 ghose:2 black:2 suggests:2 area:1 physiology:7 word:1 quenched:1 cannot:4 onto:2 collapsed:1 influence:1 equivalent:1 elongated:2 center:1 straightforward:1 feline:1 identifying:1 increment:1 ulm:1 velocity:2 santen:2 mammalian:1 persist:1 predicts:1 region:2 ensures:2 movement:1 removed:1 inhibit:1 balanced:1 environment:1 purely:1 division:1 basis:1 rightward:1 easily:1 differently:1 cat:5 activate:1 detected:6 outside:2 ability:1 transform:1 final:1 reconstruction:1 interaction:1 loop:1 insensitivity:2 description:1 competition:3 neumann:4 object:3 strong:1 soc:3 revers:2 direction:10 anatomy:1 filter:9 human:1 require:1 subdivided:1 opt:3 extension:1 leftwards:1 predict:1 pointing:1 trailing:1 geniculate:1 sensitive:14 vice:2 heiko:1 rather:2 avoid:1 zhou:2 wilson:2 yo:1 indicates:1 contrast:12 detect:7 am:3 bergen:2 pasadena:1 relation:1 selective:2 lgn:7 germany:1 among:1 orientation:7 development:1 spatial:2 integration:1 psychophysics:1 mackay:2 field:5 equal:1 elongation:1 biology:1 look:1 adelson:2 stimulus:17 develops:1 retina:1 oriented:2 phase:33 detection:22 light:2 edge:5 partial:1 daily:1 necessary:2 poggio:2 re:1 opponency:1 earlier:1 spatiotemporal:4 psychophysically:1 peak:3 sensitivity:3 potent:1 bullier:2 off:20 pool:1 possibly:1 cognitive:1 leading:2 account:2 suggesting:2 nonlinearities:1 de:1 intracortical:1 pooled:1 includes:2 explicitly:2 vi:3 depends:1 stream:1 performed:1 wave:3 parallel:1 rectifying:1 elaborated:1 bright:3 correspond:4 identify:1 yield:4 directional:1 informatik:1 hammond:4 confirmed:1 rectified:2 detector:16 energy:11 treatment:1 organized:1 response:8 furthermore:1 just:1 stage:20 binocular:1 correlation:2 ohzawa:1 contain:1 adequately:1 hence:2 inspiration:1 former:1 spatially:1 illustrated:1 white:1 excitation:1 stress:1 motion:87 image:4 consideration:1 functional:1 rl:1 insensitive:12 unoriented:18 relating:1 interpret:1 surround:1 versa:2 ai:1 pew:1 tuning:2 similarly:1 nonlinearity:1 henry:2 dot:1 moving:7 longer:1 cortex:7 inhibition:3 recent:1 showed:1 reverse:1 selectivity:2 life:1 caltech:1 additional:1 signal:46 multiple:1 full:1 match:1 cross:1 schematic:1 neuro:1 essentially:3 kernel:9 accord:1 cell:41 justified:1 background:2 receive:5 separately:1 subject:1 pooling:2 ledgments:1 architecture:1 opposite:7 shift:1 motivated:1 curiously:1 cause:2 useful:1 latency:1 clear:1 dark:3 simplest:1 generate:1 exist:1 inhibitory:1 neuroscience:1 broadly:1 four:1 threshold:1 drawn:1 luminance:5 respond:2 place:1 ruling:1 chubb:2 followed:1 occur:1 alex:1 fourier:6 mcdonnell:1 poor:1 across:1 remain:1 slightly:1 invariant:1 restricted:1 rectification:2 remains:1 sperling:4 turn:1 ordinal:1 reversal:3 studying:1 operation:1 gaussians:1 opponent:1 appropriate:1 subtracted:1 encounter:1 convolved:6 denotes:2 psychophysical:1 move:4 added:1 occurs:1 receptive:5 dependence:1 striate:2 responds:2 exhibit:1 gradient:1 separate:3 mapped:1 lateral:1 mail:1 reason:1 code:1 polarity:4 relationship:1 lg:1 acknow:1 neuron:2 synchronously:1 drift:1 concentric:2 neuroinformatik:1 connection:1 california:1 distinction:1 able:2 bar:10 usually:1 pattern:2 perception:3 program:1 built:3 technology:1 reichardt:4 relative:1 suggestion:1 analogy:1 nucleus:1 consistent:1 supported:1 side:2 institute:1 face:1 absolute:2 van:2 feedback:1 dimension:4 cortical:4 transition:2 made:1 reconstructed:2 uni:1 ignore:1 preferred:2 keep:1 hubel:2 active:2 conclude:1 stimulated:1 robust:1 ca:1 investigated:1 complex:21 did:1 neurosci:1 decrement:1 noise:1 arise:3 position:1 vim:2 third:1 offset:1 evidence:2 dl:4 visual:6 constantly:1 reversing:1 selectively:1 support:1 latter:1 alexander:1 avoiding:1 rightwards:1 |
408 | 1,372 | From Regularization Operators
to Support Vector Kernels
Alexander J. Smola
Bernhard Scholkopf
GMDFIRST
Rudower Chaussee 5
12489 Berlin, Germany
smola@first.gmd.de
Max-Planck-Institut fur biologische Kybernetik
Spemannstra.Be 38
72076 Ttibingen, Germany
bs-@mpik-tueb.mpg.de
Abstract
We derive the correspondence between regularization operators used in
Regularization Networks and Hilbert Schmidt Kernels appearing in Support Vector Machines. More specifica1ly, we prove that the Green's Functions associated with regularization operators are suitable Support Vector
Kernels with equivalent regularization properties. As a by-product we
show that a large number of Radial Basis Functions namely conditionally positive definite functions may be used as Support Vector kernels.
1 INTRODUCTION
Support Vector (SV) Machines for pattern recognition, regression estimation and operator
inversion exploit the idea of transforming into a high dimensional feature space where
they perform a linear algorithm. Instead of evaluating this map explicitly, one uses Hilbert
Schmidt Kernels k(x, y) which correspond to dot products of the mapped data in high
dimensional space, i.e.
(I)
k(x, y) = (<I>(x) ? <I>(y))
with <I> : .!Rn --* :F denoting the map into feature space. Mostly, this map and many of
its properties are unknown. Even worse, so far no general rule was available. which kernel
should be used, or why mapping into a very high dimensional space often provides good
results, seemingly defying the curse of dimensionality. We will show that each kernel
k(x, y) corresponds to a regularization operator P, the link being that k is the Green's
function of P* P (with F* denoting the adjoint operator of F). For the sake of simplicity
we shall only discuss the case of regression - our considerations, however, also hold true
foi the other cases mentioned above.
~
We start by briefly reviewing the concept of SV Machines (section 2) and of Regularization
Networks (section 3). Section 4 contains the main result stating the equivalence of both
A. J. Smola and B. Schollwpf
344
methods. In section 5, we show some applications of this finding to known SV machines.
Section 6 introduces a new class of possible SV kernels, and, finally, section 7 concludes
the paper with a discussion.
2
SUPPORT VECTOR MACHINES
The SV algorithm for regression estimation, as described in [Vapnik, 1995] and [Vapnik
et al., 1997], exploit~ the idea of computing a linear function in high dimensional feature
space F (furnished with a dot product) and thereby computing a nonlinear function in the
space of the input data !Rn. The functions take the form f(x) = (w ? <ll(x)) + b with
ell : !Rn -+ :F and w E F.
In order to infer f from a training set {(xi, Yi) I i = 1, ... , f., Xi E !Rn, Yi E IR}, one tries
to minimize the empirical risk functional Remp[f] together with a complexity term l!wll 2 ,
thereby enforcingflatness in feature space, i.e. to minimize
1
Rreg[/] = Remp[!]
l
+ Allwll 2 =f. :L;c(f(xi),yi) + Allwll 2
(2)
i=l
with c(f(xi),yi) being the cost function determining how deviations of f(xi) from the
target values Yi should be penalized, and A being a regularization constant. As shown in
[Vapnik, 1995] for the case of ?-insensitive cost functions,
c(f(x)
) = { lf(x)- yl- ? for lf(_x)- Yl ;::: e
0
otherwtse
'
(3)
'Y
(2) can be minimized by solving a quadratic programming problem formulated in terms
of dot products in :F. It turns out that the solution can be expressed in terms of Support
Vectors, w = :Ef=I Cti<ll(xi), and therefore
f(x) =
l
l
i=l
i=l
L ai(<ll(xi) ? <ll(x)) + b = L aik(xi, x) + b,
(4)
where k(xi, x) is a kernel function computing a dot product in feature space (a concept
introduced by Aizerrnan et al. [ 1964]). The coefficients ai can be found by solving a
quadratic programming problem (with Kii := k(xi, Xj) and ai = f3i - {3i):
l
minimize
!L
l
(f3i- f3i)(f3j- f3J)Kij- "?, (!3i- f3i)Yi- (f3i
i,j=l
i=l
f.
subject to
"?, f3i - !3i
=
+ f3i)e
-
(5)
0, f3i, !3i E [0, ft] ?
i=l
Note that (3) is not the only possible choice of cost functions resulting in a quadratic programming problem (in fact quadratic parts and infinities are admissible, too). For a detailed
discussion see [Smola and Scholkopf, 1998]. Also note that any continuous symmetric
function k(x, y) E L 2 ? L 2 may be used as an admissible Hilbert-Schmidt kernel if it
satisfies Mercer's condition
/ / k(x,y)g(x)g(y)dxdy;::: 0 for all g E
3
L 2 (IRn)~
(6)
REGULARIZATION NETWORKS
Here again we start with minimizing the empirical risk functional Remp[!] plus a regularization term liP /11 2 defined by a regularization operator Pin the sense of Arsenin and
From Regularization Operators to Support Vector Kernels
345
Tikhonov [1977]. Similar to (2), we minimize
Rreg[f] = Remp
1"'
~
?_-c(f(xi),yi) + -XIIPJII
?
l
~ 2
+ .\IIPJII
= f
2
(7)
i=1
Using an expansion off in terms of some symmetric function k(xi, Xj) (note here, that k
need not fulfil Mercer's condition),
f(x) =
2: aik(xi, x) + b,
(8)
and the cost function defined in (3 ), this leads to a quadratic programming problem similar
to the one for SVs: by computing Wolfe's dual (for details of the calculations see [Smola
and SchOlkopf, 1998]), and using
Dij := ((Fk)(xi, .) ? (Fk)(xj, .))
(9)
((f ? g) denotes the dot product of the functions f and g in Hilbert Space, t.e.
I !(x)g(x)dx), we get n- 1 K(i]- /3*), with f3i, f3i being the solution of
a=
l
mm1m1Ze
! 2:
l
(f3i- f3i)(f3j- {3j)(KD- 1 K)ij-
i,j=1
f_
subject to
L
L (f3i- f3i)Yi- (f3i + f3i)E
i=l
f3i - f3i = 0, f3i, f3i E [0, A]
i=1
(I 0)
Unfortunately this setting of the problem does not preserve sparsity in terms of the coefficients, as a potentially sparse decomposition in terms of f3i and f3i is spoiled by n- 1 K,
which in general is not diagonal (the expansion (4) on the other hand does typically have
many vanishing coefficients).
4
THE EQUIVALENCE OF~BOTH METHODS
Comparing (5) with (10) leads to the question if and under which condition the two methods
might be equivalent and therefore also under which conditions regularization networks
might lead to sparse decompositions (i.e. only a few of the expansion coefficients in f
would differ from zero). A sufficient condition is D = K (thus K n- 1 K = K), i.e.
(11)
Our goal now is twofold:
? Given a regularization operator P, find a kernel k such that a SV machine using k
will not only enforce flatness in feature space, but also correspond to minimizing
a regularized risk functional with P as regularization operator.
? Given a Hilbert Schmidt kernel k, find a regularization operator P such that a SV
machine using this kernel can be viewed as a Regularization Network using P.
These two problems can be solved by employing the concept of Green's functions as described in [Girosi et al., 1993]. These functions had been introduced in the context of
solving differential equations. For our purpose, it is sufficient to know that the Green's
functions Gx, (x) ofF* P satisfy
(12)
Here, 8xi (x) is the 8-distribution (not to be confused with the Kronecker symbol8ij) which
has the property that (f ? 8x.) = f(xi). Moreover we require for all Xi the projection of
Gx, (x) onto the null space of F* P to be zero. The relationship between kernels and
regularization operators is formalized in the following proposition.
A. I. Smola and B. Schtilkopf
346
Proposition 1
Let P be a regularization operator, and G be the Green's function of P* P. Then G is a
Hilbert Schmidt-Kernel such that D = K. SV machines using G minimize risk functional
(7) with Pas regularization operator.
Proof: Substituting ( 12) into GxJ (xi) = (GxJ (.) ? 8x? (.)) yields
(13)
Gxi (xi) = ( (PGx, ){.) ? (PGxJ(.)) = Gx; {xj),
hence G(xi,Xj) := Gx,{xj) is symmetric and satisfies (11). Thus the SVoptimization
problem (5) is equivalent to the regularization network counterpart ( 10). Furthermore G is
an admissible positive kernel, as it can be written as a dot product in Hilbert Space, namely
(14)
In the following we will exploit this relationship in both ways: to compute Green's functions for a given regularization operator P and to infer the regularization operator from a
given kernel k.
5 TRANSLATION INVARIANT KERNELS
Let us now more specifically consider regularization operators
multiplications in Fourier space [Girosi et al., 1993]
(Pi? Pg ) =
1
(21l')n/2
f
Jn
P that may
be written as
f{w)fJ(w) dw
P(w)
(15)
with ](w) denoting the Fourier transform of j(x), and P(w) = P( -w) real valued, nonnegative and converging uniformly to 0 for lwl --+ oo and
supp[P(w)]. Small values
of P(w) correspond to a strong attenuation of the corresponding frequencies.
n=
For regularization operators defined in Fourier Space by (15) it can be shown by exploiting
P(w) = P(-w) = P(w) that
G(Xi x) =
l
1
{27!' )n/2
f
}JR.n
eiw(x;-x) P(w)dw
(16)
is a corresponding Green's function satisfying translational invariance, i.e. G(xi,xj) =
G (Xi - xi), and G(w) P (w). For the proof, one only has to show that G satisfies (11 ).
=
This provides us with an efficient tool for analyzing SV kernels and the types of capacity
control they exhibit
Example 1 (Bq-splines)
Vapnik et al. [ 1997] propose to use Bq-splines as building blocks for kernels, i.e.
(17)
i=l
with x E !Rn. For the sake of simplicity, we consider the case n
1. Recalling the
definition
(18)
.
Bq = ?q+ 1 1[-o.5,o.5]
(? denotes the convolution and Ix the indicator function on X), we can utilize the above
result and the Fourier-Plancherel identity to construct the Fourier representation of the
corresponding regularization operator. Up to a multiplicative constant, it equals
P(w)
= k(w) = sinc(q+l)(Wi).
2
(19)
From Regularization Operators to Support W?ctor Kernels
347
This shows that only B-splines of odd order are admissible, as the even ones have negative
parts in the Fourier spectrum (which would result in an amplification of the corresponding
frequency components). The zeros ink stem from the fact that B1 has only compact support
[-(k+ 1)/2, (k+ 1)/2). By using this kernel we trade reduced computational complexity in
calculatingf(we only have to take points with llxi- xi II :S cfrom some limited neighborhood determined by c into account)for a possibly worse performance of the regularization
operator as it completely ~emovesfrequencies wp with k(wp) = 0.
Example 2 (Dirichlet kernels)
In [Vapnik et al., 1997], a class of kernels generating Fourier expansions was introduced,
k(x) = sin(2N + 1)x/2.
(20)
sinx/2
(As in example 1 we consider x E ~ 1 to avoid tedious notation.) By construction, this
kernel corresponds to P(w) = ~ L~-N 6(w- i). A regularization operator with these
properties, however; may not be desirable as it only damps a finite number offrequencies
and leaves all other frequencies unchanged which can lead to overjitting (Fig. 1).
\
!\
'i
'I
I.
I? ,\
I l
05
I
\
:\
-I
.I
'I
-?
_,
-.:.,,\ 1\ .'i
\ I.\ r
""t-4
-15
-10
-?
\I
~
10
15
Figure 1: Left: Interpolation with a Dirichlet Kernel of order N = 10. One can clearly observe the overfitting (dashed line: interpolation, solid line: original data points, connected
by lines). Right: Interpolation of the same data with a Gaussian Kernel of width CT 2 = 1.
Example 3 (Gaussian kernels)
Following the exposition of Yuille and Grzywacz [1988] as described in [Girosi et al.,
1993], one can see that for
11P!II 2 =
I
dx
L
2m
~!2m com f(x)) 2
(21)
m
with 6 2 m = 6. m and 6 2m+l
tor; we get Gaussians kernels
= V' 6. m. 6. being the Laplacian and V' the Gradient operak(x) = exp (
-~~~~2 ).
(22)
Moreover; we can provide an equivalent representation of P in terms of its Fourier properties, i.e. P(w) = exp(- u2 \kxll 2 ) up to a multiplicative constant. Training a SV machine
with Gaussian RBF kernels [Scholkopf et al., 1997] corresponds to minimizing the specific
costfunction with a regularization operator of type (21 ). This also explains the good performance ofSV machines in this case, as it is by no means obvious that choosing a flat fum:;tion
in high dimensional space will correspond to a simple function in low dimensional space,
as showed in example 2. Gaussian kernels tend to yield good performance under g'eneral
smoothness assumptions and should be considered especially if no additional knowledge
of the data is available.
A. J. Smola and B. Scholkopf
348
6 A NEW CLASS OF SUPPORT VECTOR KERNELS
We will follow the lines of Madych and Nelson [ 1990] as pointed out by Girosi et al. [ 1993].
Our main statement is that conditionally positive definite functions (c.p.d.) generate admissible SV kernels. This is very useful as the property of being c.p.d. often is easier to verify
than Mercer's condition, especially when combined with the results of Schoenberg and
Micchelli on the connection between c.p.d. and completely monotonic functions [Schoenberg, 1938, Micchelli, 1986]. Moreover c.p.d. functions lead to a class of SV kernels that
do not necessarily satisfy Mercer's condition.
Definition 1 (Conditionally positive definite functions)
A continuous function h, defined on [0, oo), is said to be conditionally positive definite
? (c.p.d.) of order m on m.n iffor any distinct points x1, ... , Xt E m.n and scalars c1, ... , Ct
the quadratic form Eri=l cicih(llxi- Xj II) is nonnegative provided that E~=l Cip(xi) =
0 for all polynomials p on m.n of degree lower than m.
Proposition 2 (c.p.d. functions and admissible kernels)
Define II~ the space of polynomials of degree lower than m on IRn. Every c.p.d. function
h of order m generates an admissible Kernel for SV expansions on the space offunctions
f orthogonal to II~ by setting k(xi, Xj) := h(llxi- Xjll 2 ).
Proof: In [Dyn, /991] and [Madych and Nelson, 1990] it was shown that c.p.d. functions
h generate semi-norms 11-llh by
(23)
Provided that the projection off onto the space ofpolynomials of degree lower than m is
zero. For these functions, this, however. also defines a dot product in some feature space.
Hence they can be used as SV kernels.
Only c.p.d. functions of order m up to 2 are of practical interest for SV methods (for
details see [Smola and Scholkopf, 1998]). Consequently, we may use kernels like the ones
proposed in [Girosi et al., 1993] as SV kernels:
k(x,y) =
e-.BIJx-yJJ2
k(x,y) =
-v'llx- Yll 2 + c2
k(x,y)
=
k(x,y) =
1
y'Jix-yJ12+c2
llx- Yll 2 ln llx- Yll
= 0);
multiquadric, (m = 1)
Gaussian, (m
inverse multiquadric, (m
thin plate splines, (m
(24)
(25)
= 0)
= 2)
(26)
(27)
7 DISCUSSION
We have pointed out a connection between SV kernels and regularization operators. As one
of the possible implications of this result, we hope that it will deepen our understanding of
SV machines and of why they have been found to exhibit high generalization ability. In
Sec. 5, we have given examples where only the translation into the regularization framework provided insight in why certain kernels are preferable to others. Capacity control is
one of the strengths of SV machines; however, this does not mean that the structure of the
learning machine, i.e. the choice of a suitable kernel for a given task, should be disregarded.
On the contrary, the rather general class of admissible SV kernels should be seen as another
strength, provided that we have a means of choosing the right kernel. The newly established
link to regularization theory can thus be seen as a tool for constructing the structure consisting of sets of functions in which the SV machine (approximately) performs structural
From Regularization Operators to Support \iector Kernels
349
risk minimization (e.g. [Vapnik, 1995]). For a treatment of SV kernels in a Reproducing
Kernel Hilbert Space context see [Girosi, 1997].
Finally one should leverage the theoretical results achieved for regularization operators
for a better understanding of SVs (and vice versa). By doing so this theory might serve
as a bridge for connecting two (so far) separate threads of machine learning. A trivial
example for such a connection would be a Bayesian interpretation of SV machines. In
this case the choice of a special kernel can be regarded as a prior on the hypothesis space
with P[f] ex exp{ ->.IIF111 2 ). A more subtle reasoning probably will be necessary for
understanding the capacity bounds [Vapnik, 1995] from a Regularization Network point
of view. Future work will include an analysis of the family of polynomial kernels, which
perform very well in Pattern Classification [SchOlkopf et al., 1995].
Acknowledgements
AS is supported by a grant of the DFG (# Ja 379/51 ). BS is supported by the Studienstiftung
des deutschen Volkes. The authors thank Chris Burges, Federico Girosi, Leo van Hemmen,
Klaus-Robert Muller and Vladimir Vapnik for helpful discussions and comments.
References
M.A. Aizerman, E. M. Braverman, and L. I. Rozonoer. Theoretical foundations of the potential function method in pattern recognition learning. Automation and Remote Control,
25:821-837, 1964.
N. Dyn. Interpolation and approximation by radial and related functions. In C.K. Chui,
L.L. Schumaker, and D.J. Ward, editors, Approximation Theory, VI, pages 211-234.
Academic Press, New York, 1991.
F. Girosi. An equivalence between sparse approximation and suppm1 vector machines. A.I.
Memo No. 1606, MIT, 1997.
F. Girosi, M. Jones, and T. Poggio. Priors, stabilizers and basis functions: From regularization to radial, tensor and additive splines. A.I. Memo No. 1430, MIT, 1993.
W.R. Madych and S.A. Nelson. Multivariate interpolation and conditionally positive definite functions. II. Mathematics of Computation, 54(189):211-230, 1990.
C. A. Micchelli. Interpolation of scattered data: distance matrices and conditionally positive definite functions. Constructive Approximation, 2:11-22, 1986.
I.J. Schoenberg. Metric spaces and completely monotone functions. Ann. of Math., 39:
811-841, 1938.
B. Scholkopf, C. Burges, and V. Vapnik. Extracting support data for a given task. In U. M.
Fayyad and R. Uthurusamy, editors, Proc. KDD I, Menlo Park, 1995. AAAI Press.
B. SchOlkopf, K. Sung, C. Burges, F. Girosi, P. Niyogi, T. Poggio, and V. Vapnik. Comparing support vector machines with gaussian kernels to radial basis function classifiers.
IEEE Trans. Sign. Processing, 45:2758-2765, 1997.
A. J. Smola and B. SchOlkopf. On a kernel-based method for pattern recognition, regression, approximation and operator inversion. Algorithmica, 1998. see also GMD
Technical Report 1997- I 064, URL: http://svm.first.gmd.de/papers.html.
V. Vapnik. The Nature of Statistical Learning Theory. Springer Verlag, New York, 1995.
V. Vapnik, S. Golowich, and A. Smola: Support vector method for function approximation,
regression estimation, and signal processing. In NIPS 9, San Mateo, CA, 1997.
A. Yuille and N. Gr:z;ywacz. The motion coherence theory. In Proceedings of the International Conference on Computer Vision, pages 344-354, Washington, D.C., 1988. IEEE
Computer Society Press.
| 1372 |@word rreg:2 briefly:1 inversion:2 polynomial:3 norm:1 tedious:1 decomposition:2 pg:1 thereby:2 solid:1 contains:1 denoting:3 comparing:2 com:1 dx:2 written:2 additive:1 kdd:1 girosi:10 offunctions:1 wll:1 leaf:1 vanishing:1 provides:2 math:1 gx:4 c2:2 differential:1 scholkopf:10 prove:1 eiw:1 mpg:1 curse:1 confused:1 provided:4 moreover:3 notation:1 null:1 finding:1 sung:1 every:1 attenuation:1 preferable:1 classifier:1 control:3 grant:1 planck:1 positive:7 kybernetik:1 analyzing:1 interpolation:6 approximately:1 might:3 plus:1 mateo:1 equivalence:3 limited:1 practical:1 block:1 definite:6 lf:2 empirical:2 projection:2 radial:4 get:2 onto:2 operator:27 risk:5 context:2 equivalent:4 map:3 simplicity:2 formalized:1 rule:1 insight:1 regarded:1 dw:2 fulfil:1 schoenberg:3 grzywacz:1 rozonoer:1 target:1 construction:1 aik:2 programming:4 us:1 spoiled:1 hypothesis:1 pa:1 wolfe:1 recognition:3 satisfying:1 ft:1 solved:1 connected:1 remote:1 trade:1 mentioned:1 transforming:1 complexity:2 rudower:1 reviewing:1 solving:3 yuille:2 serve:1 basis:3 completely:3 leo:1 distinct:1 klaus:1 neighborhood:1 choosing:2 valued:1 federico:1 ability:1 niyogi:1 ward:1 transform:1 seemingly:1 propose:1 schumaker:1 product:8 adjoint:1 amplification:1 exploiting:1 costfunction:1 generating:1 mpik:1 derive:1 oo:2 stating:1 golowich:1 ij:1 odd:1 strong:1 differ:1 explains:1 kii:1 require:1 ja:1 generalization:1 proposition:3 hold:1 considered:1 exp:3 mapping:1 substituting:1 tor:1 purpose:1 estimation:3 proc:1 bridge:1 vice:1 tool:2 hope:1 minimization:1 mit:2 clearly:1 gaussian:6 rather:1 avoid:1 fur:1 sense:1 helpful:1 typically:1 irn:2 germany:2 translational:1 dual:1 classification:1 html:1 special:1 ell:1 equal:1 construct:1 washington:1 park:1 jones:1 thin:1 future:1 minimized:1 others:1 spline:5 report:1 few:1 preserve:1 dfg:1 algorithmica:1 consisting:1 sinx:1 volkes:1 recalling:1 interest:1 cip:1 braverman:1 introduces:1 dyn:2 implication:1 necessary:1 poggio:2 institut:1 spemannstra:1 bq:3 orthogonal:1 theoretical:2 kij:1 cost:4 deviation:1 dij:1 gr:1 too:1 damp:1 sv:23 combined:1 international:1 yl:2 off:3 together:1 connecting:1 again:1 aaai:1 possibly:1 worse:2 supp:1 account:1 potential:1 de:4 sec:1 automation:1 coefficient:4 satisfy:2 explicitly:1 vi:1 multiplicative:2 try:1 tion:1 view:1 doing:1 biologische:1 start:2 minimize:5 ir:1 correspond:4 yield:2 bayesian:1 foi:1 definition:2 frequency:3 obvious:1 associated:1 proof:3 newly:1 treatment:1 remp:4 f3j:3 knowledge:1 dimensionality:1 hilbert:8 subtle:1 follow:1 furthermore:1 smola:10 hand:1 nonlinear:1 defines:1 building:1 concept:3 true:1 verify:1 counterpart:1 regularization:37 hence:2 symmetric:3 wp:2 conditionally:6 ll:4 sin:1 width:1 plate:1 performs:1 motion:1 fj:1 reasoning:1 lwl:1 consideration:1 ef:1 functional:4 insensitive:1 interpretation:1 versa:1 ai:3 llx:3 smoothness:1 fk:2 mathematics:1 pointed:2 had:1 dot:7 multivariate:1 showed:1 tikhonov:1 certain:1 verlag:1 yi:8 muller:1 seen:2 dxdy:1 additional:1 dashed:1 semi:1 ii:6 signal:1 flatness:1 desirable:1 infer:2 stem:1 technical:1 academic:1 calculation:1 schtilkopf:1 laplacian:1 converging:1 regression:5 vision:1 metric:1 kernel:52 achieved:1 c1:1 chaussee:1 probably:1 comment:1 subject:2 tend:1 contrary:1 extracting:1 structural:1 leverage:1 xj:9 f3i:22 idea:2 stabilizer:1 thread:1 url:1 iffor:1 york:2 svs:2 useful:1 detailed:1 tueb:1 gmd:3 reduced:1 generate:2 http:1 xjll:1 sign:1 shall:1 utilize:1 monotone:1 inverse:1 furnished:1 family:1 coherence:1 bound:1 ct:2 correspondence:1 quadratic:6 nonnegative:2 strength:2 infinity:1 kronecker:1 flat:1 sake:2 generates:1 fourier:8 chui:1 fayyad:1 kd:1 jr:1 eneral:1 wi:1 b:2 invariant:1 ln:1 equation:1 discus:1 turn:1 pin:1 know:1 available:2 gaussians:1 observe:1 enforce:1 appearing:1 uthurusamy:1 schmidt:5 jn:1 original:1 denotes:2 dirichlet:2 include:1 eri:1 exploit:3 especially:2 society:1 unchanged:1 ink:1 micchelli:3 tensor:1 question:1 diagonal:1 said:1 exhibit:2 gradient:1 distance:1 link:2 mapped:1 berlin:1 capacity:3 separate:1 thank:1 gxi:1 nelson:3 chris:1 trivial:1 relationship:2 minimizing:3 vladimir:1 mostly:1 unfortunately:1 robert:1 potentially:1 statement:1 negative:1 memo:2 unknown:1 perform:2 convolution:1 ctor:1 finite:1 rn:5 reproducing:1 introduced:3 namely:2 connection:3 yll:3 established:1 nip:1 trans:1 deepen:1 aizerman:1 pattern:4 sparsity:1 max:1 green:7 suitable:2 regularized:1 indicator:1 concludes:1 prior:2 understanding:3 acknowledgement:1 multiplication:1 determining:1 foundation:1 degree:3 sufficient:2 mercer:4 editor:2 pi:1 translation:2 arsenin:1 penalized:1 deutschen:1 supported:2 burges:3 sparse:3 van:1 evaluating:1 llh:1 author:1 san:1 far:2 employing:1 compact:1 bernhard:1 overfitting:1 b1:1 xi:27 spectrum:1 continuous:2 why:3 lip:1 nature:1 plancherel:1 f_:1 menlo:1 ca:1 expansion:5 necessarily:1 constructing:1 main:2 x1:1 fig:1 hemmen:1 scattered:1 ix:1 admissible:8 specific:1 xt:1 svm:1 sinc:1 studienstiftung:1 vapnik:12 disregarded:1 easier:1 expressed:1 scalar:1 u2:1 monotonic:1 springer:1 corresponds:3 satisfies:3 cti:1 goal:1 formulated:1 viewed:1 identity:1 llxi:3 exposition:1 rbf:1 twofold:1 consequently:1 ann:1 specifically:1 determined:1 uniformly:1 invariance:1 multiquadric:2 support:15 alexander:1 constructive:1 ex:1 |
409 | 1,373 | Globally Optimal On-line Learning Rules
Magnus Rattray*and David Saad t
Department of Computer Science & Applied Mathematics,
Aston University, Birmingham B4 7ET, UK.
Abstract
We present a method for determining the globally optimal on-line
learning rule for a soft committee machine under a statistical mechanics framework. This work complements previous results on
locally optimal rules, where only the rate of change in generalization error was considered. We maximize the total reduction in
generalization error over the whole learning process and show how
the resulting rule can significantly outperform the locally optimal
rule.
1
Introduction
We consider a learning scenario in which a feed-forward neural network model (the
student) emulates an unknown mapping (the teacher), given a set of training examples produced by the teacher. The performance of the student network is typically
measured by its generalization error, which is the expected error on an unseen example. The aim of training is to reduce the generalization error by adapting the
student network's parameters appropriately.
A common form of training is on-line learning, where training patterns are presented sequentially and independently to the network at each learning step. This
form of training can be beneficial in terms of both storage and computation time,
especially for large systems. A frequently used on-line training method for networks
with continuous nodes is that of stochastic gradient descent, since a differentiable
error measure can be defined in this case. The stochasticity is a consequence of
the training error being determined according to only the latest, randomly chosen, training example. This is to be contrasted with batch learning, where all the
training examples would be used to determine the training error leading to a deterministic algorithm. Finding an effective algorithm for discrete networks is less
straightforward as the error measure is not differentiable.
? rattraym@aston.ac.uk
t saadd@aston.ac.uk
Globally Optimal On-line Learning Rules
323
Often, it is possible to improve on the basic stochastic gradient descent algorithm
and a number of modifications have been suggested in the literature. At late times
one can use on-line estimates of second order information (the Hessian or its eigenvalues) to ensure asymptotically optimal performance (e.g., [1, 2]). A number of
heuristics also exist which attempt to improve performance during the transient
phase of learning (for a review, see [3]). However, these heuristics all require the
careful setting of parameters which can be critical to their performance. Moreover,
it would be desirable to have principled and theoretically well motivated algorithms
which do not rely on heuristic arguments.
Statistical mechanics allows a compact description for a number of on-line learning
scenarios in the limit of large input dimension, which we have recently employed to
propose a method for determining globally optimal learning rates for on-line gradient descent [4]. This method will be generalized here to determine globally optimal
on-line learning rules for both discrete and continuous machines. That is, rules
which provide the maximum reduction in generalization error over the whole learning process. This provides a natural extension to work on locally optimal learning
rules [5, 6], where only the rate of change in generalization error is optimized. In
fact, for simple systems we sometimes find that the locally optimal rule is also globally optimal. However, global optimization seems to be rather important in more
complex systems which are characterized by more degrees of freedom and often require broken permutation symmetries to learn perfectly. We will outline our general
formalism and consider two simple and tractable learning scenarios to demonstrate
the method.
It should be pointed out that the optimal rules derived here will often require knowledge of macroscopic properties related to the teacher's structure which would not
be known in general. In this sense these rules do not provide practical algorithms as
they stand, although some of the required macroscopic properties may be evaluated
or estimated on the basis of data gathered as the learning progresses. In any case
these rules provide an upper bound on the performance one could expect from a
real algorithm and may be instrumental in designing practical training algorithms.
2
The statistical mechanics framework
For calculating the optimal on-line learning rule we employ the statistical mechanics
description of the learning process. Under this framework, which may be employed
for both smooth [7,8] and discrete sy.stems (e.g. [9]), the learning process is captured
by a small number of self-averaging statistics whose trajectory is deterministic in the
limit of large input dimension. In this analysis the relevant statistics are overlaps
between weight vectors associated with different nodes of the student and teacher
networks. The equations of motion for the evolution of these overlaps can be written
in closed form and can be integrated numerically to describe the dynamics.
We will consider a general two-layer soft committee machine l . The desired teacher
mapping is from an N-dimensional input space
E RN onto a scalar ( E R,
which the student models through a map a(J,e) 2:~l g(Ji ?e), where g(x) is the
activation function for the hidden layer, J == {Jih<i<K is the set of input-to-hidden
adaptive weights for the K hidden nodes and the hidden-to-output weights are set
to 1. The activation of hidden node i under presentation of the input pattern eJ' is
= J i . eJ'?
denoted
e
=
xr
IThe general result presented here also applies to the discrete committee machine, but
we will limit our discussion to the soft-committee machine.
M. Rattray and D. Saad
324
Training examples are of the form (el', (I') where J.L = 1,2, ... , P. The components
of the independently drawn input vectors el' are uncorrelated random variables
with zero mean and unit variance. The corresponding output (I' is given by a deterministic teacher of a similar configuration to the student except for a possible
difference in the number M of hidden units and is of the form (I' = l:!1 g(B n . el'),
where B == {Bnh<n<M is the set of input-to-hidden adaptive weights. The activation of hidden node n under presentation of the input pattern el' is denoted
Y~ = Bn . el'. We will use indices i, j, k, I ... to refer to units in the student network and n, m, ... for units in the teacher network. We will use the commonly used
quadratic deviation E(J,e) == ~ [ a(J,e) - (]2, as the measure of disagreement between teacher and student. The most basic learning rule is to perform gradient
descent on this quantity. Performance on a typical input defines the generalization
error Eg(J) == (E(J,e?{(} through an average over all possible input vectors
e.
The general form of learning rule we will consider is,
J~+l
= J~ + ~FfJ(xl' ~I') el'
,
'N'
,." ~
(1)
where F == {Fi} depends only on the student activations and the teacher's output,
and not on the teacher activations which are unobservable. Note that gradient
descent on the error takes this general form, as does Hebbian learning and other
training algorithms commonly used in discrete machines. The optimal F can also
depend on the self-averaging statistics which describe the dynamics, since we know
how they evolve in time. Some of these would not be available in a practical application, although for some simple cases the unobservable statistics can be deduced
from observable quantities. This is therefore an idealization rather than a practical
algorithm and provides a bound on the performance of a real algorithm.
The activations are distributed according to a multivariate Gaussian with covariances: (XiXk) = Ji?Jk == Qik, (XiYn) = Ji?Bn == R in , and (YnYm) = Bn?B m == Tnm ,
measuring overlaps between student and teacher vectors. Angled brackets denote
averages over input patterns. The covariance matrix completely describes the state
of the system and in the limit of large N we can write equations of motion for each
macroscopic (the Tnm are fixed and define the teacher):
dRin
dO!
= (FiYn )
dQik
do:
= (Fixk + FkXi + FiFk)
,
(2)
where angled brackets now denote the averages over activations, replacing the averages over inputs, and 0: = J.LIN plays the role of a continuous time variable.
3
The globally optimal rule
Carrying out the averaging over input patterns one obtains an expression for the
generalization error which depends exclusively on the overlaps R,Q and T. Using
the dependence of their dynamics (Eq. 2) on F one can easily calculate the locally
optimal learning rule [5] by taking the functional derivative of dEg(F)/do: to zero,
looking for the rule that will maximize the reduction in generalization error at the
present time step. This approach has been shown to be successful in some training
scenarios but is likely to fail where the learning process is characterized by several
phases of a different natures (e.g., multilayer networks).
The globally optimal learning rule is found by minimizing the total change in generalization error over a fixed time window,
~fg(F)=
la
l
ao
dE
dgdo: =
0:
la
ao
l
?(F, 0:) do:.
(3)
Globally Optimal On-line Learning Rules
325
This is a functional of the learning rule which we minimize by a variational approach.
First we can rewrite the integrand by expanding in terms of the equations of motion,
each constrained by a Lagrange multiplier,
?(F
,a
) = ~ 8f g dRin
~ 8R . da
in
In
~ 8f g dQik
+~
8Q.
ik
Ik
+L
Vik
ik
da
(d~ik
~..\ . (d~n _ (F.-
+~
in
In
d
a
IYn
?)
- (FiXk + FkXi + FiFk?) .
(4)
a
The expression for ? still involve two multidimensional integrations over x and y,
so taking variations in F, which may depend on x and ( but not on y, we find an
expression for the optimal rule in terms of the Lagrange multipliers:
F
1 -1 _
= -x -"2v
.\y
(5)
where v = [Vij] and .\ = [..\in]. We define y to be the teacher's expected field given
the teacher's output and the student activations, which are observable quantities:
y
=/
dyyp(ylx,() .
(6)
Now taking variations in the overlaps w.r.t. the integral in Eq. (3) we find a set of
differential equations for the Lagrange multipliers1
d..\km
da
-
_ L,,\ . 8{FiYn) _ LV .. 8{Fi Xj + FjXi + FiFj)
. In 8R km
.. I}
8Rkm
In
IJ
(7)
where F takes its optimal value defined in Eq. (5). The boundary conditions for
the Lagrange multipliers are,
and
(8)
which are found by minimizing the rate of change in generalization error at ai, so
that the globally optimal solution reduces to the locally optimal solution at this
point, reflecting the fact that changes at al have no affect at other times.
IT the above expressions do not yield an explicit formula for the optimal rule then
the rule can be determined iteratively by gradient descent on the functional Llfg(F).
To determine all the quantities necessary for this procedure requires that we first
integrate the equations for the overlaps forward and then integrate the equations
for the Lagrange multipliers backwards from the boundary conditions in Eq. (8).
4
Two tractable examples
In order to apply the above results we must be able to carry out the average in
Eq. (6) and then in Eq. (7). These averages are also required to determine the
locally optimal learning rule, so that the present method can be extended to any
of the problems which have already been considered under the criteria of local
optimality. Here we present two examples where the averages can be computed
in closed form. The first problem we consider is a boolean perceptron learning a
M. Rattray and D. Saad
326
linearly separable task where we retrieve the locally optimal rule [5]. The second
problem is an over-realizable task, where a soft committee machine student learns
a perceptron with a sigmoidal response. In this example the globally optimal rule
significantly outperforms the locally optimal rule and exhibits a faster asymptotic
decay.
Boolean perceptron: For the boolean perceptron we choose the activation function g(x) = sgn(x) and both teacher and student have a single hidden node
(M = K = 1). The locally optimal rule was determined by Kinouchi and Caticha [5]
and they supply the expected teacher field given the teacher output ( sgn(y) and
the student field x (we take the teacher length T = 1 without loss of generality),
=
_ R (
y= x
Q
(~exp(-if-?)
+ ---'-----,,..-----:,erfc
(9)
(-5?)
Substituting this expression into the Lagrange multiplier dynamics in Eq. (7) shows
that the ratio of .x to v is given by .x/v = -2Q / R, and Eq. (5) then returns the
locally optimal value for the optimal rule:
F =
( fi.. exp( _ , 2 z 2 )
Vi
2.
,erfc (-~,)
(10)
This rule leads to modulated Hebbian learning and the resulting dynamics are
discussed in [5]. We also find that the locally optimal rule is retrieved when the
teacher is corrupted by output or weight noise [9].
Soft committee machine learning a continuous perceptron: In this example
the teacher is an invertible perceptron (M = 1) while the student is a soft committee
machine with an arbitrary number (K) of hidden nodes. We choose the activation
function g(x) = erf(x/v'2) for both the student and teacher since this allows the
generalization error to be determined in closed form [7]. This is an example of
an over-realizable task, since the student has greater complexity than is required
to learn the teacher's mapping. The locally optimal rule for this scenario was
determined recently [6].
Since the teacher is invertible, the expected teacher activation fi is trivially equal
to the true activation y. This leads to a particularly simple form for the dynamics
(the n suffix is dropped since there is only one teacher node),
dRi
dQik
do: = biT - Ri
do: = bibk T - Qik ,
(11)
where we have defined bi = - Ej vi/ .xj /2 and the optimal rule is given by
Fi = biy - Xi. The Lagrange multiplier dynamics in Eq. (7) then show that the
relative ratios of each Lagrange multiplier remain fixed over time, so that bi is determined by its boundary value (see Eq. (8?. It is straightforward to find solutions
for long times, since the bi approach limiting values for very small generalization
error (there are a number of possible solutions because of symmetries in the problem
but any such solution will have the same performance for long times). For example,
one possible solution is to have b1 = 1 and bi = 0 for all i f:. 1, which leads to an
exponential decay of weights associated with all but a single node. This shows how
the optimal performance is achieved when the complexity of the student matches
that of the teacher.
Figure 1 shows results for a three node student learning a continuous perceptron.
Clearly, the locally optimal rule performs poorly in comparison to the globally
Globally Optimal On-line Learning Rules
327
100
--.__ .-
10-2
?g
0.8
.....
'" ....
10-4
~n
'"
.... ....
.........
0.6
0.4
("
.
10'"
....
i ,.
0.2
.........
0
"
,
'.,
-0.2
10'"
-0.4
, ....
-0.6
10- '0
"
-0.8
10- '2
0
5
10
15
20
25
a
-1
0
5
10
15
20
25
a
Figure 1: A three node soft committee machine student learns from an continuous percept ron teacher. The figure on the left shows a log plot of the generalization error for the globally optimal (solid line) and locally optimal (dashed line)
algorithms. The figure on the right shows the student-teacher overlaps for the
locally optimal rule, which exhibit a symmetric plateau before specialization occurs. The overlaps where initialized randomly and uniformly with Qii E [0,0.5] and
~,Qi*j E [0,10- 6 ].
optimal rule. In this example the globally optimal r.ule arrived at was one in which
two nodes became correlated with the teacher while a third became anti-correlated,
showing another possible variation on the optimal rule (we determined this rule
iteratively by gradient descent in order to justify our general approach, although
the observations above show how one can predict the final result for long times).
The locally optimal rule gets caught in a symmetric plateau, characterized by a lack
of differentiation between student vectors associated with different nodes, and also
displays a slower asymptotic decay.
5
Conclusion and future work
We have presented a method for determining the optimal on-line learning rule for
a soft committee machine under a statistical mechanics framework. This result
complements previous work on locally optimal rules which sought only to optimize
the rate of change in generalization error. In this work we considered the global
optimization problem of minimizing the total change in generalization error over
the whole learning process. We gave two simple examples for which the rule could
be determined in closed form, for one of which, an over-realizable learning scenario,
it was shown how the locally optimal rule performed poorly in comparison to the
globally optimal rule. It is expected that more involved systems will show even
greater difference in performance between local and global optimization and we
are currently applying the method to more general teacher mappings. The main
technical difficulty is in computing the expected teacher activation in Eq. (6) and
this may require the use of approximate methods in some cases.
It would be interesting to compare the training dynamics obtained by the globally
optimal rules to other approaches, heuristic and principled, aimed at incorporating
information about the curvature of the error surface into the parameter modification
rule. In particular we would like to examine rules which are known to be optimal
asymptotically (e.g. [10]). Another important issue is whether one can apply these
results to facilitate the design of a practical learning algorithm.
328
M. Rattray and D. Saad
Acknowledgement This work was supported by the EPSRC grant GR/L19232.
References
[1] G. B. Orr and T. K. Leen in Advances in Neural Information Processing Systems, vol 9, eds M. C. Mozer, M. I. Jordan and T. Petsche (MIT Press, Cambridge MA, 1997) p 606.
[2] Y. LeCun, P. Y. Simard and B. Pearlmutter in Advances in Neural Information
Processing Systems, vol 5, eds S. J. Hanson, J. D. Cowan and C. 1. Giles
(Morgan Kaufman, San Mateo, CA, 1993) P 156.
[3] C. M. Bishop, Neural networks for pattern recognition, (Oxford University
Press, Oxford, 1995).
[4] D. Saad and M. Rattray, Phys. Rev. Lett. 79, 2578 (1997).
[5] O. Kinouchi and N. Caticha J. Phys. A 25, 6243 (1992).
[6] R. Vicente and N. Caticha J. Phys. A 30, L599 (1997).
[7] D. Saad and S. A. Solla, Phys. Rev. Lett. 74, 4337 (1995) and Phys. Rev. E
52 4225 {1995}.
[8] M. Biehl and H. Schwarze, J. Phys. A 28, 643 (1995).
[9] M. Biehl, P. Riegler and M. Stechert, Phys. Rev. E 52, R4624 (1995).
[10] S. Amari in Advances in Neural Information Processing Systems, vol 9, eds M.
C. Mozer, M. I. Jordan and T. Petsche (MIT Press, Cambridge MA, 1997).
| 1373 |@word seems:1 instrumental:1 km:2 bn:3 covariance:2 solid:1 carry:1 reduction:3 configuration:1 exclusively:1 outperforms:1 activation:13 written:1 must:1 plot:1 provides:2 node:13 ron:1 sigmoidal:1 differential:1 supply:1 ik:4 theoretically:1 expected:6 frequently:1 mechanic:5 examine:1 globally:17 window:1 moreover:1 kaufman:1 finding:1 differentiation:1 multidimensional:1 uk:3 unit:4 grant:1 before:1 dropped:1 local:2 limit:4 consequence:1 oxford:2 riegler:1 mateo:1 qii:1 bi:4 practical:5 lecun:1 xr:1 procedure:1 significantly:2 adapting:1 get:1 onto:1 storage:1 applying:1 optimize:1 deterministic:3 map:1 latest:1 straightforward:2 independently:2 caught:1 rule:50 retrieve:1 variation:3 limiting:1 play:1 designing:1 recognition:1 jk:1 particularly:1 role:1 epsrc:1 calculate:1 solla:1 principled:2 mozer:2 broken:1 complexity:2 dynamic:8 depend:2 carrying:1 rewrite:1 ithe:1 basis:1 completely:1 easily:1 effective:1 describe:2 whose:1 heuristic:4 biehl:2 amari:1 statistic:4 erf:1 unseen:1 final:1 differentiable:2 eigenvalue:1 propose:1 relevant:1 poorly:2 description:2 ac:2 measured:1 ij:1 progress:1 eq:11 stochastic:2 sgn:2 transient:1 require:4 ao:2 generalization:16 extension:1 considered:3 magnus:1 exp:2 mapping:4 predict:1 substituting:1 sought:1 angled:2 birmingham:1 currently:1 mit:2 clearly:1 gaussian:1 aim:1 rather:2 ej:3 jih:1 derived:1 sense:1 realizable:3 el:6 suffix:1 typically:1 integrated:1 hidden:10 unobservable:2 issue:1 denoted:2 constrained:1 integration:1 field:3 equal:1 future:1 employ:1 randomly:2 stechert:1 phase:2 attempt:1 freedom:1 bracket:2 integral:1 necessary:1 initialized:1 desired:1 formalism:1 soft:8 boolean:3 giles:1 measuring:1 deviation:1 successful:1 gr:1 teacher:31 corrupted:1 deduced:1 invertible:2 iyn:1 choose:2 derivative:1 leading:1 return:1 simard:1 de:1 orr:1 student:22 depends:2 vi:2 performed:1 closed:4 minimize:1 became:2 variance:1 emulates:1 percept:1 sy:1 gathered:1 yield:1 produced:1 trajectory:1 plateau:2 phys:7 ed:3 involved:1 associated:3 knowledge:1 reflecting:1 feed:1 response:1 leen:1 evaluated:1 generality:1 replacing:1 lack:1 defines:1 schwarze:1 facilitate:1 multiplier:7 true:1 evolution:1 symmetric:2 iteratively:2 eg:1 kinouchi:2 during:1 self:2 criterion:1 generalized:1 arrived:1 outline:1 demonstrate:1 pearlmutter:1 performs:1 motion:3 variational:1 recently:2 fi:5 common:1 functional:3 ji:3 b4:1 discussed:1 numerically:1 refer:1 cambridge:2 ai:1 trivially:1 mathematics:1 pointed:1 stochasticity:1 surface:1 curvature:1 multivariate:1 retrieved:1 scenario:6 captured:1 morgan:1 greater:2 employed:2 determine:4 maximize:2 dashed:1 desirable:1 reduces:1 stem:1 hebbian:2 smooth:1 technical:1 faster:1 characterized:3 match:1 long:3 lin:1 qi:1 basic:2 multilayer:1 sometimes:1 qik:2 achieved:1 macroscopic:3 appropriately:1 saad:6 dri:1 cowan:1 jordan:2 backwards:1 xj:2 affect:1 gave:1 perfectly:1 reduce:1 vik:1 whether:1 motivated:1 expression:5 specialization:1 hessian:1 involve:1 ylx:1 aimed:1 locally:19 outperform:1 exist:1 estimated:1 rattray:5 discrete:5 write:1 vol:3 drawn:1 asymptotically:2 idealization:1 saadd:1 bit:1 layer:2 bound:2 display:1 quadratic:1 ri:1 integrand:1 argument:1 optimality:1 separable:1 department:1 according:2 beneficial:1 describes:1 remain:1 rev:4 modification:2 equation:6 committee:9 fail:1 know:1 tractable:2 available:1 apply:2 disagreement:1 petsche:2 batch:1 slower:1 ensure:1 calculating:1 especially:1 erfc:2 already:1 quantity:4 occurs:1 dependence:1 exhibit:2 gradient:7 length:1 index:1 ratio:2 minimizing:3 design:1 unknown:1 perform:1 upper:1 observation:1 descent:7 anti:1 extended:1 looking:1 rn:1 arbitrary:1 david:1 complement:2 required:3 optimized:1 rkm:1 hanson:1 able:1 suggested:1 pattern:6 critical:1 overlap:8 natural:1 rely:1 difficulty:1 improve:2 aston:3 biy:1 review:1 literature:1 acknowledgement:1 evolve:1 determining:3 asymptotic:2 relative:1 loss:1 expect:1 permutation:1 interesting:1 lv:1 integrate:2 degree:1 vij:1 uncorrelated:1 supported:1 perceptron:7 taking:3 fg:1 distributed:1 boundary:3 dimension:2 lett:2 stand:1 forward:2 commonly:2 adaptive:2 san:1 caticha:3 approximate:1 compact:1 observable:2 obtains:1 deg:1 global:3 sequentially:1 b1:1 xi:1 continuous:6 learn:2 nature:1 expanding:1 ca:1 symmetry:2 complex:1 da:3 main:1 linearly:1 whole:3 noise:1 ule:1 explicit:1 exponential:1 xl:1 late:1 third:1 learns:2 xiyn:1 formula:1 tnm:2 bishop:1 showing:1 decay:3 incorporating:1 likely:1 lagrange:8 ffj:1 scalar:1 applies:1 ma:2 presentation:2 careful:1 change:7 vicente:1 determined:8 except:1 contrasted:1 typical:1 averaging:3 uniformly:1 justify:1 total:3 la:2 modulated:1 correlated:2 |
410 | 1,374 | Just One View:
Invariances in Inferotemporal Cell Thning
Maximilian Riesenhuber
Tomaso Poggio
Center for Biological and Computational Learning and
Department of Brain and Cognitive Sciences
Massachusetts Institute of Techno)ogy, E25-201
Cambridge, MA 02139
{max,tp }@ai.mit.edu
Abstract
In macaque inferotemporal cortex (IT), neurons have been found to respond selectively to complex shapes while showing broad tuning ("invariance") with respect to stimulus transformations such as translation
and scale changes and a limited tuning to rotation in depth. Training
monkeys with novel, paperclip-like objects, Logothetis et al. 9 could investigate whether these invariance properties are due to experience with
exhaustively many transformed instances of an object or if there are mechanisms that allow the cells to show response invariance also to previously
unseen instances of that object. They found object-selective cells in anterior IT which exhibited limited invariance to various transformations
after training with single object views. While previous models accounted
for the tuning of the cells for rotations in depth and for their selectivity to a specific object relative to a population of distractor objects,14,1
the model described here attempts to explain in a biologically plausible
way the additional properties of translation and size invariance. Using
the same stimuli as in the experiment, we find that model IT neurons
exhibit invariance properties which closely parallel those of real neurons.
Simulations show that the model is capable of unsupervised learning of
view-tuned neurons.
We thank Peter Dayan, Marcus Dill, Shimon Edelman, Nikos Logothetis, Jonathan Mumick and
Randy O'Reilly for useful discussions and comments.
216
M. RiesenhuberandT. Poggio
1 Introduction
Neurons in macaque inferotemporal cortex (IT) have been shown to respond to views of
complex objects,8 such as faces or body parts, even when the retinal image undergoes size
changes over several octaves, is translated by several degrees of visual angle 7 or rotated in
depth by a certain amount 9 (see [13] for a review).
These findings have prompted researchers to investigate the physiological mechanisms
underlying these tuning properties. The original model 14 that led to the physiological
experiments of Logothetis et al. 9 explains the behavioral view invariance for rotation in
depth through the learning and memory of a few example views, each represented by a
neuron tuned to that view. Invariant recognition for translation and scale transformations
have been explained either as a result of object-specific learning4 or as a result of a
normalization procedure ("shifter") that is applied to any image and hence requires only
one object-view for recognition. 12
A problem with previous experiments has been that they did not illuminate the mechanism
underlying invariance since they employed objects (e.g., faces) with which the monkey was
quite familiar, having seen them numerous times under various transformations. Recent
experiments by Logothetis et al. 9 addressed this question by training monkeys to recognize novel objects ("paperclips" and amoeba-like objects) with which the monkey had no
previous visual experience. After training, responses of IT cells to transformed versions of
the training stimuli and to distractors of the same type were collected. Since the views the
monkeys were exposed to during training were tightly controlled, the paradigm allowed to
estimate the degree of invariance that can be extracted from just one object view.
In partiCUlar, Logothetis et al. 9 tested the cells' responses to rotations in depth, translation
and size changes. Defining "in variance" as yielding a higher response to test views than
to distractor objects, they report 9 ,10 an average rotation invariance over 30?, translation
invariance over ?2?, and size invariance of up to ?1 octave around the training view.
These results establish that there are cells showing some degree of invariance even after
training with just one object view, thereby arguing against a completely learning-dependent
mechanisms that requires visual experience with each transformed instance that is to be
recognized. On the other hand, invariance is far from perfect but rather centered around the
object views seen during training.
2
The Model
Studies of the visual areas in the ventral stream of the macaque visual system8 show a
tendency for cells higher up in the pathway (from VI over V2 and V4 to anterior and
posterior IT) to respond to increasingly complex objects and to show increasing invariance
to transformations such as translations, size changes or rotation in depth. 13
We tried to construct a model that explains the receptive field properties found in the
experiment based on a simple feedforward model. Figure 1 shows a cartoon of the model:
A retinal input pattern leads to excitation of a set of "VI" cells, in the figure abstracted
as having derivative-of-Gaussian receptive field profiles. These "VI" cells are tuned to
simple features and have relatively small receptive fields. While they could be cells from a
variety of areas, e.g., VI or V2 (cf. Discussion), for simplicity, we label them as "VI" cells
(see figure). Different cells differ in preferred feature, e.g., orientation, preferred spatial
frequency (scale), and receptive field location. "VI" cells of the same type (i.e., having
the same preferred stimulus, but of different preferred scale and receptive field location)
feed into the same neuron in an intermediate layer. These intermediate neurons could be
complex cells in VI or V2 or V4 or even posterior IT: we label them as "V4" cells, in the
Just One View: Invariances in Inferotemporal Cell Tuning
217
same spirit in which we labeled the neurons feeding into them as "VI" units. Thus, a "V4"
cell receives inputs from "VI" cells over a large area and different spatial scales ([8] reports
an average receptive field size in V4 of 4.4? of visual angle, as opposed to about I ? in VI;
for spatial frequency tuning, [3] report an average FWHM of 2.2 octaves, compared to 1.4
(foveally) to 1.8 octaves (parafoveally) in VI 5). These "V4" cells in turn feed into a layer
of "IT" neurons, whose invariance properties are to be compared with the experimentally
observed ones.
Retina
Figure 1: Cartoon of the model. See text for explanation.
A crucial element of the model is the mechanism an intermediate neuron uses to pool the
activities of its afferents. From the computational point of view, the intermediate neurons
should be robust feature detectors, i.e., measure the presence of specific features without
being confused by clutter and context in the receptive field. More detailed considerations
(Riesenhuber and Poggio, in preparation) show that this cannot be achieved with a response
function that just summates over all the afferents (cf. Results). Instead, intermediate
neurons in our model perform a "max" operation (akin to a "Winner-Take-AII") over all
their afferents, i.e., the response ofan intermediate neuron is determined by its most strongly
excited afferent. This hypothesis appears to be compatible with recent data,15 that show
that when two stimuli (gratings of different contrast and orientation) are brought into the
recepti ve field of a V4 cell, the cell's response tends to be close to the stronger of the two
individual responses (instead of e.g., the sum as in a linear model).
218
M. Riesenhuber and T. Poggio
Thus, the response function
0i
of an intennediate neuron i to stimulation with an image v
IS
(1)
with Ai the set of afferents to neuron i, aU) the receptive field center of afferent j, v a(j) the
(square-nonnalized) image patch centered at aU) that corresponds in size to the receptive
field, ~j (also square-nonnalized) of afferent j and "." the dot product operation.
Studies have shown that V 4 neurons respond to features of "intennediate" complexity
such as gratings, corners and crosses. 8 In V4 the receptive fields are comparatively large
(4.4 0 of visual angle on average8 ), while the preferred stimuli are usually much smaller. 3
Interestingly, cells respond independently of the location of the stimulus within the receptive
field. Moreover, average V4 receptive field size is comparable to the range of translation
invariance of IT cells (:S ?2?) observed in the experiment. 9 For afferent receptive fields
~j, we chose features similar to the ones found for V 4 cells in the visual system: 8 bars
(modeled as second derivatives of Gaussians) in two orientations, and "corners" of four
different orientations and two different degrees of obtuseness. This yielded a total of lO
intennediate neurons. This set of features was chosen to give a compact and biologically
plausible representation. Each intennediate cell received input from cells with the same
type of preferred stimulus densely covering the visual field of 256 x 256 pixels (which thus
would correspond to about 4.40 of visual angle, the average receptive field size in V48 ),
with receptive field sizes of afferent cells ranging from 7 to 19 pixels in steps of 2 pixels.
The features used in this paper represent the first set of features tried, optimizing feature
shapes might further improve the model's performance.
The response tj oftop layer neuron j with connecting weights Wj to the intennediate layer
was set to be a Gaussian, centered on Wj,
tj
= ~exp (_IIO;~jI12)
271'0'2
(2)
0'
?
where is the excitation of the intennediate layer and 0' the variance of the Gaussian, which
was chosen based on the distribution of responses (for section 3.1) or learned (for section
3.2).
The stimulus images were views of 21 randomly generated "paperclips" of the type used in
the physiology experiment. 9 Distractors were 60 other paperclip images generated by the
same method. Training size was 128 x 128 pixels.
3
3.1
Results
Invariance of Representation
In a first set of simulations we investigated whether the proposed model could indeed
account for the observed invariance properties. Here we assumed that connection strengths
from the intennediate layer cells to the top layer had already been learned by a separate
process, allowing us to focus on the tolerance of the representation to the above-mentioned
transformations and on the selectivity of the top layer cells.
To {,?,tablish the tuning properties of view-tuned model neurons, the connections Wj between
the intermediate layer and top layer unit j were set to be equal to the excitation 0training in
the intermediate layer caused by the training view. Figure 2 shows the "tuning curve" for
rotation in depth and Fig. 3 the response to changes in stimulus size of one such neuron . The
neuron shows rotation invariance (i.e., producing a higher response than to any distractor)
over about 44 0 and invariance to scale changes over the whole range tested. For translation
Just One View: Invariances in InJerotemporal Cell Tuning
219
1~----------------Q.J
U'J
oc: O.
0.5
Q..
til
~
60
80
100
120
angle
20
40
60
distractor
Figure 2: Responses of a sample top layer neuron to different views of the training stimulus and to distractors.
The left plot shows the rotation tuning curve, with the training view (900 view) shown in the middle image over
the plot. The neighboring images show the views of the paperclip at the borders of the rotation tuning curve,
which are located where the response to the rotated clip falls below the response to the best distractor (shown in
the plot on the right). The neuron exhibits broad rotation tuning over more than 40? .
(not shown), the neuron showed invariance over translations of ?96 pixels around the center
in any direction, corresponding to ? 1.7? of visual angle.
The average invariance ranges for the 21 tested paperclips were 35? of rotation angle, 2.9
octaves of scale invariance and ? 1.8 0 of translation invariance. Comparing this to the
experimentally observed 10 30 0 ,2 octaves and ?2?, resp., shows a very good agreement of
the invariance properties of model and experimental neurons.
3.2
Learning
In the previous section we assumed that the connections from the intermediate layer to a
view-tuned neuron in the top layer were pre-set to appropriate values. In this section, we
investigate whether the system allows unsupervised learning of view-tuned neurons.
Since biological plausibility of the learning algorithm was not our primary focus here,
we chose a general, rather abstract learning algorithm, viz. a mixture of Gaussians model
trained with the EM algorithm. Our model had four neurons in the top level, the stimuli were
views of four paperclips, randomly selected from the 21 paperclips used in the previous
experiments. For each clip, the stimulus set contained views from 17 different viewpoints,
spanning 34 0 of viewpoint change. Also, each clip was included at 11 different scales in
the stimulus set, covering a range of two octaves of scale change.
=
Connections Wi and variances O'i, i
1, ... ,4, were initialized to random values at the
beginning of training. After a few iterations of the EM algorithm (usually less than 30), a
stationary state was reached, in which each model neuron had become tuned to views of one
paperclip: For each paperclip, all rotated and scaled views were mapped to (i.e., activated
most strongly) the same model neuron and views of different paperclips were mapped to
different neurons. Hence, when the system is presented with multiple views of different
objects, receptive fields of top level neurons self-organize in such a way that different
neurons become tuned to different objects.
M. Riesenhuber and T. Poggio
220
1r-----------------~
0.8
0.5
0.6
0.4
100
0.20
stimulus size
200
20
40
60
distractor
Figure 3: Responses of the same top layer neuron as in Fig. 2 to scale changes of the training stimulus and to
distractors. The left plot shows the size tuning curve, with the training size (128 x 128 pixels) shown in the
middle image over the plot. The neighboring images show scaled versions of the paperclip. Other elements as in
Fig. 2. The neuron exhibits scale invariance over more than 2 octaves.
4 Discussion
Object recognition is a difficult problem because objects must be recognized irrespective
of position, size, viewpoint and illumination. Computational models and engineering
implementations have shown that most of the required invariances can be obtained by
a relatively simple learning scheme, ba<;ed on a small set of example views. 14 ,17 Quite
sensibly, the visual system can also achieve some significant degree of scale and translation
invariance from just one view. Our simulations show that the maximum response function is
a key component in the performance ofthe model. Without it - i.e., implementing a direct
convolution of the filters with the input images and a subsequent summation - invariance
to rotation in depth and translation both decrease significantly. Most dramatically, however,
invariance to scale changes is abolished completely, due to the strong changes in afferent
cell activity with changing stimulus size. Taking the maximum over the afferents, as in our
model, always picks the best matching filter and hence produces a more stable response. We
expect a maximum mechanism to be essential for recognition-in-context, a more difficult
task and much more common than the recognition of isolated objects studied here and in
the related psychophysical and physiological experiments.
The recognition of a specific paperclip object is a difficult, subordinate level classification
task. It is interesting that our model sol ves it well and with a performance closely resembling
the physiological data on the same task. The model is a more biologically plausible and
complete model than previous ones14, 1 but it is still at the level of a plausibility proof rather
than a detailed physiological model. It suggests a maximum-like response of intermediate
cells as a key mechanism for explaining the properties of view-tuned IT cells, in addition
to view-based representations (already described in (1,9]).
Neurons in the intermediate layer currently use a very simple set of features . While this
appears to be adequate for the class of paperclip objects, more complex filters might be
necessary for more complex stimulus classes like faces. Consequently, future work will
aim to improve the filtering step ofthe model and to test it on more real world stimuli. One
can imagine a hierarchy of cell layers, similar to the "S" and "C" layers in Fukushima's
Just One View: Invariances in In!erotemporal Cell Tuning
221
Neocognitron,6 in which progressively more complex features are synthesized from simple
ones. The corner detectors in our model are likely candidates for such a scheme. We are
currently investigating the feasibility of such a hierarchy of feature detectors.
The demonstration that unsupervised learning of view-tuned neurons is possible in this
representation (which is not clear for related view-based models 14 , 1) shows that different
views of one object tend to form distinct clusters in the response space of intermediate
neurons. The current learning algorithm, however, is not very plausible, and more realistic
learning schemes have to be explored, as, for instance, in the attention-based model of
Riesenhuber and Dayan 16 which incorporated a learning mechanism using bottom-up and
top-down pathways. Combining the two approaches could also demonstrate how invariance
over a wide range of transformations can be learned from several example views, as in the
case of familiar stimuli. We also plan to simulate detailed physiological implementations
of several aspects of the model such as the maximum operation (for instance comparing
nonlinear dendritic interactions l l with recurrent excitation and inhibition). As it is, the
model can already be tested in additional physiological experiments, for instance involving
partial occlusions.
References
[1] Bricolo. E, Poggio, T & Logothetis, N (1997). 3D object recognition: A model of view-tuned
neurons. In Advances In Neural Information Processing 9,41-47. MIT Press.
[2] Biilthoff, H & Edelman, S (1992). Psychophysical support for a two-dimensional view interpolation theory of object recognition. Proc. Nat. Acad. Sci. USA 89, 60-64.
[3] Desimone, R & Schein, S (1987). Visual properties of neurons in area V4 of the macaque:
Sensitivity to stimulus fonn . 1. Neurophys. 57, 835-868.
[4] Foldiak, P (1991). Learning invariance from transfonnation sequences. Neural Computation 3,
194-200.
[5] Foster, KH, Gaska, JP, Nagler, M & Pollen, DA (1985). Spatial and temporal selectivity of
neurones in visual cortical areas VI and V2 of the macaque monkey. 1. Phy. 365,331-363 .
[6] Fukushima, K (1980). Neocognitron: A self-organizing neural network model for a mechanism
of pattern recognition unaffected by shift in position. Biological Cybernetics 36, 193-202.
[7] Ito, M, Tamura, H, Fujita, I & Tanaka, K (1995). Size and position invariance of neuronal
responses in monkey inferotemporal cortex. 1. Neurophys. 73, 218-226.
[8] Kobatake, E & Tanaka, K (1995). Neuronal selectivities to complex object features in the ventral
visual pathway of the macaque cerebral cortex 1. Neurophys., 71, 856-867.
[9] Logothetis, NK, Pauls, J & Poggio, T (1995). Shape representation in the inferior temporal
cortex of monkeys. Current Biology, 5, 552-563.
[10] Nikos Logothetis, personal communication.
[11] Mel, BW, Ruderman, DL & Archie, KA (1997). Translation-invariant orientation tuning in visual
'complex' cells could derive from intradendritic computations. Manuscript in preparation.
[12] Olshausen, BA, Anderson, CH & Van Essen, DC (1993). A neurobiological model of visual
attention and invariant pattern recognition based on dynamic routing of infonnation. 1. Neurosci. 13,4700-4719.
[13] Perret, D & Oram, M (1993). Neurophysiology of shape processing. Image Vision Comput. 11,
317-333.
[14] Poggio, T & Edelman, S (1990). A Network that learns to recognize 3D objects. Nature 343,
263-266.
[15] Reynolds, JH & Desimone, R (1997). Attention and contrast have similar effects on competitive
interactions in macaque area V4. Soc. Neurosc. Abstr. 23,302.
[16] Riesenhuber, M & Dayan, P (1997). Neural models for part-whole hierarchies. In Advances In
Neural Information Processing 9, 17-23. MIT Press.
[17] Ullman, S (1996). High-level vision: Object recognition and visual cognition. MIT Press.
| 1374 |@word neurophysiology:1 version:2 middle:2 stronger:1 simulation:3 tried:2 excited:1 pick:1 fonn:1 thereby:1 phy:1 tuned:11 interestingly:1 perret:1 reynolds:1 current:2 comparing:2 anterior:2 neurophys:3 ka:1 must:1 subsequent:1 realistic:1 shape:4 plot:5 progressively:1 stationary:1 selected:1 beginning:1 location:3 direct:1 become:2 edelman:3 pathway:3 behavioral:1 indeed:1 tomaso:1 distractor:6 brain:1 increasing:1 confused:1 underlying:2 moreover:1 intennediate:7 monkey:8 finding:1 transformation:7 temporal:2 sensibly:1 scaled:2 unit:2 producing:1 organize:1 engineering:1 tends:1 acad:1 interpolation:1 might:2 chose:2 au:2 studied:1 suggests:1 limited:2 fwhm:1 range:5 arguing:1 procedure:1 area:6 physiology:1 significantly:1 reilly:1 matching:1 pre:1 cannot:1 close:1 context:2 ogy:1 center:3 resembling:1 attention:3 independently:1 simplicity:1 population:1 resp:1 imagine:1 logothetis:8 hierarchy:3 techno:1 us:1 hypothesis:1 agreement:1 element:2 paperclip:14 recognition:11 located:1 labeled:1 observed:4 bottom:1 wj:3 decrease:1 sol:1 mentioned:1 complexity:1 exhaustively:1 personal:1 dynamic:1 trained:1 exposed:1 completely:2 translated:1 aii:1 various:2 represented:1 distinct:1 quite:2 whose:1 plausible:4 unseen:1 sequence:1 interaction:2 product:1 neighboring:2 combining:1 organizing:1 achieve:1 kh:1 bricolo:1 cluster:1 abstr:1 produce:1 perfect:1 rotated:3 object:31 derive:1 recurrent:1 received:1 strong:1 soc:1 grating:2 differ:1 direction:1 closely:2 iio:1 filter:3 centered:3 routing:1 implementing:1 subordinate:1 explains:2 feeding:1 dill:1 biological:3 dendritic:1 summation:1 around:3 exp:1 cognition:1 ventral:2 proc:1 label:2 currently:2 infonnation:1 mit:4 brought:1 gaussian:3 always:1 aim:1 rather:3 focus:2 viz:1 contrast:2 learning4:1 dayan:3 dependent:1 transformed:3 selective:1 pixel:6 fujita:1 classification:1 orientation:5 plan:1 spatial:4 field:18 construct:1 equal:1 having:3 cartoon:2 biology:1 broad:2 unsupervised:3 future:1 report:3 stimulus:21 few:2 retina:1 randomly:2 recognize:2 tightly:1 ve:1 individual:1 densely:1 familiar:2 occlusion:1 bw:1 fukushima:2 attempt:1 investigate:3 essen:1 mixture:1 yielding:1 activated:1 tj:2 desimone:2 capable:1 partial:1 necessary:1 experience:3 poggio:8 initialized:1 isolated:1 schein:1 instance:6 tp:1 archie:1 sensitivity:1 v4:11 pool:1 connecting:1 e25:1 opposed:1 cognitive:1 corner:3 derivative:2 til:1 ullman:1 account:1 retinal:2 caused:1 afferent:11 vi:12 stream:1 view:43 reached:1 competitive:1 parallel:1 square:2 variance:3 correspond:1 ofthe:2 intradendritic:1 researcher:1 cybernetics:1 unaffected:1 explain:1 detector:3 ed:1 against:1 frequency:2 proof:1 oram:1 massachusetts:1 distractors:4 appears:2 feed:2 manuscript:1 higher:3 response:22 strongly:2 anderson:1 just:8 hand:1 receives:1 ruderman:1 nonlinear:1 undergoes:1 olshausen:1 usa:1 effect:1 hence:3 during:2 self:2 inferior:1 covering:2 excitation:4 mel:1 oc:1 nagler:1 octave:8 neocognitron:2 complete:1 demonstrate:1 image:12 ranging:1 consideration:1 novel:2 common:1 rotation:13 stimulation:1 winner:1 jp:1 cerebral:1 synthesized:1 significant:1 cambridge:1 ai:2 tuning:15 had:4 dot:1 stable:1 cortex:5 inhibition:1 inferotemporal:5 posterior:2 recent:2 showed:1 foldiak:1 optimizing:1 selectivity:4 randy:1 certain:1 seen:2 additional:2 nikos:2 employed:1 recognized:2 paradigm:1 multiple:1 plausibility:2 cross:1 controlled:1 feasibility:1 involving:1 vision:2 iteration:1 represent:1 normalization:1 achieved:1 cell:37 tamura:1 addition:1 addressed:1 crucial:1 exhibited:1 comment:1 tend:1 transfonnation:1 spirit:1 presence:1 feedforward:1 intermediate:12 variety:1 shift:1 whether:3 akin:1 f:1 peter:1 neurones:1 adequate:1 dramatically:1 useful:1 detailed:3 clear:1 amount:1 clutter:1 clip:3 key:2 four:3 changing:1 sum:1 angle:7 respond:5 patch:1 comparable:1 layer:18 yielded:1 activity:2 strength:1 aspect:1 simulate:1 relatively:2 department:1 smaller:1 increasingly:1 em:2 wi:1 biologically:3 explained:1 invariant:3 previously:1 turn:1 mechanism:9 operation:3 gaussians:2 v2:4 appropriate:1 original:1 top:9 cf:2 neurosc:1 establish:1 comparatively:1 psychophysical:2 question:1 already:3 receptive:16 primary:1 illuminate:1 exhibit:3 thank:1 separate:1 mapped:2 sci:1 collected:1 spanning:1 marcus:1 shifter:1 modeled:1 prompted:1 demonstration:1 difficult:3 ba:2 implementation:2 perform:1 allowing:1 neuron:41 convolution:1 riesenhuber:6 defining:1 nonnalized:2 incorporated:1 communication:1 dc:1 required:1 connection:4 pollen:1 learned:3 amoeba:1 tanaka:2 macaque:7 bar:1 usually:2 pattern:3 below:1 max:2 memory:1 explanation:1 scheme:3 improve:2 numerous:1 irrespective:1 text:1 review:1 relative:1 expect:1 abolished:1 interesting:1 filtering:1 degree:5 foster:1 viewpoint:3 translation:13 lo:1 compatible:1 accounted:1 allow:1 jh:1 institute:1 fall:1 explaining:1 face:3 taking:1 wide:1 tolerance:1 van:1 curve:4 depth:8 cortical:1 world:1 far:1 biilthoff:1 compact:1 preferred:6 neurobiological:1 abstracted:1 investigating:1 assumed:2 nature:1 robust:1 investigated:1 complex:9 da:1 did:1 neurosci:1 whole:2 border:1 paul:1 profile:1 allowed:1 body:1 neuronal:2 fig:3 position:3 comput:1 candidate:1 ito:1 learns:1 shimon:1 down:1 specific:4 showing:2 explored:1 physiological:7 dl:1 essential:1 nat:1 illumination:1 maximilian:1 nk:1 led:1 likely:1 visual:18 contained:1 ch:1 corresponds:1 extracted:1 ma:1 consequently:1 change:11 experimentally:2 included:1 determined:1 total:1 invariance:38 tendency:1 experimental:1 selectively:1 support:1 jonathan:1 preparation:2 tested:4 |
411 | 1,375 | Classification by Pairwise Coupling
*
Stanford University
and
TREVOR HASTIE
ROBERT TIBSHIRANI t
University of Toronto
Abstract
We discuss a strategy for polychotomous classification that involves
estimating class probabilities for each pair of classes, and then coupling the estimates together. The coupling model is similar to the
Bradley-Terry method for paired comparisons. We study the nature of the class probability estimates that arise, and examine the
performance of the procedure in simulated datasets. The classifiers
used include linear discriminants and nearest neighbors: application to support vector machines is also briefly described.
1
Introduction
We consider the discrimination problem with J{ classes and N training observations.
The training observations consist of predictor measurements x = (Xl, X2, ... Xp) on
p predictors and the known class memberships. Our goal is to predict the class
membership of an observation with predictor vector Xo
Typically J{ -class classification rules tend to be easier to learn for J{ = 2 than for
f{ > 2 - only one decision boundary requires attention. Friedman (1996) suggested
the following approach for the the K-class problem: solve each of the two-class
problems, and then for a test observation, combine all the pairwise decisions to
form a J{ -class decision. Friedman's combination rule is quite intuitive: assign to
the class that wins the most pairwise comparisons.
Department of Statistics, Stanford University, Stanford California 94305;
trevor@playfair.stanford.edu
bepartment of Preventive Medicine and Biostatistics, and Department of Statistics;
tibs@utstat.toronto.edu
508
T. Hastie and R. Tibshirani
Friedman points out that this rule is equivalent to the Bayes rule when the class
posterior probabilities Pi (at the test point) are known :
argm~[Pd
= argma~[LI(pd(Pi + Pj) > Pj/(Pi + Pj?]
Jti
Note that Friedman's rule requires only an estimate of each pairwise decision. Many
(pairwise) classifiers provide not only a rule, but estimated class probabilities as well.
In this paper we argue that one can improve on Friedman's procedure by combining
the pairwise class probability estimates into a joint probability estimate for all J{
classes.
This leads us to consider the following problem. Given a set of events AI, A 2 , ... A.K,
some experts give us pairwise probabilities rij = Prob(AilA or Aj) . Is there a set
of probabilities Pi = Prob(Ai) that are compatible with the 1'ij?
In an exact sense, the answer is no. Since Prob( A d A i or Aj) = Pj /(Pi + pj) and
2: Pi = 1, we are requiring that J{ -1 free parameters satisfy J{ (/{ -1) /2 constraints
and , this will not have a solution in general. For example, if the 1'ij are the ijth
entries in the matrix
.
( 0.1
0.6
0.9 0.4)
. 0.7
0.3
.
then they are not compatible with any pi's. This is clear since
but also r31 > .5.
(1)
r12
> .5 and 1'23 > .5,
The model Prob(A i IAi or Aj) = Pj /(Pi + pj) forms the basis for the BradleyTerry model for paired comparisons (Bradley & Terry 1952) . In this paper we
fit this model by minimizing a Kullback- Leibler distance criterion to find the best
approximation foij = pd('Pi + pj) to a given set of 1'il's. We carry this out at each
predictor value x, and use the estimated probabilities to predict class membership
at x.
In the example above, the solution is p = (0.47, 0.25, 0.28). This solution makes
qualitative sense since event Al "beats" A 2 by a larger margin than the winner of
any of the other pairwise matches.
Figure 1 shows an example of these procedures in action . There are 600 data points
in three classes, each class generated from a mixture of Gaussians. A linear discriminant model was fit to each pair of classes, giving pairwise probability estimates
1'ij at each x. The first panel shows Friedman's procedure applied to the pairwise
rules. The shaded regions are areas of indecision , where each class wins one vote.
The coupling procedure described in the next section was then applied , giving class
probability estimates f>(x) at each x. The decision boundaries resulting from these
probabilities are shown in the second panel. The procedure has done a reasonable
job of resolving the confusion, in this case producing decision boundaries similar
to the three-class LDA boundaries shown in panel 3. The numbers in parentheses
above the plots are test-error rates based on a large test sample from the same population. Notice that despite the indeterminacy, the max-wins procedure performs
no worse than the coupling procedure. and both perform better than LDA. Later
we show an example where the coupling procedure does substantially better than
max-wms.
Classification by Pairwise Coupling
509
Pairwise LDA + Max (0.132)
Pairwise LOA + Coupling (0.136)
3?Class LOA (0.213)
Figure 1: A three class problem, with the data in each class generated from a mixture
of Gaussians. The first panel shows the maximum-win procedure. The second panel
shows the decision boundary from coupling of the pairwise linear discriminant rules
based on d in (6). The third panel shows the three-class LDA boundaries. Test-error
rates are shown in parentheses.
This paper is organized as follows. The coupling model and algorithm are given
in section 2. Pairwise threshold optimization, a key advantage of the pairwise
approach, is discussed in section 3. In that section we also examine the performance
of the various methods on some simulated problems, using both linear discriminant
and nearest neighbour rules. The final section contains some discussion .
2
Coupling the probabilities
Let the probabilities at feature vector x be p(x) = (PI (x) , ... PK (x)). In this section
we drop the argument x , since the calculations are done at each x separately.
\Ve assume that for each i -# j, there are nij observations in the training set and
from these we have estimated conditional probabilities Tij = Prob( iii or j).
Our model is
J..Lij
Binomial( nij , J-Lij )
Pi
Pi
+ Pj
(2)
or equivalently
log J-Lij
= log (Pi ) -
log (Pi
+ Pj),
(3)
a log-nonlinear model.
We wish to find Pi'S so that the Uij'S are close to the Tij'S. There are K - 1
independent parameters but K(I{ - 1)/2 equations, so it is not possible in general
to find .Pi's so that {iij = Tij for all i, j.
Therefore we must settle for {iij'S that are close to the observed Tij'S. Our closeness
criterion is the average (weighted) Kullback-Leibler distance between Tij and J-Lij :
(4)
T. Hastie and R. libshirani
510
and we find p to minimize this function .
This model and criterion is formally equivalent to the Bradley-Terry model for
preference data. One observes a proportion fij of nij preferences for item i, and
the sampling model is binomial, as in (2) . If each of the fij were independent, then
R(p) would be equivalent to the log-likelihood under this model. However our fij
are not independent as they share a common training set and were obtained from
a common set of classifiers. Furthermore the binomial models do not apply in this
case; the fij are evaluations of functions at a point, and the randomness arises in
the way these functions are constructed from the training data. We include the nij
as weights in (4); this is a crude way of accounting for the different precisions in
the pairwise probability estimates.
The score (gradient) equations are:
Lnijj1ij
=
Lnijfij;
jti
subject to
L Pi = 1.
(5)
i= 1,2 .... K
j#i
We use the following iterative procedure to compute the iN's:
Algorithm
1. Start with some guess for the
Pi,
and corresponding
Pij.
2. Repeat (i = 1,2, . .. , K, 1, ... ) until convergence:
renormalize the
Pi,
and recompute the
Pij.
The algorithm also appears in Bradley & Terry (1952). The updates in step 2 attempt to modify p so that the sufficient statistics match their expectation, but go
only part of the way. We prove in Hastie & Tibshirani (1996) that R(p) increases
at each step. Since R(p) is bounded above by zero, the procedure converges. At
convergence, the score equations are satisfied, and the PijS and p are consistent.
This algorithm is similar in flavour to the Iterative Proportional Scaling (IPS) procedure used in log-linear models. IPS has a long history, dating back to Deming &
Stephan (1940). Bishop, Fienberg & Holland (1975) give a modern treatment and
many references.
The resulting classification rule is
(6)
Figure 2 shows another example similar to Figure 1, where we can compare the
performance of the rules d and d. The hatched area in the top left panel is an
indeterminate region where there is more than one class achieving max(pd. In the
top right panel the coupling procedure has resolved this indeterminacy in favor of
class 1 by weighting the various probabilities. See the figure caption for a description
of the bottom panels.
Classification by Pairwise Coupling
511
PallWlse LOA + Max (0.449)
Pairwise LOA + Coupling (0 358)
LOA (0.457)
aDA (0.334)
Figure 2: A three class problem similar to that in figure 1, with the data in each
class generated from a mixture of Gaussians. The first panel shows the maximumwins procedure d). The second panel shows the decision boundary from coupling of
the pairwise linear discriminant rules based on d in (6). The third panel shows the
three-class LDA boundaries, and the fourth the QDA boundaries. The numbers in
the captions are the error rates based on a large test set from the same population.
3
Pairwise threshold optimization
As pointed out by Friedman (1996), approaching the classification problem in a
pairwise fashion allows one to optimize the classifier in a way that would be computationally burdensome for a J< -class classifier . Here we discuss optimization of
the classification threshold.
=
For each two class problem, let logit Pij(X) dij(x). Normally we would classify to
class i if d ij (x) > O. Suppose we find that d ij (x) > tij is better. Then we define
dij (x) = d ij (x) - tij, and hence pij (x) = logiC 1 di/x). We do this for all pairs, and
then apply the coupling algorithm to the P~j (x) to obtain probabilities pi( x) . In this
way we can optimize over J?J< - 1)/2 parameters separately, rather than optimize
jointly over J< parameters. With nearest neigbours, there are other approaches to
threshold optimization, that bias the class probability estimates in different ways.
See Hastie & Tibshirani (1996) for details. An example of the benefit of threshofd
optimization is given next.
Example: ten Gaussian classes with unequal covariance
In this simulated example taken from Friedman (1996), there are 10 Gaussian classes
in 20 dimensions. The mean vectors of each class were chosen as 20 independent
uniform [0,1] random variables . The covariance matrices are constructed from
eigenvectors whose square roots are uniformly distributed on the 20-dimensional
unit sphere (subject to being mutually orthogonal) , and eigenvalues uniform on
[0.01,1.01]. There are 100 observations per class in the training set, and 200 per
T. Hastie and R ..1ibshirani
512
class in the test set. The optimal decision boundaries in this problem are quadratic,
and neither linear nor nearest-neighbor methods are well-suited. Friedman states
that the Bayes error rate is less than 1%.
Figure 3 shows the test error rates for linear discriminant analysis, J -nearest neighbor and their paired versions using threshold optimization. We see that the coupled
classifiers nearly halve the error rates in each case. In addition, the coupled rule
works a little better than Friedman's max rule in each task. Friedman (1996) reports a median test error rate of about. 16% for his thresholded version of pairwise
nearest neighbor.
Why does the pairwise t.hresholding work in this example? We looked more closely
at the pairwise nearest neighbour rules rules that were constructed for this problem.
The thresholding biased the pairwise distances by about 7% on average. The average
number of nearest neighbours used per class was 4.47 (.122), while t.he standard Jnearest neighbour approach used 6.70 (.590) neighbours for all ten classes. For all
ten classes, the 4.47 translates into 44.7 neighbours. Hence relative to t.he standard
J - NN rule, the pairwise rule, in using the threshold optimization to reduce bias, is
able to use about six times as many near neighbours.
It)
C\I
ci
o
C\I
ci
T,
, II
1
o
:
c....l..."
ci
I
!
I
1
T
I
,
1.'
c....l..."
J-nn
nn/max
nn/coup
Ida
Ida/max
Ida/coup
Figure 3: Test errors for 20 simulations of ten-class Gaussian example.
4
Discussion
Due to lack of space, there are a number of issues that we did not discuss here. In
Hastie & Tibshirani (1996), we show the relationship between the pairwise coupling
and the max-wins rule: specifically, if the classifiers return 0 or Is rather than
probabilities, the two rules give the same classification. We also apply the pairwise
coupling procedure to nearest neighbour and support vector machines. In the latter
case, this provides a natural way of extending support vector machines, which are
defined for two-class problems, to multi-class problems.
Classification by Pairwise Coupling
513
The pairwise procedures, both Friedman 's max-win and our coupling, are most
likely to offer improvements when additional optimization or efficiency gains are
possible in the simpler 2-class scenarios. In some situations they perform exactly
like the multiple class classifiers. Two examples are: a) each of the pairwise rules
are based on QDA: i.e. each class modelled by a Gaussian distribution with separate covariances, and then the rijS derived from Bayes rule; b) a generalization
of the above, where the density in each class is modelled in some fashion, perhaps
nonparametrically via density estimates or near-neighbor methods, and then the
density estimates are used in Bayes rule.
Pairwise LDA followed by coupling seems to offer a nice compromise between LDA
and QDA, although the decision boundaries are no longer linear. For this special
case one might derive a different coupling procedure globally on the logit scale ,
which would guarantee linear decision boundaries. Work of this nature is currently
in progress with Jerry Friedman.
Acknowledgments
We thank Jerry Friedman for sharing a preprint of his pairwise classification paper
with us, and acknowledge helpful discussions with Jerry, Geoff Hinton, Radford Neal
and David Tritchler. Trevor Hastie was partially supported by grant DMS-9504495
from the National Science Foundation, and grant ROI-CA-72028-01 from the National Institutes of Health. Rob Tibshirani was supported by the Natural Sciences
and Engineering Research Council of Canada and the iRIS Centr of Excellence.
References
Bishop, Y., Fienberg, S. & Holland, P. (1975), Discrete multivariate analysis, MIT
Press, Cambridge.
Bradley, R. & Terry, M. (1952), 'The rank analysis of incomplete block designs . i.
the method of paired comparisons', Biometrics pp . 324-345 .
Deming, W. & Stephan, F . (1940), 'On a least squares adjustment of a sampled
frequency table when the expected marginal totals are known', A nn. Math .
Statist. pp. 427-444.
Friedman, J . (1996), Another approach to polychotomous classification, Technical
report, Stanford University.
Hastie, T. & Tibshirani, R. (1996), Classification by pairwise coupling, Technical
report, University of Toronto.
| 1375 |@word version:2 briefly:1 proportion:1 seems:1 logit:2 neigbours:1 simulation:1 accounting:1 covariance:3 carry:1 contains:1 score:2 bradley:5 ida:3 must:1 plot:1 drop:1 update:1 discrimination:1 guess:1 item:1 argm:1 recompute:1 provides:1 math:1 toronto:3 preference:2 simpler:1 constructed:3 qualitative:1 prove:1 combine:1 deming:2 excellence:1 pairwise:35 expected:1 hresholding:1 examine:2 nor:1 multi:1 globally:1 little:1 estimating:1 bounded:1 panel:12 biostatistics:1 substantially:1 guarantee:1 exactly:1 classifier:8 normally:1 unit:1 grant:2 producing:1 engineering:1 modify:1 despite:1 might:1 shaded:1 acknowledgment:1 block:1 procedure:18 area:2 indeterminate:1 close:2 optimize:3 equivalent:3 go:1 attention:1 rule:23 his:2 population:2 suppose:1 exact:1 caption:2 observed:1 bottom:1 tib:1 preprint:1 rij:1 utstat:1 region:2 observes:1 pd:4 compromise:1 efficiency:1 basis:1 resolved:1 joint:1 geoff:1 various:2 quite:1 whose:1 stanford:5 solve:1 larger:1 indecision:1 favor:1 statistic:3 jointly:1 final:1 ip:2 advantage:1 eigenvalue:1 combining:1 intuitive:1 description:1 convergence:2 extending:1 converges:1 coupling:23 derive:1 nearest:9 ij:6 progress:1 indeterminacy:2 job:1 involves:1 fij:4 closely:1 settle:1 assign:1 generalization:1 roi:1 predict:2 currently:1 r31:1 council:1 weighted:1 mit:1 gaussian:4 rather:2 derived:1 improvement:1 rank:1 likelihood:1 sense:2 burdensome:1 helpful:1 membership:3 nn:5 typically:1 uij:1 issue:1 classification:13 special:1 marginal:1 sampling:1 nearly:1 report:3 wms:1 modern:1 neighbour:8 ve:1 national:2 friedman:15 attempt:1 evaluation:1 mixture:3 orthogonal:1 biometrics:1 incomplete:1 qda:3 renormalize:1 nij:4 classify:1 ijth:1 ada:1 entry:1 predictor:4 uniform:2 dij:2 answer:1 density:3 together:1 polychotomous:2 satisfied:1 worse:1 expert:1 return:1 li:1 satisfy:1 later:1 root:1 start:1 bayes:4 minimize:1 il:1 square:2 modelled:2 randomness:1 history:1 halve:1 sharing:1 trevor:3 pp:2 frequency:1 coup:2 dm:1 di:1 gain:1 sampled:1 treatment:1 organized:1 back:1 appears:1 iai:1 done:2 furthermore:1 until:1 nonlinear:1 lack:1 nonparametrically:1 aj:3 lda:7 perhaps:1 requiring:1 hence:2 jerry:3 leibler:2 neal:1 iris:1 criterion:3 jti:2 confusion:1 performs:1 common:2 discriminants:1 winner:1 discussed:1 he:2 measurement:1 cambridge:1 ai:2 pointed:1 longer:1 posterior:1 multivariate:1 scenario:1 additional:1 ibshirani:1 ii:1 resolving:1 multiple:1 technical:2 match:2 calculation:1 offer:2 long:1 sphere:1 paired:4 parenthesis:2 expectation:1 addition:1 separately:2 median:1 biased:1 subject:2 tend:1 near:2 iii:1 stephan:2 fit:2 hastie:9 approaching:1 reduce:1 translates:1 six:1 action:1 tij:7 clear:1 eigenvectors:1 ten:4 statist:1 r12:1 notice:1 estimated:3 tibshirani:7 per:3 discrete:1 key:1 threshold:6 achieving:1 pj:10 neither:1 thresholded:1 prob:5 fourth:1 reasonable:1 decision:11 scaling:1 flavour:1 followed:1 quadratic:1 constraint:1 x2:1 argument:1 department:2 combination:1 aila:1 rob:1 xo:1 fienberg:2 taken:1 computationally:1 equation:3 mutually:1 discus:3 gaussians:3 apply:3 preventive:1 binomial:3 top:2 include:2 medicine:1 giving:2 looked:1 strategy:1 gradient:1 win:6 distance:3 separate:1 thank:1 simulated:3 argue:1 discriminant:5 relationship:1 minimizing:1 equivalently:1 robert:1 design:1 perform:2 observation:6 datasets:1 acknowledge:1 beat:1 situation:1 hinton:1 canada:1 david:1 pair:3 california:1 unequal:1 able:1 suggested:1 max:10 terry:5 event:2 natural:2 improve:1 coupled:2 health:1 lij:4 dating:1 nice:1 relative:1 proportional:1 foundation:1 pij:5 xp:1 sufficient:1 consistent:1 thresholding:1 pi:20 share:1 compatible:2 loa:5 repeat:1 free:1 supported:2 bias:2 institute:1 neighbor:5 benefit:1 distributed:1 boundary:12 dimension:1 hatched:1 kullback:2 logic:1 iterative:2 why:1 table:1 nature:2 learn:1 ca:1 did:1 pk:1 arise:1 fashion:2 iij:2 precision:1 wish:1 xl:1 crude:1 third:2 weighting:1 bishop:2 closeness:1 consist:1 ci:3 margin:1 easier:1 suited:1 likely:1 adjustment:1 partially:1 holland:2 radford:1 conditional:1 goal:1 specifically:1 uniformly:1 total:1 vote:1 playfair:1 formally:1 support:3 latter:1 arises:1 |
412 | 1,376 | Agnostic Classification of Markovian
Sequences
Ran EI-Yaniv
Shai Fine
Naftali Tishby*
Institute of Computer Science and Center for Neural Computation
The Hebrew University
Jerusalem 91904, Israel
E-Dlail: {ranni,fshai,tishby}Ocs.huji.ac.il
Category: Algorithms.
Abstract
Classification of finite sequences without explicit knowledge of their
statistical nature is a fundamental problem with many important
applications. We propose a new information theoretic approach
to this problem which is based on the following ingredients: (i) sequences are similar when they are likely to be generated by the same
source; (ii) cross entropies can be estimated via "universal compression"; (iii) Markovian sequences can be asymptotically-optimally
merged.
With these ingredients we design a method for the classification of
discrete sequences whenever they can be compressed. We introduce
the method and illustrate its application for hierarchical clustering
of languages and for estimating similarities of protein sequences.
1
Introd uction
While the relationship between compression (minimal description) and supervised
learning is by now well established, no such connection is generally accepted for
the unsupervised case. Unsupervised classification is still largely based on ad-hock
distance measures with often no explicit statistical justification. This is particularly true for unsupervised classification of sequences of discrete symbols which is
encountered in numerous important applications in machine learning and data mining, such as text categorization, biological sequence modeling, and analysis of spike
trains.
The emergence of "universal" (Le. asymptotically distribution independent) se?Corresponding author.
R. EI-Yaniv, S. FineandN. TlShby
466
quence compression techniques suggests the existence of "universal" classification
methods that make minimal assumptions about the statistical nature of the data.
Such techniques are potentially more robust and appropriate for real world applications.
In this paper we introduce a specific method that utilizes the connection between
universal compression and unsupervised classification of sequences. Our only underlying assumption is that the sequences can be approximated (in the information
theoretic sense) by some finite order Markov sources. There are three ingredients to
our approach. The first is the assertion that two sequences are statistically similar
if they are likely to be independently generated by the same source. This likelihood
can then be estimated, given a typical sequence of the most likely joint source, using
any good compression method for the sequence samples. The third ingredient is a
novel and simple randomized sequence merging algorithm which provably generates
a typical sequence of the most likely joint source of the sequences, under the above
Markovian approximation assumption.
Our similarity measure is also motivated by the known "two sample problem"
[Leh59] of estimating the probability that two given samples are taken from the
same distribution. In the LLd. (Bernoulli) case this problem was thoroughly investigated and the. optimal statistical test is given by the sum of the empirical cross
entropies between the two samples and their most likely joint -source. We argue
that this measure can be extended for arbitrary order Markov sources and use it to
construct and sample the most likely joint source.
The similarity measure and the statistical merging algorithm can be naturally combined into classification algorithms for sequences. Here we apply the method to
hierarchical clustering of short text segments in 18 European languages and to evaluation of similarities of protein sequences. A complete analysis of the method, with
further applications, will be presented elsewhere [EFT97].
2
Measuring the statistical similarity of sequences
Estimating the statistical similarity of two individual sequences is traditionally done
by training a statistical model for each sequence and then measuring the likelihood
of the other sequence by the model. Training a model entails an assumption about
the nature of the noise in the data and this is the rational behind most "edit
distance" measures, even when the noise model is not explicitly stated.
Estimating the log-likelihood of a sequence-sample over a discrete alphabet E by
a statistical model can be done through the Cross Entropy or Kullback-Leibler
Divergence[CT91] between the sample empirical distribution p and model distribution q, defined as:
DKL (pllq) =
L P (0-) log P((:?
q
.
(1)
uEE
The KL-divergence, however, has some serious practical drawbacks. It is nonsymmetric and unbounded unless the model distribution q is absolutely continuous
with respect to p (Le. q = 0 ::::} P = 0). The KL-divergence is therefore highly sensitive to low probability events under q. Using the "empirical" (sample) distributions
for both p and q can result in very unreliable estimates of the true divergences. Essentially, D K L [Pllq] measures the asymptotic coding inefficiency when coding the
sample p with an optimal code for the model distribution q.
The symmetric divergence, i.e. D (p, q)
= DKL [Pllq] + DKL [qllp],
suffers from
Agnostic Classification of Markovian Sequences
467
similar sensitivity problems and lacks the clear statistical meaning.
2.1
The "two sample problem"
Direct Bayesian arguments, or alternately the method of types [CK81], suggest that
the probability that there exists one source distribution M for two independently
drawn samples, x and y [Leh59], is proportional to
!
dJ-L (M) Pr (xIM) . Pr (yIM) =
!
dJ-L (M) . 2-(lzIDKdp",IIM1+lyIDKL[P1l1IM]), (2)
where dJ-L(M) is a prior density of all candidate distributions, pz and Py are the
empirical (sample) distributions, and Ixl and Iyl are the corresponding sample sizes.
For large enough samples this integral is dominated (for any non-vanishing prior)
by the maximal exponent in the integrand, or by the most likely joint source of x
and y, M>.., defined as
M>.. = argmin {lxIDKL (PzIIM') + IYIDKL (pyIlM')}.
(3)
M'
where 0 ~ A = Ixl/(lxl + Iyl) ~ 1 is the sample mixture ratio. The convexity of the
KL-divergence guarantees that this minimum is unique and is given by
M>.. = APz + (1 - A) PY'
the A - mixture of pz and py.
The similarity measure between two samples, d(x, y), naturally follows as the minimal value of the above exponent. That is,
Definition 1 The similarity measure, d(x, y) = V>..(Pz,Py), of two samples x and
y, with empirical distributions pz and Py respectively, is defined as
d(x, y) = V>..(Pz,Py) = ADKL (PzIIM>..) + (1- A) DKL (pyIIM>..)
(4)
where M>.. is the A-mixture of pz and Py.
The function V>.. (p, q) is an extension of the Jensen-Shannon divergence (see e.g.
[Lin91]) and satisfies many useful analytic properties, such as symmetry and boundedness on both sides by the L1-norm, in addition to its clear statistical meaning.
See [Lin91, EFT97] for a more complete discussion of this measure.
2.2
Estimating the V>.. similarity measure
The key component of our classification method is the estimation of V>.. for individual finite sequences, without an explicit model distribution.
Since cross entropies, D K L, express code-length differences, they can be estimated
using any efficient compression algorithm for the two sequences. The existence
of "universal" compression methods, such as the Lempel-Ziv algorithm (see e.g.
[CT91]) which are provably asymptotically optimal for any sequence, give us the
means for asymptotically optimal estimation of V>.., provided that we can obtain a
typical sequence of the most-likely joint source, M >...
We apply an improvement of the method of Ziv and Merhav [ZM93] for the estimation of the two cross-entropies using the Lempel-Ziv algorithm given two sample
sequences [BE97]. Notice that our estimation of V>.. is as good as the compression
method used, namely, closer to optimal compression yields better estimation of the
similarity measure.
It remains to show how a typical sequence of the most-likely joint source can be
generated.
468
3
R. El-Yaniv, S. Fine and N. Tishby
Joint Sources of Markovian Sequences
In this section we first explicitly generalize the notion of the joint statistical source to
finite order Markov probability measures. We identify the joint source of Markovian
sequences and show how to construct a typical random sample of this source.
More precisely, let x and y be two sequences generated by Markov processes with
distributions P and Q, respectively. We present a novel algorithm for the merging
the two sequences, by generating a typical sequence of an approximation to the
most likely joint source of x and y. The algorithm does not require the parameters
of the true sources P and Q and the computation of the sequence is done directly
from the sequence samples x and y.
As before, r; denotes a finite alphabet and P and Q denote two ergodic Markov
sources over r; of orders Kp and KQ, respectively. By equation 3, the :A-mixture
joint source M>.. of P and Q is M>.. = argminM' :ADKdPIIM')+(I-:A)DKdQIIM') ,
where for sequences DKdPIIM) = limsuPn-too ~ L:zE!:n P(x) log
The following theorem identifies the joint source of P and Q.
:1:))'
Theorem 1 The unique :A-mixture joint source M>.. of P and Q, of order K =
max {K p, K Q}, is given by the following conditional distribution. For each s E
r;K,aEE,
(1 - :A)Q(s)
:AP(s)
M>..(als) = :AP(s)
+ (1 _ :A)Q(s) P(als) + :AP(s) + (1- :A)Q(s) Q(als)
.
This distribution can be naturally extended to n sources with priors :At, ... ,:An.
3.1
The "sequence merging" algorithm
The above theorem can be easily translated into an algorithm. Figure 1 describes a
randomized algorithm that generates from the given sequences x and y, an asymptotically typical sequence z of the most likely joint source, as defined by Theorem
1, of P and Q.
Initialization:
? z [OJ = choose a symbol from x with probability ,x or y with probability 1 - ,x
? i = 0
Loop:
Repeat until the approximation error is lower then a prescribe threshold
? s", := max length suffix of z appearing somewhere in x
?
Sy
:= max length suffix of
? A(,x
,
S
""
Sy
} -
z
appearing somewhere in y
>.Pr .. (s .. )
>.Pr .. (s.,)+(l->.) Pr ll(s\I)
? r = choose x with probability A(,x, s"" Sy} or y with probability 1-A(,x, S"" S1/}
? r (Sr) = randomly choose one of the occurrences of Sr in r
? z [i + 1) = the symbol appearing immediately after r (Sr) at r
? i=i+1
End Repeat
Figure 1: The most-likely joint source algorithm
Agnostic Classification of Markovian Sequences
469
Notice that the algorithm is completely unparameterized, even the sequence alphabets, which may differ from one sequence to another, are not explicitly needed. The
algorithm can be efficiently implemented by pre-preparing suffix trees for the given
sequences, and the merging algorithm is naturally generalizable to any number of
sequences.
4
Applications
There are several possible applications of our sequence merging algorithm and similarity measure. Here we focus on three possible applications: the source merging
problem, estimation of sequence similarity, and bottom-up sequence-classification.
These algorithms are different from most existing approaches because they rely
only on the sequenced data, similar to universal compression, without explicit modeling assumptions. Further details, analysis, and applications of the method will be
presented elsewhere [EFT97].
4.1
Merging and synthesis of sequences
An immediate application of the source merging algorithm is for synthesis of typical
sequences of the joint source from some given data sequences, without any access
to an explicit model of the source.
To illustrate this point consider the sequence in Figure 2. This sequence was randomly generated, character by character, from two natural excerpts: a 47,655character string from Dickens' Tale of Two Cities, and a 59,097-character string
from Twain's The King and the Pauper.
Do your way to her breast, and sent a treason's sword- and not empty.
"I am particularly and when the stepped of his ovn commits place. No; yes,
of course, and he passed behind that by turns ascended upon him, and my bone
to touch it, less to say: 'Remove thought, everyone! Guards! In miness?"
The books third time. There was but pastened her unave misg his ruined head
than they had knovn to keep his saw whether think" The feet our grace he
called offer information?
[Twickens, 1997]
Figure 2: A typical excerpt of random text generated by the "joint source" of
Dickens and Twain.
4.2
Pairwise similarity of proteins
The joint source algorithm, combined with the new similarity measure, provide
natural means for computing the similarity of sequences over any alphabet. In this
section we illustrate this application l for the important case of protein sequences
(sequences over the set of the 20 amino-acids).
From a database of all known proteins we selected 6 different families and within
each family we randomly chose 10 proteins. The families chosen are: Chaperonin,
MHC1, Cytochrome, Kinase, Globin Alpha and Globin Beta. Our pairwise distances between all 60 proteins were computed using our agnostic algorithm and are
depicted in the 6Ox60 matrix of Figure 3. As can be seen, the algorithm succeeds to
IThe protein results presented here are part of an ongoing work with G. Yona and E .
Ben-Sasson.
R. El-Yaniv, S. Fine and N. TlShby
470
identify the families (the success with the Kinase and Cytochrome families is more
limited).
PairwIse Distances of Protein Sequences
chaperonin
MHC I
cytochrome
kinase
globin a
globin b
Figure 3: A 60x60 symmetric matrix representing the pairwise distances, as computed by our agnostic algorithm, between 60 proteins, each consecutive 10 belong
to a different family. Darker gray represent higher similarity.
In another experiment we considered all the 200 proteins of the Kinase family and
computed the pairwise distances of these proteins using the agnostic algorithm.
For comparison we computed the pairwise similarities of these sequences using the
widely used Smith-Waterman algorithm (see e.g. [HH92]).2 The resulting agnostic
similarities, computed with no biological information whatsoever, are very similar to
the Smith-Waterman similarities. 3 Furthermore, our agnostic measure discovered
some biological similarities not detected by the Smith-Waterman method.
4.3
Agnostic classification of languages
The sample of the joint source of two sequences can be considered as their "average"
or "centroid", capturing a mixture of their statistics. Averaging and measuring distance between objects are sufficient for most standard clustering algorithms such as
bottom-up greedy clustering, vector quantization (VQ), and clustering by deterministic annealing. Thus, our merging method and similarity measure can be directly
applied for the classification of finite sequences via standard clustering algorithms.
To illustrate the power of this new sequence clustering method we give the result of a
rudimentary linguistic experiment using a greedy bottom-up (conglomerative) clustering of short excerpts (1500 characters) from eighteen languages. Specifically, we
took sixteen random excerpts from the following Porto-Indo-European languages:
Afrikaans, Catalan, Danish, Dutch, English, Flemish, French, German, Italian,
Latin, Norwegian, Polish, Portuguese, Spanish, Swedish and Welsh, together with
2we applied the Smith-Waterman for computing local-alignment costs using the stateof-the-art blosum62 biological cost matrix.
3These results are not given here due to space limitations and will be discussed
elsewhere.
Agnostic Classification of Markovian Sequences
471
two artificial languages: Esperanto and Klingon4.
The resulting hierarchical classification tree is depicted in Figure 4. This entirely
unsupervised method, when applied to these short random excerpts, clearly agrees
with the "standard" philologic tree of these languages, both in terms of the grouping
and the levels of similarity (depth of the split) of the languages (the Polish-Welsh
"similarity" is probably due to the specific transcription used).
Figure 4: Agnostic bottom-up greedy clustering of eighteen languages
Acknowledgments
We sincerely thank Ran Bachrach and Golan Yona for helpful discussions. We also
thank Sageev Oore for many useful comments.
References
[BE97] R. Bachrach and R. EI-Yaniv, An Improved Measure of Relative Entropy
Between Individual Sequences, unpublished manuscript.
[CK81] 1. Csiszar and J . Krorner. Information Theory: Coding Theorems for Discrete Memoryless Systems, Academic Press, New-York 1981.
[CT91] T. M. Cover and J. A. Thomas. Elements of Information Theory, John
Wiley & Sons, New-York 1991.
[EFT97] R. EI-Yaniv, S. Fine and N. Tishby. Classifying Markovian Sources, in
preparations, 1997.
[HH92] S. Henikoff and J . G. Henikoff (1992) . Amino acid substitution matrices
from protein blocks. Proc. Natl. Acad. Sci. USA 89, 10915-10919.
[Leh59] E. L. Lehmann. Testing Statistical Hypotheses, John Wiley & Sons, NewYork 1959.
[Lin91] J. Lin, 1991. Divergence measures based on the Shannon entropy. IEEE
Transactions on In/ormation Theory, 37(1):145-15l.
[ZM93] J . Ziv and N. Merhav, 1993. A Measure of Relative Entropy Between
Individual Sequences with Application to Universal Classification, IEEE
Transactions on In/ormation Theory, 39(4).
4Klingon is a synthetic language that was invented for the Star-Trek TV series.
| 1376 |@word uee:1 compression:10 norm:1 boundedness:1 inefficiency:1 substitution:1 series:1 existing:1 portuguese:1 john:2 analytic:1 remove:1 greedy:3 selected:1 vanishing:1 short:3 smith:4 unbounded:1 guard:1 direct:1 beta:1 introduce:2 pairwise:6 x60:1 provided:1 ocs:1 estimating:5 underlying:1 agnostic:11 israel:1 argmin:1 string:2 generalizable:1 whatsoever:1 guarantee:1 before:1 local:1 flemish:1 acad:1 ap:3 chose:1 initialization:1 suggests:1 limited:1 statistically:1 practical:1 unique:2 acknowledgment:1 testing:1 block:1 empirical:5 universal:7 mhc:1 thought:1 pre:1 protein:13 suggest:1 py:7 argminm:1 deterministic:1 center:1 jerusalem:1 independently:2 ergodic:1 bachrach:2 immediately:1 his:3 notion:1 traditionally:1 justification:1 prescribe:1 hypothesis:1 element:1 approximated:1 particularly:2 ze:1 database:1 bottom:4 invented:1 ormation:2 ran:2 convexity:1 mine:1 segment:1 iyl:2 ithe:1 upon:1 completely:1 translated:1 easily:1 joint:20 newyork:1 alphabet:4 train:1 kp:1 detected:1 artificial:1 widely:1 dickens:2 say:1 compressed:1 statistic:1 think:1 emergence:1 sequence:65 henikoff:2 took:1 propose:1 maximal:1 loop:1 description:1 yaniv:6 xim:1 empty:1 categorization:1 generating:1 ben:1 object:1 illustrate:4 tale:1 ac:1 implemented:1 differ:1 foot:1 merged:1 drawback:1 porto:1 require:1 biological:4 extension:1 considered:2 consecutive:1 estimation:6 proc:1 sensitive:1 edit:1 him:1 saw:1 agrees:1 city:1 clearly:1 linguistic:1 focus:1 improvement:1 quence:1 bernoulli:1 likelihood:3 polish:2 centroid:1 sense:1 twain:2 am:1 helpful:1 el:2 suffix:3 qllp:1 her:2 italian:1 provably:2 classification:17 ziv:4 stateof:1 exponent:2 art:1 construct:2 preparing:1 unsupervised:5 serious:1 randomly:3 divergence:8 individual:4 welsh:2 mining:1 highly:1 evaluation:1 alignment:1 mixture:6 behind:2 csiszar:1 natl:1 integral:1 closer:1 unless:1 tree:3 minimal:3 modeling:2 markovian:9 assertion:1 cover:1 measuring:3 cost:2 kq:1 tishby:4 too:1 optimally:1 my:1 combined:2 thoroughly:1 synthetic:1 density:1 fundamental:1 randomized:2 huji:1 sensitivity:1 synthesis:2 together:1 choose:3 book:1 hh92:2 coding:3 star:1 explicitly:3 ad:1 bone:1 shai:1 il:1 acid:2 largely:1 efficiently:1 sy:3 yield:1 identify:2 yes:1 generalize:1 bayesian:1 suffers:1 whenever:1 danish:1 definition:1 naturally:4 rational:1 knowledge:1 manuscript:1 higher:1 supervised:1 improved:1 swedish:1 done:3 catalan:1 furthermore:1 until:1 ei:4 touch:1 hock:1 lack:1 french:1 gray:1 usa:1 true:3 symmetric:2 leibler:1 memoryless:1 ll:1 spanish:1 naftali:1 theoretic:2 complete:2 l1:1 rudimentary:1 meaning:2 novel:2 nonsymmetric:1 he:2 belong:1 discussed:1 language:10 dj:3 had:1 access:1 entail:1 similarity:23 success:1 aee:1 seen:1 minimum:1 ii:1 apz:1 academic:1 cross:5 offer:1 lin:1 dkl:4 oore:1 sageev:1 breast:1 essentially:1 dutch:1 represent:1 globin:4 sequenced:1 addition:1 fine:4 annealing:1 source:33 sr:3 probably:1 comment:1 sent:1 latin:1 iii:1 trek:1 enough:1 split:1 whether:1 motivated:1 introd:1 passed:1 york:2 generally:1 useful:2 se:1 clear:2 category:1 notice:2 estimated:3 discrete:4 express:1 key:1 threshold:1 drawn:1 asymptotically:5 sum:1 lehmann:1 place:1 family:7 utilizes:1 excerpt:5 capturing:1 entirely:1 encountered:1 precisely:1 your:1 dominated:1 generates:2 integrand:1 argument:1 tv:1 describes:1 lld:1 character:5 son:2 s1:1 pr:5 taken:1 equation:1 vq:1 remains:1 sincerely:1 turn:1 german:1 needed:1 end:1 apply:2 hierarchical:3 appropriate:1 lempel:2 yim:1 appearing:3 pllq:3 occurrence:1 existence:2 thomas:1 denotes:1 clustering:9 cytochrome:3 somewhere:2 commits:1 spike:1 grace:1 distance:7 thank:2 sci:1 stepped:1 argue:1 code:2 length:3 relationship:1 ratio:1 hebrew:1 merhav:2 potentially:1 stated:1 design:1 kinase:4 markov:5 eighteen:2 finite:6 waterman:4 immediate:1 extended:2 norwegian:1 head:1 discovered:1 arbitrary:1 namely:1 unpublished:1 kl:3 connection:2 established:1 alternately:1 max:3 oj:1 everyone:1 power:1 event:1 natural:2 rely:1 representing:1 numerous:1 identifies:1 text:3 prior:3 asymptotic:1 relative:2 limitation:1 proportional:1 ingredient:4 ixl:2 sixteen:1 sword:1 sufficient:1 classifying:1 elsewhere:3 course:1 repeat:2 english:1 side:1 institute:1 depth:1 world:1 author:1 transaction:2 alpha:1 kullback:1 transcription:1 unreliable:1 keep:1 continuous:1 nature:3 robust:1 symmetry:1 investigated:1 european:2 noise:2 amino:2 yona:2 darker:1 wiley:2 ct91:3 explicit:5 indo:1 candidate:1 third:2 theorem:5 specific:2 jensen:1 symbol:3 pz:6 grouping:1 exists:1 uction:1 quantization:1 merging:10 entropy:8 depicted:2 lxl:1 likely:12 satisfies:1 conditional:1 king:1 typical:9 specifically:1 ruined:1 averaging:1 called:1 accepted:1 succeeds:1 shannon:2 absolutely:1 preparation:1 ongoing:1 |
413 | 1,377 | Bidirectional Retrieval from Associative
Memory
Friedrich T. Sommer and Gunther Palm
Department of Neural Information Processing
University of Ulm, 89069 Ulm, Germany
{sommer,palm}~informatik.uni-ulm.de
Abstract
Similarity based fault tolerant retrieval in neural associative memories (N AM) has not lead to wiedespread applications. A drawback of the efficient Willshaw model for sparse patterns [Ste61,
WBLH69], is that the high asymptotic information capacity is of
little practical use because of high cross talk noise arising in the
retrieval for finite sizes. Here a new bidirectional iterative retrieval
method for the Willshaw model is presented, called crosswise bidirectional (CB) retrieval, providing enhanced performance. We discuss its asymptotic capacity limit, analyze the first step, and compare it in experiments with the Willshaw model. Applying the very
efficient CB memory model either in information retrieval systems
or as a functional model for reciprocal cortico-cortical pathways
requires more than robustness against random noise in the input:
Our experiments show also the segmentation ability of CB-retrieval
with addresses containing the superposition of pattens, provided
even at high memory load.
1
INTRODUCTION
From a technical point of view neural associative memories (N AM) provide data
storage and retrieval. Neural models naturally imply parallel implementation of
storage and retrieval algorithms by the correspondence to synaptic modification
and neural activation. With distributed coding of the data the recall in N AM
models is fault tolerant: It is robust against noise or superposition in the addresses
and against local damage in the synaptic weight matrix. As biological models N AM
F. T. Sommer and G. Palm
676
have been proposed as general working schemes of networks of pyramidal cells in
many places of the cortex.
An important property of a NAM model is its information capacity, measuring
how efficient the synaptic weights are used. In the early sixties Steinbuch realized
under the name "Lernmatrix" a memory model with binary synapses which is now
known as Wills haw model [Ste6I, WBLH69]. The great variety of NAM models
proposed since then, many triggered by Hopfield's work [Hop82], do not reach the
high asymptotic information capacity of the Willshaw model.
For finite network size, the Willshaw model does not optimally retrieve the
stored information, since the inner product between matrix colum and input
pattern determines the activity for each output neuron independently. For autoassociative pattern completion iterative retrieval can reduce cross talk noise
[GM76, GR92, PS92, SSP96]. A simple bidirectional iteration - as in bidirectional
associative memory (BAM) [Kos87] - can, however, not improve heteroassociative
pattern mapping. For this task we propose CB-retrieval where each retrieval step
forms the resulting activity pattern in an autoassociative process that uses the connectivity matrix twice before thresholding, thereby exploiting the stored information
more efficiently.
2
WILLSHAW MODEL AND CB EXTENSION
Here pattern mapping tasks XV -+ yV are considered for a set of memory patterns:
{(XV,yV): XV E {O,I}n,yv E {o,I}m,v = I, ... ,M}. The number of I-components
in a pattern is called pattern activity. The Willshaw model works efficiently, if
the memories are sparse, i.e., if the memory patterns have the same activities:
Ixvi = 2:~=I xi = a,lyvl = 2::1 Yi
b V v with a ? nand b ?m. During
learning the set of memory patterns is transformed to the weight matrix by
=
Cij = min(I,
L xiv}) = supxiy'j?
V
V
For a given initial pattern XJ1. the retrieval yields the output pattern yJ1. by forming
in each neuron the dendritic sum [C xJ1.]j = 2:i Ci/if and by calculating the activity
value by threshold comparison
yj = H([C xJ1.j j - 9) Vj,
(1)
with the global threshold value 9 and H(x) denoting the Heaviside function.
For finite sizes and with high memory load, Le., 0? PI := Prob[Cij = 1] ? 0.5),
the Willshaw model provides no tolerance with respect to errors in the address, see
Fig. 1 and 2. A bidirectional iteration of standard simple retrieval (1), as proposed in
BAM models [Kos87], can therefore be ruled out for further retrieval error reduction
[SP97j. In the energy function of the Willshaw BAM
E(x,y)
=-
LCijXiYj
ij
+ 8' LXi + 8
i
LYj
j
we now indroduce a factor accounting for the magnitudes of dendritic potentials at
acti vated neurons
(2)
Bidirectional Retrieval from Associative Memory
677
Differentiating the energy function (2) yields the gradient descent equations
yr W
=
+ L 'LCi j CikXi Yk
H( [CxU
k
- 8 )
(3)
i
"'-v---'
=:Wjk
X~ew
H( [CT y); +
=
L "LPiiCljYi Xl I
8' )
(4)
i
--------=:wfr
As new terms in (3) and (4) sums over pattern components weighted with the
quantities wjk and wft occur. wjk is the overlap between the matrix columns j and
k conditioned by the pattern X, which we call a conditioned link between y-units.
Restriction on the conditioned link terms yields a new iterative retrieval scheme
which we denote as crosswise bidirectional (eB) retrieval
y(r+ I)i
=
'L Cij[CT y(r-I))i -
H(
8)
(5)
L Cij[Cx(r-I))j - 8')
(6)
iEx(r)
H(
iEy(r)
For r = 0 pattern y(r:-I) has to be replaced by H([Cx(O)] - 0), for r > 2 Boolean
ANDing with results from timestep r - 1 can be applied which has been shown to
improve iterative retrieval in the Willshaw model for autoassociation [SSP96].
3
MODEL EVALUATION
Two possible retrieval error types can be distinguished: a "miss" error converts a
I-entry in Y~ to '0' and a "add" error does the opposite.
35
30
25
20
15
10
5
0
].
"2.
,."
2.
].
40
simple r. add error ... . .
C8-r. add error CB-r. miss error .....
5
10
15
20
25
30
Figure 1: Mean retrieval error rates for n = 2000, M = 15000, a = b = 10
corresponding to a memory load of H = 0.3. The x-axes display the address
activity: lilLl = 10 corresponds to a errorfree learning pattern, lower activities are
due to miss errors, higher activities due to add errors. Left: Theory - Add errors
for simple retrieval, eq. (7) (upper curve) and lower bound for the first step of
CB-retrieval, eq. (9). Right: Simulations - Errors for simple and CB retrieval.
The analysis of simple retrieval from the address i~ yields with optimal threshold
setting 0 = k the add error rate, i.e, the expectation of spurious ones:
& = (m - b)Prob
[r ~ k] ,
(7)
F. T. Sommer and G. Palm
678
with the binomial random variable Prob[r=l] = B(Lit'I,Pt}I, where B(n,p), :=
(7)pl(1 - p)n-l. a denotes the add error rate and k = lit'l - a the number of
correct 1-s in the address.
For the first step of CB-retrieval a lower bound of the add error rate a(l) can be
derived by the analysis of CB-retrieval with fixed address x(O) = iIJ. and the perfect
learning pattern ylJ. as starting patterns in the y-Iayer. In this case the add error
rate is:
(8)
where the random variables rl and r2 have the distributions:
Prob [rl = lib] = B(k, PI), and Prob [r2 = 1] = B(ab, (PI )2) l" Thus,
a(l) ~ (m - b)
k
L B(k, PdsBS [ab, (PI )2, (k -
s)b) ,
(9)
8=0
where BS [n,p, t] := L:~t B(n,p), is the binomial sum.
In Fig. 1 the analytic results for the first step (7) and (9) can be compared with
simulations (left versus right diagram) . In the experiments simple retrieval is performed with threshold () = k. CB-retrieval is iterated in the y-Iayer (with fixed
address x) starting with three randomly chosen 1-s from the simple retrieval result
yt'. The iteration is stopped, if a stable pattern at threshold e = bk is reached.
The memory capacity can be calculated per pattern component under the assumption that in the memory patterns each component is independent, i.e., the probabilities for a 1 are p = a/n or q = b/m respectively, and the probabilities of an add
and a miss error are simply the renormalized rates denoted by a', {3' and a', {3' for
x-patterns and by,', 6' for y-patterns. The information about the stored pattern
contained in noisy initial or retrieved patterns is then given by the transinformation t(p,a',{3') := i(p) -i(p,a',{3'), where i(p) is the Shannon information, and
i (p, a', {3') the conditional information. The heteroassociative mapping is evaluated
by the output capacity: A(a', {3') := Mm t(q, ,', 6')/mn (in units bit/synapse). It
depends on the initial noise since the performance drops with growing initial errors
and assumes the maximum, if no fault tolerance is provided, that is, with noiseless
initial patterns, see Fig. 2. Autoassociative completion of a distorted x-pattern is
evaluated by the completion capacity: C(a', {3') := Mn(t(p, a', {3')-t(p, a', {3'))/mn.
A BAM maps and completes at the same time and should be therefore evaluated
by the search capacity S := C + A.
The asymptotic capacity of the Willshaw model is strikingly high: The completion
capacity (for autoassociation) is C+ = In[2] /4, the mapping capacity (for heteroassociation with input noise) is A+ = In[2] /2 bit/syn [Pal91]' leading to a value for
the search capacity of (3 In[2])/4 = 0.52 bit/syn. To estimate S for general retrieval
procedures one can consider a recognition process of stored patterns in the whole
space of sparse initial patterns; an initial pattern is "recognized", if it is invariant under a bidirectional retrieval cycle. The so-called recognition capacity of this
process is an upper bound of the completion capacity and it had been determined
as In [2J/2, see [PS92]. This is achieved again with parameters M, p, q providing
A = In[2] /2 yielding In[2] bit/syn as upper bound of the asymptotic search capacity. In summary, we know about the asymptotic search capacity of the CB-model:
0.52 ::; S+ ::; 0.69 bit/syn. For experimental results, see Fig. 4.
Bidirectional Retrieval from Associative Memory
4
679
EXPERIMENTAL RESULTS
The CB model has been tested in simulations and compared with the Willshaw
model (simple retrieval) for addresses with random noise (Fig. 2) and for addresses
composed by two learning patterns (Fig. 3). In Fig. 2 the widely enlarged range of
high qualtity retrieval in the CB-model is demonstrated for different system sizes.
output miss errors
6
5
4
3
2
1
0
10
simple r. .. ".
:
CB?r. - .. : "
6
4
.'
, .. "...
10 15 20 25 30
.'
, .. ~ '"
.......
::'
2
5
7
6
5
4
3
2
1
0
8
0
5
10 15 20 25 30
7
6
5
4
3
2
101214161820
0
101214161820
output add errors
6
5
4
3
r--..--.--,--.----.---,
14
12
simple r. " ."
CB?r.-
10
8
/"
2
6
4
2
0
','
1
o ~'--'--'--'--1----'-""'-'-----'
5
rcr-..---,----.~----,
10 15 20 25 30
10
10
8
8
6
6
.,.,.; , ?.-:: .... os?,?
.:'1 ...... ?
/",
L-.C..:....I..!.-...I.-_~=:::::::..J
5
4
4
2
0
2
0
1015202530
101214161820
L..:....L-I.........."--I
101214161820
transinformation in output pattern (bit)
50 ;c;::r;:::::r:::=r::=7---"J 100
45 L
~
100 r-r--,-,--,--,1 00 r-r--.-.-T'"""1
~
35
30
25
...... .
~
15
10
5
s imple r. ..".
?CB-r. -
".
20
o '---''---'----1---'----'---'
5
10 15 20 25 30
~
~
60
60
~
40
20
20
o
5
10 15 20 25 30
Fig. 2: Retrieval from addresses with random
noise. The x-axis labeling is as in Fig. 1. Small
system with n = 100, M = 35 (left), system size
as in Fig. 1, two trials (right). Output activities
adjusted near Iyl = k by threshold setting.
101214161820
0
L-.I.---L.--1-.I..--J
101214161820
Fig. 3: Retrieval from addresses
composed by two learning patterns. Parameters as in right column of Fig. 2, explanation of left
and right column, see text.
In Fig. 3 the address contains one learning pattern and I-components of a second
learning pattern successively added with increasing abscissae. On the right end
of each diagram both patterns are completely superimposed. Diagrams in the left
column show errors and transinformation, if retrieval results are compared with
the learning pattern which is for li~ I < 20 dominantly addressed. Simple retrieval
errors behave similiar as for random noise in the address (Fig. 2) while the error
level of CB-retrieval raises faster if more than 7 adds from the second pattern are
present. Diagrams in the right column show the same quantities, if the retrieval
result is compared with the closest of the two learning patterns. It can be observed
i) that a learning pattern is retrieved even if the address is a complete superposition and ii) if the second pattern is almost complete in the address the retrieved
pattern corresponds in some cases to the second pattern. However, in all cases CBretrieval yields one of the learning pattern pairs and it could be used to generate
a good address for further retrieval of the other by deletion of the corresponding
I-components in the original address.
680
F. T. Sommer and G. Palm
0.48
0.46
0.44
0.42
0.4
output c. .....
searchc. -
0.38 .
8
.....
10 12 14 16 18
Fig. 4: Output and search capacity of CB retrieval in
bit/syn with x-axis labeling as in Fig. 2 for n = m = 2000,
a = b = 10 M = 20000. The difference between both curves
is the contribution due to x-pattern completion, the completion capacity C. It is zero for Ix(O}1 = 10, if the initial
pattern is errorfree.
The search capacity of the CB model in Fig. 4 is close to the theoretical expectations
from Sect. 3, increasing with input noise due to the address completion.
5
SPARSE CODING
To apply the proposed N AM model, for instance, in information retrieval, a coding
of the data to be accessed into sparse binary patterns is required. A useful extraction
of sparse features should take account of statistical data properties and the way the
user is acting on them. There is evidence from cognitive psychology that such a
coding is typically quite easy to find. The feature encoding, where a person is
extracting feature sets to characterize complex situations by a few present features,
is one of the three basic classes of cognitive processes defined by Sternberg [Ste77].
Similarities in the data are represented by feature patterns having a large number
of present features in common, that is a high overlap: o(x, x'} := L:i XiX'i' For text
retrieval word fragments used in existing indexing techniques can be directly taken
as sparse binary features [Geb87]. For image processing sparse coding strategies
[Zet90], and neural models for sparse feature extraction by anti-Hebbian learning
[F6l90] have been proposed. Sparse patterns extracted from different data channels
in heterogeneous data can simply be concatenated and processed simultaneously in
N AM. If parts of the original data should be held in a conventional memory, also
these addresses have to be represented by distributed and sparse patterns in order
to exploit the high performance of the proposed NAM.
6
CONCLUSION
A new bidirectional retrieval method (CB-retrieval) has been presented for the Willshaw neural associative memory model. Our analysis of the first CB-retrieval step
indicates a high potential for error reduction and increased input fault tolerance.
The asymptotic capacity for bidirectional retrieval in the binary Willshaw matrix
has been determined between 0.52 and 0.69 bit/syn. In experiments CB-retrieval
showed significantly increased input fault tolerance with respect to the standard
model leading to a practical information capacity in the order of the theoretical
expectations (0.5 bit/syn). Also the segmentation ability of CB-retrieval with ambiguous addresses has been shown. Even at high memory load such input patterns can be decomposed and corresponding memory entries returned individually.
The model improvement does not require sophisticated individual threshold setting
[GW95], strategies proposed for BAM like more complex learning procedures, or
"dummy augmentation" in the pattern coding [WCM90, LCL95].
The demonstrated performance of the CB-model encourages applications as massively parallel search strategies in Information Retrieval. The sparse coding requirement has been briefly discussed regarding technical strategies and psychological plausibility. Biologically plausible variants of CB-retrieval contribute to more
Bidirectional Retrieval from Associative Memory
681
refined cell assembly theories, see [SWP98].
Acknowledgement: One of the authors (F.T.S.) was supported by grant S0352/3-1
of the Deutsche Forschungsgemeinschaft.
References
[F6190]
P. F6ldiak. Forming sparse representations by local anti-hebbian learning.
Biol. Cybern., 64:165-170, 1990.
[Geb87]
F. Gebhardt. Text signatures by superimposed coding of letter triplets and
quadruplets. Information Systems, 12(2):151-156, 1987.
[GM76]
A.R. Gardner-Medwin. The recall of events through the learning of associations between their parts. Proceedings of the Royal Society of London B,
194:375-402, 1976.
[GR92]
W.G. Gibson and J. Robinson. Statistical analysis of the dynamics of a sparse
associative memory. Neural Networks, 5:645-662, 1992.
[GW95]
B. Graham and D. Willshaw. Improving recall from an associative memory.
Biological Cybernetics, 72:337-346, 1995.
[Hop82]
J.J. Hopfield. Neural networks and physical systems with emergent collective
computational abilities. Proceedings of the National Academy of Sciences,
USA, 79, 1982.
[Kos87]
B. Kosko. Adaptive bidirectional associative memories.
26(23):4947-4971, 1987.
[LCL95]
C.-S. Leung, L.-W. Chan, and E. Lai. Stability, capacity and statistical dynamics of second-order bidirectional associative memory. IEEE Trans. Syst,
Man Cybern., 25(10):1414-1424, 1995.
[Pal91]
G. Palm. Memory Capacities of Local Rules for Synaptic Modification. Concepts in Neuroscience, 2:97-128, 1991.
[PS92]
G. Palm and F. T. Sommer. Information capacity in recurrent McCulloch-Pitts
networks with sparsely coded memory states. Network, 3:1-10, 1992.
[SP97]
F. T. Sommer and G. Palm. Improved bidirectional retrieval of sparse patterns
stored by Hebbian learning. Submitted to Neural Networks, 1997.
[SSP96]
F. Schwenker, F. T. Sommer, and G. Palm. Iterative retrieval of sparsely coded
associative memory patterns. Neural Networks, 9(3) :445 - 455, 1996.
[Ste61]
K. Steinbuch. Die Lernmatrix. Kybernetik, 1:36-45, 1961.
[Ste77]
R. J. Sternberg. Intelligence, information processing and analogical reasoning.
Hillsdale, NJ, 1977.
[SWP98]
F. T. Sommer, T. Wennekers, and G. Palm. Bidirectional completion of Cell
Assemblies in the cortex. In Computational Neuroscience: Trends in Research.
Plenum Press, 1998.
Applied Optics,
[WBLH69] D. J. Willshaw, O. P. Buneman, and H. C. Longuet-Higgins. Nonholographic
associative memory. Nature, 222:960-962, 1969.
[WCM90]
Y. F. Wang, J. B. Cruz, and J. H. Mulligan. Two coding stragegies for bidirectional associative memory. IEEE Trans. Neural Networks, 1(1):81-92, 1990.
[Zet90]
C. Zetsche. Sparse coding: the link between low level vision and associative
memory. In R. Eckmiller, G. Hartmann, and G. Hauske, editors, Parallel
Processing in Neural Systems a.nd Computers. Elsevier Science Publishers B.
V. (North Holland), 1990.
| 1377 |@word trial:1 briefly:1 nd:1 simulation:3 heteroassociative:2 accounting:1 thereby:1 reduction:2 initial:8 contains:1 fragment:1 denoting:1 existing:1 activation:1 cruz:1 analytic:1 drop:1 intelligence:1 yr:1 reciprocal:1 provides:1 contribute:1 accessed:1 acti:1 pathway:1 abscissa:1 growing:1 decomposed:1 little:1 lib:1 increasing:2 provided:2 deutsche:1 mcculloch:1 nj:1 willshaw:16 unit:2 grant:1 before:1 local:3 xv:3 limit:1 kybernetik:1 encoding:1 xiv:1 gr92:2 twice:1 eb:1 autoassociation:2 range:1 practical:2 yj:1 procedure:2 gibson:1 significantly:1 word:1 close:1 storage:2 applying:1 cybern:2 restriction:1 conventional:1 map:1 demonstrated:2 yt:1 starting:2 independently:1 rule:1 higgins:1 nam:3 retrieve:1 stability:1 plenum:1 enhanced:1 anding:1 pt:1 ulm:3 user:1 us:1 trend:1 recognition:2 steinbuch:2 gw95:2 sparsely:2 observed:1 wang:1 cycle:1 sect:1 yk:1 imple:1 dynamic:2 renormalized:1 signature:1 raise:1 iyl:1 completely:1 strikingly:1 hopfield:2 emergent:1 represented:2 talk:2 london:1 labeling:2 refined:1 quite:1 widely:1 plausible:1 ability:3 noisy:1 associative:16 triggered:1 propose:1 product:1 haw:1 hop82:2 academy:1 analogical:1 wjk:3 exploiting:1 requirement:1 perfect:1 recurrent:1 completion:9 ij:1 eq:2 drawback:1 correct:1 hillsdale:1 require:1 biological:2 dendritic:2 adjusted:1 extension:1 pl:1 mm:1 considered:1 great:1 cb:26 mapping:4 pitt:1 early:1 superposition:3 individually:1 weighted:1 wft:1 iex:1 ax:1 derived:1 ylj:1 improvement:1 superimposed:2 indicates:1 am:6 elsevier:1 leung:1 typically:1 nand:1 spurious:1 transformed:1 germany:1 hartmann:1 denoted:1 extraction:2 having:1 lit:2 patten:1 few:1 randomly:1 composed:2 simultaneously:1 national:1 individual:1 replaced:1 ab:2 rcr:1 evaluation:1 sixty:1 yielding:1 held:1 ruled:1 theoretical:2 stopped:1 psychological:1 wfr:1 column:5 transinformation:3 boolean:1 instance:1 increased:2 measuring:1 entry:2 optimally:1 stored:5 characterize:1 person:1 connectivity:1 again:1 augmentation:1 successively:1 containing:1 cognitive:2 leading:2 li:1 syst:1 account:1 potential:2 de:1 coding:10 north:1 depends:1 performed:1 view:1 analyze:1 reached:1 yv:3 parallel:3 contribution:1 efficiently:2 yield:5 iterated:1 informatik:1 lci:1 cybernetics:1 submitted:1 synapsis:1 reach:1 synaptic:4 against:3 energy:2 naturally:1 recall:3 segmentation:2 syn:7 sophisticated:1 bidirectional:18 higher:1 improved:1 synapse:1 evaluated:3 working:1 o:1 lernmatrix:2 usa:1 name:1 xj1:3 concept:1 during:1 quadruplet:1 encourages:1 ambiguous:1 die:1 complete:2 reasoning:1 image:1 common:1 functional:1 rl:2 physical:1 discussed:1 association:1 had:1 stable:1 similarity:2 cortex:2 add:12 closest:1 showed:1 chan:1 retrieved:3 massively:1 binary:4 fault:5 yi:1 recognized:1 ii:1 hebbian:3 technical:2 faster:1 plausibility:1 cross:2 retrieval:56 lai:1 coded:2 buneman:1 variant:1 basic:1 heterogeneous:1 noiseless:1 expectation:3 vision:1 iteration:3 achieved:1 cell:3 addressed:1 diagram:4 pyramidal:1 completes:1 publisher:1 call:1 extracting:1 near:1 forschungsgemeinschaft:1 easy:1 variety:1 psychology:1 opposite:1 inner:1 reduce:1 regarding:1 returned:1 autoassociative:3 useful:1 processed:1 generate:1 neuroscience:2 arising:1 per:1 dummy:1 eckmiller:1 threshold:7 gunther:1 timestep:1 convert:1 sum:3 prob:5 letter:1 distorted:1 place:1 almost:1 graham:1 bit:9 bound:4 ct:2 display:1 correspondence:1 activity:9 occur:1 optic:1 sternberg:2 min:1 c8:1 palm:10 department:1 modification:2 b:1 biologically:1 invariant:1 indexing:1 taken:1 equation:1 errorfree:2 discus:1 know:1 end:1 yj1:1 apply:1 distinguished:1 robustness:1 lxi:1 original:2 binomial:2 denotes:1 sommer:9 assumes:1 assembly:2 calculating:1 exploit:1 concatenated:1 society:1 added:1 realized:1 quantity:2 damage:1 strategy:4 gradient:1 link:3 capacity:24 providing:2 cij:4 f6ldiak:1 implementation:1 collective:1 upper:3 neuron:3 kosko:1 finite:3 descent:1 behave:1 similiar:1 anti:2 situation:1 bk:1 pair:1 required:1 friedrich:1 deletion:1 robinson:1 address:21 trans:2 pattern:55 bam:5 royal:1 memory:32 explanation:1 overlap:2 event:1 mn:3 scheme:2 improve:2 imply:1 gardner:1 axis:2 text:3 acknowledgement:1 asymptotic:7 versus:1 thresholding:1 editor:1 pi:4 summary:1 supported:1 dominantly:1 cortico:1 differentiating:1 sparse:16 distributed:2 tolerance:4 curve:2 calculated:1 cortical:1 xix:1 author:1 adaptive:1 uni:1 global:1 tolerant:2 xi:1 iayer:2 search:7 iterative:5 triplet:1 channel:1 nature:1 robust:1 longuet:1 improving:1 complex:2 vj:1 whole:1 noise:10 enlarged:1 fig:17 iij:1 lyj:1 heteroassociation:1 xl:1 ix:1 load:4 vated:1 r2:2 evidence:1 ci:1 magnitude:1 conditioned:3 cx:2 simply:2 forming:2 contained:1 holland:1 corresponds:2 determines:1 extracted:1 conditional:1 man:1 determined:2 acting:1 miss:5 called:3 experimental:2 shannon:1 ew:1 hauske:1 heaviside:1 tested:1 biol:1 |
414 | 1,378 | Using Expectation to Guide Processing:
A Study of Three Real-World Applications
Shumeet 8aluja
Justsystem Pittsburgh Research Center &
School of Computer Science, Carnegie Mellon University
baluja@cs.cmu.edu
Abstract
In many real world tasks, only a small fraction of the available inputs are important
at any particular time. This paper presents a method for ascertaining the relevance
of inputs by exploiting temporal coherence and predictability. The method proposed in this paper dynamically allocates relevance to inputs by using expectations
of their future values. As a model of the task is learned, the model is simultaneously extended to create task-specific predictions of the future values of inputs.
Inputs which are either not relevant, and therefore not accounted for in the model,
or those which contain noise, will not be predicted accurately. These inputs can be
de-emphasized, and, in turn, a new, improved, model of the task created. The techniques presented in this paper have yielded significant improvements for the
vision-based autonomous control of a land vehicle, vision-based hand tracking in
cluttered scenes, and the detection of faults in the etching of semiconductor wafers.
1 Introduction
In many real-world tasks, the extraneous information in the input can be easily confused
with the important features, making the specific task much more difficult. One of the
methods in which humans function in the presence of many distracting features is to selectively attend to only portions of the input signal. A means by which humans select where
to focus their attention is through the use of expectations. Once the important features in
the current input are found, an expectation can be formed of what the important features in
the next inputs will be, as well as where they will be. The importance of features must be
determined in the context of a specific task; different tasks can require the processing of
different subsets of the features in the same input.
There are two distinct uses of expectations. Consider Carnegie Mellon's Navlab autonomous navigation system. The road-following module [Pomerleau, 1993] is separate from
the obstacle avoidance modules [Thorpe, 1991]. One role of expectation, in which unexpected features are de-emphasized, is appropriate for the road-following module in which
the features to be tracked, such as lane-markings, appear in predictable locations. This use
of expectation removes distractions from the input scene. The second role of expectation,
to emphasize unexpected features, is appropriate for the obstacle avoidance modules. This
use of expectation emphasizes unanticipated features of the input scene.
2 Architectures for Attention
In many studies of attention, saliency maps (maps which indicate input relevance) have
been constructed in a bottom-up manner. For example, in [Koch & Ullman, 1985], a
s. Baluja
860
saliency map, which is not task-specific, is created by emphasizing inputs which are different from their neighbors. An alternate approach, presented in [Clark & Ferrier, 1992],
places mUltiple different, weighted, task-specific feature detectors around the input image.
The regions of the image which contain high weighted sums of the detected features are
the portion of the scene which are focused upon. Top-down knowledge of which features
are used and the weightings of the features is needed to make the procedure task-specific.
In contrast, the goal of this study is to learn which task-specific features are relevant without requiring top-down knowledge.
In this study, we use a method based on Input Reconstruction Reliability Estimation
(IRRE) [Pomerleau, 1993] to detennine which portions of the input are important for the
task. IRRE uses the hidden units of a neural network (NN) to perfonn the desired task and
to reconstruct the inputs. In its original use, IRRE estimated how confident a network's
outputs were by measuring the similarity between the reconstructed and current inputs.
Figure 1(Left) provides a schematic ofIRRE. Note that the weights between the input and
hidden layers are trained to reduce both task and reconstruction error.
Because the weights between the input and hidden layers are trained to reduce both task
and reconstruction error, a potential drawback of IRRE is the use of the hidden layer to
encode all of the features in the image, rather than only the ones required for solving the
particular task [Pomerleau, 1993]. This can be addressed by noting the following: if a
strictly layered (connections are only between adjacent layers) feed-forward neural network can solve a given task, the activations of the hidden layer contain, in some fonn, the
important infonnation for this task from the input layer. One method of detennining what
is contained in the hidden layer is to attempt to reconstruct the original input image, based
solely upon the representation developed in the hidden layer. Like IRRE, the input image
is reconstructed from the activations of the units in the hidden layer. Unlike IRRE, the hidden units are not trained to reduce reconstruction error, they are only trained to solve the
panicular task. The network's allocation of its limited representation capacity at the hidden layer is an indicator of what it deems relevant to the task. Information which is not relevant to the task will not be encoded in the hidden units. Since the reconstruction of the
inputs is based solely on the hidden units' activations, and the irrelevant portions of the
input are not encoded in the hidden units' activations, the inputs which are irrelevant to the
task cannot be reconstructed. See Figure I(Right).
By measuring which inputs can be reconstructed accurately, we can ascertain which inputs
the hidden units have encoded to solve the task. A synthetic task which demonstrates this
idea is described here. Imagine being given a lOxlO input retina such as shown in
Figure 2a&b. The task is to categorize many such examples into one of four classes.
Because of the random noise in the examples, the simple underlying process, of a cross
being present in one of four locations (see Figure 2c), is not easily discernible, although it
is the feature on which the classifications are to be based. Given enough examples, the NN
will be able to solve this task. However, even after the model of the task is learned, it is
difficult to ascertain to which inputs the network is attending. To detennine this, we can
freeze the weights in the trained network and connect a input-reconstruction layer to the
hidden units, as shown in Figure 1(Right). After training these connections, by measuring
where the reconstruction matches the actual input, we can detennine which inputs the network has encoded in its hidden units, and is therefore attending. See Figure 2d.
weights
trained to
reduce task
error only
weights
trained to reduce
reconstruction
error only.
error.
Figure 1:
weights
trained to
reduce task
error only.
(Left) IRRE. (Right) Modified IRRE.
"-
weights
trained to reduce
reconstruction
error only.
Using Expectation to Guide Processing
B:
861
C:
D:
2
3
4
+
ir+
Figure 2: (A & B): Samples of training data (cross appears in position 4 & 1 respectively). Note the large
amounts of noise. (C): The underlying process puts a cross in one of these four locations. (D): The black
crosses are where the reconstruction matched the inputs; these correspond exactly to the underlying process.
IRRE and this modified IRRE are related to auto-encoding networks [Cottrell, 1990] and
principal components analysis (PeA). The difference between auto-encoding networks
and those employed in this study is that the hidden layers of the networks used here were
trained to perfonn well on the specific task, not to reproduce the inputs accurately.
2.1 Creating Expectations
A notion of time is necessary in order to focus attention in future frames. Instead of reconstructing the current input, the network is trained to predict the next input; this corresponds to changing the subscript in the reconstruction layer of the network shown in
Figure 1(Right) from t to t+ 1. The prediction is trained in a supervised manner, by using
the next set of inputs in the time sequence as the target outputs. The next inputs may contain noise or extraneous features. However, since the hidden units only encode infonnation
to solve the task, the network will be unable to construct the noise or extraneous features
in its prediction.
To this point, a method to create a task-specific expectation of what the next inputs will be
has been described. As described in Section 1, there are two fundamentally different ways
in which to interpret the difference between the expected next inputs and the actual next
inputs. The first interpretation is that the difference between the expected and the actual
inputs is a point of interest because it is a region which was not expected. This has applications in anomaly detection; it will be explored in Section 3.2. In the second interpretation, the difference between the expected and actual inputs is considered noise. Processing
should be de-emphasized from the regions in which the difference is large. This makes the
assumption that there is enough infonnation in the previous inputs to specify what and
where the important portions of the next image will be. As shown in the road-following
and hand-tracking task, this method can remove spurious features and noise.
3 Real-World Applications
1bree real-world tasks are discussed in this section. The first, vision-based road following,
shows how the task-specific expectations developed in the previous section can be used to
eliminate distractions from the input. The second, detection of anomalies in the plasmaetch step of wafer fabrication, shows how expectations can be used to emphasize the unexpected'features in the input. The third, visual hand-tracking, demonstrates how to incorporate a priori domain knowledge about expectations into the NN.
3.1 Application 1: Vision-Based Autonomous Road Following
In the domain of autonomous road following, the goal is to control a robot vehicle by analyzing the image of the road ahead. The direction of travel should be chosen based on the
location of important features like lane markings and r~ad edges. On highways and dirt
roads, simple techniques, such as feed-forward NNs, have worked well for mapping road
images to steering commands [Pomerleau, 1993]. However, on city streets, where there
are distractions like old lane-.narkings, pedestrians, and heavy traffic, these methods fail.
The purpose of using attention in this domain is to eliminate features of the road which the
NN may mistake as lane markings. Approximately 1200 images were gathered from a
862
S. Baluja
E
F
Figure 3: (Top) : Four samples of training images . Left most
shows the position of the lane-marking which was hand-marked.
(Right): In each triplet: Left: raw input imagtt. Middle: the
network's prediction of the inputs at time t; this prediction was
made by a network with input ofimaget_I ' Right: a pixel-by-pixel
filtered image (see text). This image is used as the input to the NN.
G
camera mounted on the left side of the CMU-Navlab 5 test vehicle, pointed downwards
and slightly ahead of the vehicle. The car was driven through city and residential neighborhoods around Pittsburgh, PA. The images were gathered at 4-5 hz. The images were
subsampled to 30x32 pixels. In each of these images, the horizontal position of the lane
marking in the 20th row of the input image was manually identified. The task is to produce
a Gaussian of activation in the outputs centered on the horizontal position of the lane
marking in the 20th row of the image, given the entire input image. Sample images and
target outputs are shown in Figure 3. In this task, the ANN can be confused by road edges
(Figure 3a), by extraneous lane markings (Figure 3b), and reflections on the car itself
(since the camera was positioned on the side of the car), as shown in Figure 3c.
The network architecture shown in Figure 4 was used; this is the same architecture as in
Figure l(right) with the feedback shown. The feedback is used during both training and
simulation. In each time-step, a steering direction and a prediction of the next inputs is
produced. For each time-step, the magnitude of the difference' between the input's
expected value (computed in the previous time-step) and its actual value is computed.
Each input pixel can be moved towards its background value l in proportion to this difference-value. The larger the difference value, the more weight is given to the background
value. If the difference value is small, the actual inputs are used. This has the effect of deemphasizing the unexpected inputs.
The results of using this method were very promising. The lane tracker removed distracting features from the images. In Figure 3G, a distracting lane-marking is removed: the
lane marker on the right was correctly tracked in images before the distractor lane-marker
appeared. In Figure 3F, a passing car is de-emphasized: the network does not have a model
to predict the movement of passing cars, since these are not relevant for the lane-marker
detection task. In Figure 3E, the side of the road appears brighter than expected; therefore
it is de-emphasized. Note that the expectation-images (shown in the middle of each triplet
weights
trained to
reduce task
error only.
Delayed I time step
I ...~~....
__
weights
--- trained to reduce
prediction
error only.
m~tii;:mrt;,._--1 Weight bkgd.
and actual
inputs according to
difference ima e.
Figure 4:
t - - - t Difference
between
inputs t &
predicted in uts
Architecture used
to track the lane
marking in
cluttered scenes.
Signal Transfer
(Connections are not trainable)
I. A simple estimate of the background value for each pixel is its average activation across the training set. For
the road-following domain, it is possible to use a background activation of 0.0 (when the entire image is scaled to
activations of +1.0 to -1.0) since the road often appears as intermediate grays.
Using Expectation to Guide Processing
863
in Figure 3) show that the expected lane-marker and road edge locations are not precisely
defined. This is due to the training method, which attempts to model the many possible
transitions from one time step to the next to account for inter- and intra-driver variability
with a limited training set [Baluja, 1996].
In summary, by eliminating the distractions in the input images, the lane-tracker with the
attention mechanisms improved performance by 20% over the standard lane-tracker, measured on the difference between the estimated and hand-marked position of the lanemarker in each image. This improvement was seen on multiple runs, with random initial
weights in the NN and different random translations chosen for the training images.
3.2 Application 2: Fault Detection in the Plasma-Etch Wafer Fabrication
Plasma etch is one of the many steps in the fabrication of semiconductor wafers. In this
study, the detection of four faults was attempted. Descriptions of the faults can be found in
[Baluja, 1996][Maxion, 1996]. For the experiments conducted here, only a single sensor
was used, which measured the intensity of light emitted from the plasma at the 520nm
wavelength. Each etch was sampled once a second, providing approximately 140 samples
per wafer wavefonn. The data-collection phase of this experiment began on October 25,
1994, and continued until April 4, 1995. The detection of faults is a difficult problem
because the contamination of the etch chamber and the degradation parts keeps the sensor's outputs, even for fault-free wafers, changing over time. Accounting for machine state
should help the detection process.
Expectation is used as follows: Given the waveform signature of waferT_I' an expectation
of waferT can be fonned. The input to the prediction-NN is the wavefonn signature of
waferT_I; the output is the prediction of the signature of waferT. The target output for each
example is the signature of the next wafer in sequence (the full 140 parameters). Detection
of the four faults is done with a separate network which used as input: the expectation of
the wafer's wavefonn, the actual wafer's wavefonn, and the point-by-point difference of
the two. In this task, the input is not filtered as in the driving domain described previously;
the values of the point-by-point difference vector are used as extra inputs.
The perfonnance of many methods and architectures were compared on this task, details
can be found in [Baluja, 1996]. The results using the expectation based methods was a
98.7% detection rate, 100% classification rate on the detected faults (detennining which of
the four types of faults the detected fault was), and a 2.3% false detection rate. For comparison, a simple perceptron had an 80% detection rate, and a 40% false-detection rate. A
fully-connected network which did not consider the state of the machine achieved a 100%
detection rate, but a 53% false detection rate. A network which considered state by using
the last-previous no-fault wafer for comparison with the current wafer (instead of an
expectation for the current wafer) achieved an 87.9% detection rate, and a 1.5% falsedetection rate. A variety of neural and non-neural methods which examined the differences between the expected and current wafer, as well those which examined the differences between the last no-fault wafer and the current wafer, perfonned poorly. In
summary, methods which did not use expectations were unable to obtain the false-positives and detection rates of the expectation-based methods.
3.3 Application 3: Hand-Tracking in Cluttered Scenes
In the tasks described so far, the transition rules were learned by the NN. However, if the
transition rules had been known a priori, processing could have been directed to only the
relevant regions by explicitly manipulating the expectations. The ability to incorporate a
priori rules is important in many vision-based tasks. Often the constraints about the environment in which the tracking is done can be used to limit the portions of the input scene
which need to be processed. For example, consider visually tracking a person's hand.
Given a fast camera sampling rate, the person's hand in the current frame will be close to
864
S. Baluja
B.
A.
Figure 5:
Typical input images used for
the hand-tracking experiments.
The target is to track the
subject's right hand. Without
expectation, in (A) both hands
were found in X outputs, and
the wrong hand was found in
the Y outputs. In (8) Subject's
right hand and face found in
the X outputs.
where it appeared in the previous frame. Although a network can learn this constraint by
developing expectations of future inputs (as with the NN architecture shown in Figure 4),
training the expectations can be avoided by incorporating this rule directly.
In this task, the input layer is a 48*48 image. There are two output layers of 48 units; the
desired outputs are two gaussians centered on the (X,Y) position of the hand to be tracked.
See Figure 5. Rather than creating a saliency map based upon the difference between the
actual and predicted inputs, as was done with autonomous road following, the saliency
map was explicitly created with the available domain knowledge. Given the sampling rate
of the camera and the size of the hand in the image, the salient region for the next timestep was a circular region centered on the estimated location of the hand in the previous
image. The activations of the inputs outside of the salient region were shifted towards the
background image. The activations inside the salient region were not modified. After
applying the saliency map to the inputs, the filtered inputs were fed into the NN.
This system was tested in very difficult situations; the testing set contained images of a
person moving both of his hands and body throughout the sequence (see Figure 5). Therefore, both hands and body are clearly visible in the difference images used as input into
the network. All training was done on much simpler training sets in which only a single
hand was moving. To gauge the perfonnance of an expectation-based system, it was compared to a system which used the following post-processing heuristics to account for temporal coherence. First, before a gaussian was fit to either of the output layers, the
activation of the outputs was inversely scaled with the distance away from the location of
the hand in the previous time step. This reduces the probability of detecting a hand in a
location very different than the previous detection. This helps when both hands are
detected, as shown in Figure 5. The second heuristic was that any predictions which differ
from the previous prediction by more than half of the dimension of the output layer were
ignored, and the previous prediction used instead. See Table I for the results. In summary,
by using the expectation based methods, perfonnance improved from 66% to 90% when
tracking the left hand, and 52% to 91 % when tracking the right hand.
Table I: Performance: Number of frames in which each hand was located (283 total images).
Method
No Heuristics, No Expect.
Heuristics
Expectation
Expectation + Heuristics
Target: Find Left Hand
Target: Find Right Hand
Which Hand Was Found
Which Hand Was Found
% Correct
L
52%
66%
91%
90%
146
187
258
256
R
44
22
3
3
None
% Correct
L
93
74
22
24
16%
52%
90%
91%
143
68
3
2
R
47
147
255
257
None
93
68
25
24
[Nowlan & Platt, 1995] presented a convolutional-NN based hand-tracker which used separate NNs for intensity and differences images with a rule-based integration of the multiple network outputs. The integration of this expectation-based system should improve the
performance of the difference-image NN.
Using Expectation to Guide Processing
865
4 Conclusions
A very closely related procedure to the one described in this paper is the use of Kalman
Filters to predict the locations of objects of interest in the input retina. For example, Dickmanns uses the prediction of the future state to help guide attention by controlling the
direction of a camera to acquire accurate position of landmarks [Dickmanns, 1992].
Strong models of the vehicle motion, the appearance of objects of interest (such as the
road, road-signs, and other vehicles), and the motion of these objects are encoded in the
system. The largest difference in their system and the one presented here is the amount of
a priori knowledge that is used. Many approaches which use Kalman Filters require a
large amount of problem specific information for creating the models. In the approach presented in this paper, the main object is to automatically learn this information from examples. First, the system must learn what the important features are, since no top-down
information is assumed. Second, the system must automatically develop the control strategy from the detected features. Third, the system must also learn a model for the movements of all of the relevant features.
In deciding whether the approaches described in this paper are suitable to a new problem,
two criteria must be considered. First, if expectation is to be used to remove distractions
from the inputs, then given the current inputs, the activations of the relevant inputs in the
next time step must be predictable while the irrelevant inputs are either unrelated to the
task or are unpredictable. In many visual object tracking problems, the relevant inputs are
often predictable while the distractions are not. In the cases in which the distractions are
predictable, if they are unrelated to the main task, these methods can work. When using
expectation to emphasize unexpected or potentially anomalous features, the activations of
the relevant inputs should be unpredictable while the irrelevant ones are predictable. This
is often the case for anomaly/fault detection tasks. Second, when expectations are used as
a filter, it is necessary to explicitly define the role of the expected features. In particular, it
is necessary to define whether the expected features should be considered relevant or irrelevant, and therefore, whether they should be emphasized or de-emphasized, respectively.
We have demonstrated the value of using task-specific expectations to guide processing in
three real-world tasks. In complex, dynamic, environments, such as driving, expectations
are used to quickly and accurately discriminate between the relevant and irrelevant features. For the detection of faults in the plasma-etch step of semiconductor fabrication,
expectations are used to account for the underlying drift of the process. Finally, for visionbased hand-tracking, we have shown that a priori knowledge about expectations can be
easily integrated with a hand-detection model to focus attention on small portions of the
scene, so that distractions in the periphery can be ignored.
Acknowledgments
The author would like to thank Dean Pomerleau, Takeo Kanade, Tom Mitchell and Tomaso Poggio
for their help in shaping this work.
References
Baluja, S. 1996, Expectation-Based Selective Attention. Ph.D. Thesis, School of Computer Science, CMU.
Clark, J. & Ferrier, N (1992), Attentive Visual Servoing, in: Active Vision. Blake & Yuille, (MIT Press) 137-154.
Cottrell, G.W., 1990, Extracting Features from Faces using Compression Network, Connectionist Models, Morgan Kaufmann 328-337.
Dickmanns, 1992, Expectation-based Dynamic Scene Understanding, in: Active Vision. A. Blake & A.Yuille,
MIT Press .
Koch, C. & Ullman, S. (1985) "Shifts in Selective Visual Attention: Towards the Underlying Neural Circuitry",
in: Human Neurobiology 4 (1985) 219-227.
Maxion, R. (1995) The Semiconductor Wafer Plasma-Etch Data Set.
Nowlan, S. & Platt, J., 1995, "A Convolutional Neural Network Hand Tracker". NIPS 7. MIT Press . 901-908.
Pomerleau, D.A., 1993. Neural Network Perception for Mobile Robot Guidance, Kluwer Academic.
Thorpe, C., 1991, Outdoor Visual Navigation for Autonomous Robots, in: Robotics and Autonomous Systems 7.
| 1378 |@word middle:2 eliminating:1 compression:1 proportion:1 simulation:1 accounting:1 fonn:1 deems:1 initial:1 current:9 ferrier:2 nowlan:2 activation:13 must:6 takeo:1 cottrell:2 visible:1 discernible:1 remove:3 half:1 filtered:3 provides:1 detecting:1 location:9 simpler:1 constructed:1 driver:1 inside:1 manner:2 inter:1 expected:10 tomaso:1 distractor:1 automatically:2 actual:9 unpredictable:2 confused:2 underlying:5 matched:1 unrelated:2 what:6 developed:2 perfonn:2 temporal:2 exactly:1 demonstrates:2 scaled:2 wrong:1 control:3 unit:11 platt:2 appear:1 before:2 positive:1 attend:1 shumeet:1 mistake:1 semiconductor:4 limit:1 encoding:2 analyzing:1 subscript:1 solely:2 approximately:2 black:1 examined:2 dynamically:1 limited:2 directed:1 acknowledgment:1 camera:5 testing:1 procedure:2 road:18 cannot:1 close:1 layered:1 put:1 context:1 applying:1 map:6 demonstrated:1 center:1 dean:1 attention:10 cluttered:3 focused:1 x32:1 attending:2 avoidance:2 continued:1 rule:5 his:1 notion:1 autonomous:7 imagine:1 target:6 controlling:1 anomaly:3 mrt:1 us:3 pa:1 located:1 bottom:1 role:3 module:4 region:8 connected:1 movement:2 removed:2 contamination:1 predictable:5 environment:2 dynamic:2 signature:4 trained:14 solving:1 yuille:2 upon:3 easily:3 distinct:1 fast:1 detected:5 neighborhood:1 outside:1 encoded:5 larger:1 solve:5 heuristic:5 reconstruct:2 ability:1 itself:1 sequence:3 reconstruction:11 relevant:12 poorly:1 description:1 moved:1 exploiting:1 produce:1 object:5 help:4 develop:1 measured:2 school:2 strong:1 c:1 predicted:3 indicate:1 differ:1 direction:3 waveform:1 drawback:1 correct:2 closely:1 filter:3 pea:1 centered:3 human:3 require:2 strictly:1 koch:2 around:2 blake:2 considered:4 tracker:5 visually:1 deciding:1 mapping:1 predict:3 circuitry:1 driving:2 purpose:1 estimation:1 travel:1 infonnation:3 highway:1 largest:1 create:2 city:2 gauge:1 weighted:2 mit:3 clearly:1 sensor:2 gaussian:2 modified:3 rather:2 mobile:1 command:1 encode:2 focus:3 improvement:2 contrast:1 nn:12 eliminate:2 entire:2 integrated:1 hidden:18 spurious:1 manipulating:1 reproduce:1 etch:6 selective:2 pixel:5 classification:2 extraneous:4 priori:5 integration:2 once:2 construct:1 sampling:2 manually:1 future:5 connectionist:1 fundamentally:1 thorpe:2 retina:2 simultaneously:1 delayed:1 subsampled:1 ima:1 phase:1 attempt:2 detection:21 interest:3 circular:1 intra:1 navigation:2 light:1 accurate:1 edge:3 necessary:3 poggio:1 allocates:1 perfonnance:3 old:1 desired:2 guidance:1 obstacle:2 measuring:3 subset:1 fabrication:4 conducted:1 connect:1 synthetic:1 nns:2 confident:1 person:3 deemphasizing:1 quickly:1 thesis:1 nm:1 creating:3 ullman:2 account:3 potential:1 tii:1 de:6 pedestrian:1 explicitly:3 ad:1 vehicle:6 traffic:1 portion:7 irre:10 formed:1 ir:1 convolutional:2 kaufmann:1 correspond:1 saliency:5 gathered:2 raw:1 accurately:4 emphasizes:1 produced:1 none:2 detector:1 attentive:1 sampled:1 mitchell:1 knowledge:6 car:5 ut:1 shaping:1 positioned:1 appears:3 feed:2 supervised:1 tom:1 specify:1 improved:3 april:1 done:4 until:1 hand:33 horizontal:2 marker:4 gray:1 effect:1 contain:4 requiring:1 adjacent:1 during:1 criterion:1 distracting:3 motion:2 reflection:1 image:36 dirt:1 began:1 tracked:3 detennining:2 discussed:1 interpretation:2 kluwer:1 interpret:1 mellon:2 significant:1 freeze:1 pointed:1 had:2 reliability:1 moving:2 robot:3 similarity:1 irrelevant:6 driven:1 periphery:1 dickmanns:3 fault:14 seen:1 morgan:1 steering:2 employed:1 signal:2 multiple:3 full:1 reduces:1 match:1 academic:1 cross:4 post:1 schematic:1 prediction:13 anomalous:1 vision:7 expectation:43 cmu:3 achieved:2 robotics:1 navlab:2 background:5 addressed:1 extra:1 unlike:1 hz:1 subject:2 emitted:1 extracting:1 presence:1 noting:1 intermediate:1 visionbased:1 enough:2 variety:1 fit:1 brighter:1 architecture:6 identified:1 reduce:9 idea:1 shift:1 whether:3 passing:2 ignored:2 amount:3 ph:1 processed:1 shifted:1 sign:1 estimated:3 correctly:1 track:2 per:1 carnegie:2 wafer:16 four:7 salient:3 changing:2 timestep:1 fraction:1 sum:1 residential:1 ascertaining:1 run:1 place:1 throughout:1 coherence:2 layer:17 yielded:1 ahead:2 precisely:1 worked:1 constraint:2 scene:9 lane:17 marking:9 according:1 alternate:1 developing:1 across:1 ascertain:2 reconstructing:1 slightly:1 wavefonn:4 making:1 previously:1 turn:1 fail:1 mechanism:1 needed:1 loxlo:1 fed:1 available:2 detennine:3 gaussians:1 away:1 appropriate:2 chamber:1 original:2 top:4 strategy:1 distance:1 separate:3 unable:2 thank:1 capacity:1 street:1 landmark:1 kalman:2 providing:1 acquire:1 difficult:4 october:1 potentially:1 pomerleau:6 situation:1 extended:1 variability:1 neurobiology:1 unanticipated:1 frame:4 intensity:2 drift:1 required:1 trainable:1 connection:3 learned:3 nip:1 able:1 perception:1 appeared:2 perfonned:1 suitable:1 indicator:1 improve:1 inversely:1 created:3 auto:2 text:1 understanding:1 fully:1 expect:1 allocation:1 mounted:1 clark:2 land:1 heavy:1 row:2 translation:1 summary:3 accounted:1 last:2 free:1 guide:6 side:3 perceptron:1 neighbor:1 face:2 feedback:2 dimension:1 world:6 transition:3 forward:2 made:1 collection:1 fonned:1 avoided:1 author:1 far:1 reconstructed:4 emphasize:3 keep:1 active:2 pittsburgh:2 assumed:1 triplet:2 table:2 promising:1 learn:5 transfer:1 kanade:1 complex:1 domain:6 did:2 main:2 justsystem:1 noise:7 body:2 downwards:1 predictability:1 position:7 outdoor:1 weighting:1 third:2 down:3 emphasizing:1 specific:12 emphasized:7 explored:1 incorporating:1 false:4 importance:1 magnitude:1 wavelength:1 appearance:1 visual:5 unexpected:5 contained:2 tracking:11 corresponds:1 goal:2 marked:2 ann:1 towards:3 baluja:8 determined:1 typical:1 principal:1 degradation:1 total:1 discriminate:1 plasma:5 attempted:1 selectively:1 select:1 distraction:8 categorize:1 relevance:3 incorporate:2 tested:1 |
415 | 1,379 | Shared Context Probabilistic Transducers
Yoshua Bengio*
Dept. IRO ,
Universite de Montreal,
Montreal (QC) , Canada, H3C 3J7
bengioyOiro.umontreal.ca
Samy Bengio t
Microcell Labs,
1250 , Rene Levesque Ouest,
Montreal (QC) , Canada, H3B 4W8
samy.bengioOmicrocell.ca
Jean-Fran~ois Isabelle t
Microcell Labs,
1250, Rene Levesque Ouest ,
Montreal (QC), Canada, H3B 4W8
jean-francois.isabelleCmicrocell.ca
Yoram Singer
AT&T Laboratories,
Murray Hill , NJ 07733, USA,
singerOresearch.att.com
Abstract
Recently, a model for supervised learning of probabilistic transducers represented by suffix trees was introduced. However, this algorithm tends to build very large trees, requiring very large amounts
of computer memory. In this paper, we propose anew, more compact, transducer model in which one shares the parameters of distributions associated to contexts yielding similar conditional output
distributions . We illustrate the advantages of the proposed algorithm with comparative experiments on inducing a noun phrase
recogmzer .
1
Introduction
Learning algorithms for sequential data modeling are important in many applications such as natural language processing and time-series analysis, in which one has
to learn a model from one or more sequences of training data. Many of these algorithms can be cast as weighted transducers (Pereira, Riley and Sproat, 1994), which
associate input sequences to output sequences, with weights for each input/output
* Yoshua Bengio is also with AT&T Laboratories, Holmdel, NJ 07733, USA.
t This work was performed while Samy Bengio was at INRS-Telecommunication, Iledes-Soeurs, Quebec, Canada, H3E IH6
t This work was performed while Jean-Franc;ois Isabelle was at INRS-Telecommunication, Ile-des-Soeurs, Quebec, Canada, H3E IH6
y. Bengio, S. Bengio, J-F. Isabelle and Y. Singer
410
sequence pair. When these weights are interpreted as probabilities, such models are
called probabilistic transducers. In particular, a probabilistic transducer can represent the conditional probability distribution of output sequences given an input
sequence. For example, algorithms for combining several transducers were found
useful in natural language and speech processing (Riley and Pereira, 1994). Very
often, weighted transducers use an intermediate variable that represents "context",
such as the state variable of Hidden Markov Models (Baker, 1975; Jelinek, 1976). A
particular type of weighted transducer, called Input/Output Hidden Markov Model,
is one in which the input-to-context distribution and context-to-output distribution
are represented by flexible parameterized models (such as neural networks) (Bengio
and Frasconi, 1996) . In this paper, we will study probabilistic transducers with a
deterministic input-to-state mapping (i.e., a function from the past input subsequence to the current value of the context variable). One such transducer is the one
which assigns a value of the context variable to every value of the past input subsequence already seen in the data. This input-to-state mapping can be efficiently
represented by a tree. Such transducers are called suffix tree transducers (Singer,
1996).
A problem with suffix tree transducers is that they tend to yield very large trees
(whose size may grow as O(n 2 ) for a sequence of data of length n). For example, in
the application studied in this paper, one obtains trees requiring over a gigabyte of
memory. Heuristics may be used to limit the growth of the tree (e.g., by limiting
the maximum depth of the context, i.e., of the tree, and by limiting the maximum
number of contexts, i.e., nodes of the tree). In this paper, instead, we propose a
new model for a probabilistic transducer with deterministic input-to-state function
in which this function is compactly represented, by sharing parameters of contexts
which are associated to similar output distributions. Another way to look at the
proposed algorithm is that it searches for a clustering of the nodes of a suffix tree
transducer. The data structure that represents the contexts is not anymore a tree
but a single-root acyclic directed graph.
2
Background: Suffix Tree Probabilistic Transducers
The learning algorithm for suffix tree probabilistic transducers (Singer, 1996) constructs the model P(Yilxi) from discrete input sequences xi
{Xl,X2, ... ,Xn }
to output sequences yi = {Y1, Y2, ... , Yn}, where Xt are elements of a finite alphabet
Ein. This distribution is represented by a tree in which each internal node may have
a child for every element of Ein, therefore associating a label E Ein to each arc. A
node at depth d is labeled with the sequence ut of labels on arcs from root to node,
corresponding to a particular input context, e.g., at some position n in the sequence
a context of length d is the value ut of the preceding subsequence x~_d l' Each
node at depth d is therefore associated with a model of the output distribution in
this context, P(Yn IX~-d+1 = ut) (independent of n).
=
To obtain a local output probability for Yn (i.e., given xi), one follows the longest
possible path from the root to a node a depth d according to the labels x n , xn -1,
... Xn-d+1. The local output probability at this node is used to model Yn' Since
p(yflxf) can always be written n~=1 P(Yn Ixi)' the overall input/output conditional distribution can be decomposed, according to this model, as follows:
T
p(yflxf)
= II P(YnIX~_d(x~)+l)'
(1)
n=1
where d(xi) is the depth of the node of the tree associated with the longest suffix
uf = x~_d+1 of xi. Figure 1 gives a simple example of a suffix tree transducer.
411
Shared Context Probabilistic Transducers
Pfaj"O.S
Pfb)~OJ
P(c)", 0 2
p(a)",O.6
p(b) _O.)
p(c:)= O I
11
Figure 1: ExampJe of suffix tree transducer (Singer, 1996). The input alphabet,
E in = {O, I} and the output aJphabet, EotJt = {a, b, c}. For instance, P(aIOOllO) =
P(alllO) = 0.5 .
3
Proposed Model and Learning Algorithm
In the model proposed here, the input/output conditional distribution p(yT IxI)
is represented by a single-root acyclic directed graph. Each node of this graph is
associated with a set of contexts C node = {(jt?}, corresponding to all the paths i
(of various lengths di ) from the root of the tree to this node. All these contexts are
associated with the same local output distribution P(Yn Ix? has a suffix in Cnode).
Like in suffix tree transducers, each internal node may have a child for every element
of E in . The arc is labeled with the corresponding element of ~in . Also like in suffix
tree transducers, to obtain P(Ynlx~), one follows the path from the root to the
deepest node called deepest(x?) according to the labels Xn, Xn-l, etc .. . The local
output distribution at this node is used to predict Yn or its probability. The overall
conditional distribution is therefore given by
T
P(yilxf) =
II P(Ynldeepest(x~))
(2)
n=l
where the set of contexts Cdeepe3t(x~) associated to the deepest node deepest(xl)
contains a suffix of x? The model can be used both to compute the conditional
probability of a given input/output sequence pair, or to guess an output sequence
given an input sequence. Note that the input variable can contain delayed values
of the output variable (as in Variable Length Markov Models).
3.1
Proposed Learning Algorithm
We present here a constructive learning algorithm for building the graph of the
model and specify which data points are used to update each local output model
(associated to nodes of the graph). The algorithm is on-line and operates according to two regimes: (1) adding new nodes and simply updating the local output
distributions at existing nodes, and (2) merging parts of the graph which represent
similar distributions. If there are multiple sequences in the training data they are
concatenated in order to obtain a single input/output sequence pair.
(1) After every observation (xn, Yn), the algorithm updates the output distributions
y. Bengio, S. Bengio, J-F. Isabelle and Y. Singer
412
of the nodes for which Cnode(x~) contains a suffix of Xl, possibly adding new nodes
(with labels x~_d.) until xl E Cnode for some node.
(2) Every Tmerge observations, the algorithm attempts to merge sub-graphs which
are found similar enough, by comparing the N (N - 1) /2 pairs of sub-graphs rooted
at the N nodes that have seen at least minn observations. Merging two subgraphs
is equivalent to forcing them to share parameters (as well as reducing the size of
the representation of the distribution). A merge is performed between the graphs
rooted at nodes a and b if Ll(a, b) < mina and the merge succeeds. The details of
the similarity measure and merging algorithm are given in the next subsections.
3.2
Similarity Measure Between Rooted Subgraphs
In order to compare (asymmetrically) output distributions P(yla) and P(ylb) at
two nodes a and b, one can use the Kullback-Liebler divergence:
_
~
P(ylb)
(3)
Ii. L(a, b)
L...J P(ylb) log P(yla)
=
yEE out
However, we want to compare the whole acyclic graphs rooted at these 2 nodes. In
order to do so, let us define the following. Let s be a string of input labels, and b
a node. Define desc(b, s) as the most remote descendant of b obtained by following
from b the arcs whose labels correspond to the sequence s. Let descendents(a) be
the set of strings obtained by following the arcs starting from node a until reaching
the leaves which have a as an ancestor. Let P(sla) be the probability offollowing the
arcs according to string s, starting from node a. This distribution can be estimated
by counting the relative number of descendents through each of the children of each
node.
To compare the graphs rooted at two nodes a and b, we extend the KL divergence
by weighing each of the descendents of a, as follows:
L
W K L(a, b) =
P(sla)K L(desc(a, s), desc(b, s))
(4)
de3cendent8 (a)
Finally, to obtain a symmetric measure, we define
Ll(a,b) = WKL(a,b) + WKL(b,a)
(5)
that is used in the merge phase of the constructive learning algorithm to decide
whether the subgraphs rooted at a and b should be merged.
3
3.3
E
Merging Two Rooted Subgraphs
If Ll (a, b) < mina (a predefined threshold) we want to merge the two subgraphs
rooted at a and b and create a new subgraph rooted at c. The local output distribution at c is obtained from the local output distributions at a and b as follows:
P(Yn Ic) = P(Ynla)P(ala or b)
where we define
+ P(Ynlb)P(bla or b)
ad(a)
P(ala or b) = ad(a) + ad(b) ,
(6)
(7)
where d(a) is the length ofthe longest path from the root to node a, and a represents
a prior parameter (between 0 and 1) on the depth of the acyclic graphs . This prior
parameter can be used to induce a prior distribution over possible rooted acyclic
graphs structures which favors smaller graphs and shorter contexts (see the mixture
of probabilistic transducers of (Singer, 1996)).
The merging algorithm can then be summarized as follows:
Shared Context Probabilistic Transducers
413
? The parents of a and b become parents for c.
? Some verifications are made to prevent merges which would yield to cycles in the
graph. The nodes a and b are not merged if they are parents of one another.
? We make each child of a a child of c. For each child u of b (following an arc labeled
l), look for the corresponding child v of c (also following the arc labeled l) . If there
is no such child, and u is not a parent of c, make u a new child of c. Else, if u and
v are not parents of each other, recursively merge them.
? Delete nodes a and b, as well as all the links from and to these nodes.
This algorithm is symmetric with respect to a and b except when a merge cannot
be done because a and b are parents of one another. In this case, an asymmetric
decision must be taken: we chose to keep only a and reject b. Figure 2 gives a
simple example of merge.
Figure 2: This figure shows how two nodes are merged. The result is no longer a
tree, but a directed graph. Some verifications are done to avoid cycles in the graph.
Each node can have multiple labels, corresponding to the multiple possible paths
from the root to the node.
4
Comparative Experiments
We compared experimentally our model to the one proposed in (Singer, 1996) on
mixtures of suffix tree transducers, using the same task. Given a text where each
word is assigned an appropriate part-of-speech value (verb, noun, adjective, etc),
the task is to identify the noun phrases in the text. The UPENN tree-bank corpus
database was used in these experiments. The input vocabulary size, IEinl
41,
is the number of possible part-of-speech tags, and the output vocabulary size is
IEouti = 2. The model was trained over 250000 marked tags, constraining the
tree to be of maximal depth 15. The model was then tested (freezing the model
structure and its parameters) over 37000 other tags. Using the mixture of suffix tree
transducers (Singer, 1996) and thresholding the output probability at 0.5 to take
output decisions, yielded an accuracy rate of 97.6% on the test set, but required
over 1 gigabyte of computer memory.
=
To make interesting comparisons with the shared context transducers, we chose
the following experimental scheme. Not only did we fix the maximal depth of the
directed graph to 15, but we also fixed the maximal number of allocated nodes, i.e.,
simulating fixed memory resources. When this number was reached, we froze the
structure but continued to update the parameters of the model until the end of the
training database was reached. For the shared context version, whenever a merge
freed some nodes, we let the graph grow again to its maximal node size. At the end
of this process, we evaluated the model on the test set.
Y. Bengio, S. Bengio, J-F. Isabelle and Y. Singer
414
We tested this method for various values of the maximum number of nodes in the
graph. For each experiment, we tried different values of the other parameters (the
similarity threshold min~ for merging, the minimum number of observations miIln
at a node before it can be considered for a merge, and the delay Tmerge between two
merging phases), and we picked the one which performed the best on the training
set. Results are reported in figure 3.
maximal
number
of nodes
20
50
100
500
1000
2000
5000
with
merge
(%)
0.762
0.827
0.861
0.924
0.949
0.949
0.952
without
merge
(%)
0.584
0.624
0.727
0.867
0.917
0.935
0.948
005
09
085
01
075
07
085
01
055
Figure 3: This figure shows the generalization accuracy rate of a transducer with
merges (shared contexts graph) against one without merges (suffix tree), with different maximum number of nodes. The maximum number of nodes are in a logarithmic scale, and the accuracy rates are expressed in relative frequency of correct
classification.
As can be seen from the results, the accuracy rate over the test set is better for
transducers with shared contexts than without. More precisely, the gain is greater
when the maximum number of nodes is smaller. When we fix the maximum number
of nodes to a very small value (20), a shared context transducer performs 1.3 times
better (in classification error) than a non-shared one. This gain becomes smaller
and smaller as the maximum size increases. Beyond a certain maximum size, there
is almost no gain, and one could probably observe a loss for some large sizes. We
also need to keep in mind that the larger the transducer is, the slower the program
to create the shared context transducer is, compared to the non-shared one. Finally,
it is interesting to note that using only 5000 nodes, we were able to obtain 95.2%
accuracy, which is only 2.4% less than those obtained with no constraint on the
number of nodes.
5
Conclusion
In this paper, we have presented the following:
? A new probabilistic model for probabilistic transducers with deterministic
input-to-state function, represented by a rooted acyclic directed graph with
nodes associated to a set of contexts and children associated to the different
input symbols. This is a generalization of the suffix tree transducer.
? A constructive learning algorithm for this model, based on construction and
merging phases. The merging is obtained by clustering parts of the graph
which represent a similar conditional distribution.
? Experimental results on a natural-language task showing that when the size
of the graph is constrained, this algorithm performs better than the purely
constructive (no merge) suffix tree algorithm.
Shared Context Probabilistic Transducers
4}5
References
Baker, J. (1975). Stochastic modeling for automatic speech understanding. In
Reddy, D., editor, Speech Recognition, pages 521-542. Academic Press, New
York.
Bengio, S. and Bengio, Y. (1996). An EM algorithm for asynchronous input/output
hidden markov models. In Proceedings of the International Conference on Neural Information Processing, Honk Kong .
Bengio, Y. and Frasconi, P. (1996). Input/Output HMMs for sequence processing.
IEEE Transactions on Neural Networks , 7(5):1231-1249.
Jelinek, F. (1976). Continuous speech recognition by statistical methods. Proceedings of the IEEE, 64:532-556 .
Pereira, F. , Riley, M., and Sproat, R. (1994). Weighted rational transductions and
their application to human language processing. In ARPA Natural Language
Processing Workshop.
Riley, M. and Pereira, F. (1994). Weighted-finite-automata tools with applications
to speech and language processing. Technical Report Technical Memorandum
11222-931130-28TM, AT&T Bell Laboratories.
Singer, Y . (1996). Adaptive mixtures of probabilistic transducers. In Mozer, M.,
Touretzky, D., and Perrone, M., editors, Advances in Neural Information Processing Systems 8. MIT Press, Cambridge, MA.
| 1379 |@word kong:1 version:1 tried:1 recursively:1 series:1 att:1 contains:2 ala:2 past:2 existing:1 current:1 com:1 comparing:1 written:1 must:1 update:3 leaf:1 guess:1 weighing:1 node:51 become:1 transducer:37 descendant:1 upenn:1 decomposed:1 becomes:1 baker:2 interpreted:1 string:3 nj:2 w8:2 every:5 growth:1 yn:9 before:1 local:8 tends:1 limit:1 path:5 merge:13 chose:2 studied:1 hmms:1 directed:5 bla:1 bell:1 reject:1 word:1 induce:1 cannot:1 context:28 yee:1 equivalent:1 deterministic:3 yt:1 starting:2 automaton:1 qc:3 assigns:1 subgraphs:5 continued:1 gigabyte:2 memorandum:1 limiting:2 construction:1 ixi:2 samy:3 associate:1 element:4 recognition:2 updating:1 asymmetric:1 labeled:4 database:2 cycle:2 remote:1 mozer:1 trained:1 purely:1 compactly:1 represented:7 various:2 alphabet:2 jean:3 whose:2 heuristic:1 larger:1 favor:1 h3c:1 advantage:1 sequence:18 propose:2 maximal:5 combining:1 subgraph:1 inducing:1 parent:6 francois:1 comparative:2 illustrate:1 montreal:4 ouest:2 ois:2 merged:3 correct:1 honk:1 stochastic:1 human:1 pfb:1 fix:2 generalization:2 desc:3 considered:1 ic:1 mapping:2 predict:1 yilxi:1 label:8 create:2 tool:1 weighted:5 ynix:1 mit:1 j7:1 always:1 reaching:1 avoid:1 longest:3 suffix:19 hidden:3 ancestor:1 overall:2 classification:2 flexible:1 noun:3 constrained:1 construct:1 frasconi:2 represents:3 look:2 yoshua:2 report:1 franc:1 divergence:2 delayed:1 phase:3 attempt:1 inr:2 mixture:4 yielding:1 predefined:1 shorter:1 tree:29 delete:1 froze:1 instance:1 arpa:1 modeling:2 phrase:2 riley:4 delay:1 reported:1 international:1 probabilistic:15 again:1 possibly:1 de:2 summarized:1 descendent:3 ad:3 performed:4 root:8 picked:1 lab:2 reached:2 accuracy:5 efficiently:1 yield:2 correspond:1 ofthe:1 identify:1 cnode:3 liebler:1 touretzky:1 sharing:1 whenever:1 against:1 frequency:1 universite:1 associated:10 di:1 gain:3 rational:1 subsection:1 ut:3 supervised:1 specify:1 done:2 evaluated:1 until:3 freezing:1 usa:2 building:1 requiring:2 y2:1 contain:1 assigned:1 symmetric:2 laboratory:3 ll:3 rooted:11 mina:2 hill:1 performs:2 recently:1 umontreal:1 extend:1 isabelle:5 rene:2 cambridge:1 automatic:1 language:6 similarity:3 longer:1 etc:2 forcing:1 certain:1 yi:1 seen:3 minimum:1 greater:1 preceding:1 ii:3 multiple:3 technical:2 academic:1 ylb:3 ile:1 represent:3 background:1 want:2 else:1 grow:2 allocated:1 probably:1 tend:1 quebec:2 counting:1 intermediate:1 bengio:14 sproat:2 enough:1 constraining:1 associating:1 tm:1 whether:1 speech:7 york:1 useful:1 amount:1 estimated:1 discrete:1 threshold:2 sla:2 prevent:1 freed:1 graph:24 parameterized:1 telecommunication:2 almost:1 decide:1 fran:1 decision:2 holmdel:1 yielded:1 precisely:1 constraint:1 x2:1 tag:3 min:1 ih6:2 uf:1 according:5 perrone:1 smaller:4 em:1 taken:1 resource:1 reddy:1 singer:11 mind:1 end:2 observe:1 appropriate:1 simulating:1 anymore:1 slower:1 clustering:2 yoram:1 concatenated:1 murray:1 build:1 already:1 link:1 iro:1 length:5 minn:1 observation:4 markov:4 arc:8 finite:2 y1:1 wkl:2 yla:2 verb:1 canada:5 introduced:1 cast:1 pair:4 kl:1 required:1 merges:3 beyond:1 able:1 regime:1 program:1 adjective:1 oj:1 memory:4 natural:4 scheme:1 text:2 prior:3 understanding:1 deepest:4 relative:2 loss:1 interesting:2 acyclic:6 verification:2 thresholding:1 editor:2 bank:1 share:2 asynchronous:1 jelinek:2 depth:8 xn:6 vocabulary:2 made:1 adaptive:1 transaction:1 compact:1 obtains:1 kullback:1 keep:2 anew:1 corpus:1 xi:4 subsequence:3 search:1 continuous:1 learn:1 ca:3 did:1 whole:1 child:10 transduction:1 ein:3 levesque:2 position:1 pereira:4 sub:2 xl:4 ix:2 xt:1 jt:1 showing:1 symbol:1 workshop:1 sequential:1 adding:2 merging:9 logarithmic:1 simply:1 expressed:1 ma:1 conditional:7 marked:1 shared:12 experimentally:1 except:1 operates:1 reducing:1 called:4 asymmetrically:1 experimental:2 succeeds:1 internal:2 constructive:4 dept:1 tested:2 |
416 | 138 | 402
MODELING THE OLFACTORY BULB
- COUPLED NONLINEAR OSCILLATORS
Zhaoping Lit
J. J. Hopfield?
t Division of Physics, Mathematics and Astronomy
?Division of Biology, and Division of Chemistry and Chemical Engineering
t? California Institute of Technology, Pasadena, CA 91125, USA
? AT&T Bell Laboratories
ABSTRACT
The olfactory bulb of mammals aids in the discrimination of
odors. A mathematical model based on the bulbar anatomy and
electrophysiology is described. Simulations produce a 35-60 Hz
modulated activity coherent across the bulb, mimicing the observed
field potentials. The decision states (for the odor information)
here can be thought of as stable cycles, rather than point stable
states typical of simpler neuro-computing models. Analysis and
simulations show that a group of coupled non-linear oscillators are
responsible for the oscillatory activities determined by the odor input, and that the bulb, with appropriate inputs from higher centers,
can enhance or suppress the sensitivity to partiCUlar odors. The
model provides a framework in which to understand the transform
between odor input and the bulbar output to olfactory cortex.
1. INTRODUCTION
The olfactory system has a simple cortical intrinsic structure (Shepherd 1979),
and thus is an ideal candidate to yield insight on the principles of sensory information processing. It includes the receptor cells, the olfactory bulb, and the olfactory
cortex receiving inputs from the bulb (Figure [1]). Both the bulb and the cortex
exhibit similar 35-90 Hz rhythmic population activity modulated by breathing. Efforts have been made to model the bulbar information processing function (Freeman
1979b, 1979c; Freeman and Schneider 1982; Freeman and Skarda 1985; Baird 1986j
Skarda and Freeman 1987), which is still unclear (Scott 1986). The bulbar position
in the olfactory pathway, and the linkage of the oscillatory activity with the sniff
cycles suggest that the bulb and the oscillation play important roles in the olfactory
information processing. We will examine how the bulbar oscillation pattern, which
can be thought of as the decision state about odor information, originates and how
it depends on the input odor. We then show that with appropriate inputs from the
higher centers, the bulb can suppress or enhance the its sensitivity to particular
odors. Much more details of our work are described in other two papers (Li and
Hopfield 1988a, 1988b).
403
Modeling the Olfactory Bulb-Coupled Nonlinear Oscillators
The olfactory bulb has mainly the excitatory mitral and the inhibitory granule
cells located on different parallel lamina. Odor receptors effectively synapse on the
mitral cells which interact locally with the granule cells and carry the bulbar outputs
(Fig 1, Shepherd 1979). A rabbit has about 50,000 mitral, and
10,000,000 granule
cells (Shepherd 1979). With short odor pulses, the receptor firing rate increases in
time, and terminates quickly after the odor pulse terminates (Getchell and Shepherd
1978). Most inputs from higher brain centers are directed to the granule cells, and
little is know about them. The surface EEG wave (generated by granule activities,
Freeman 1978j Freeman and Schneider 1982), depending on odor stimulations and
animal motivation, shows a high amplitude oscillation arising during the inhalation
and stopping early in the exhalation. The oscillation is an intrinsic property of the
bulb itself, and is influenced by central inputs (Freeman 1979aj Freeman and Skarda
1985). It has a peak frequency (which is the same across the bulb) in the range of
35-90 Hz, and rides on a slow background wave phase locked with the respiratory
wave.
"'-J
2. MODEL ORGANIZATION
we only include ( N excitatory) mitral and ( M
For simplicity,
inhibitory )
Iodor" + Ibaekgrotmd,i,
granule cells in the model. The Receptor input I is Ii
for 1 ... , N, a superposition of an odor signal Iodor and a background input
Ibaekgrov,nd. Iodor > 0 increases in time during inhalation, and return exponentially during exhalation toward the ambient. The central input to the granule
cells is vector Ie with components Ie,j for 1 < j < M. For now, it is assumed that
Ie = 0.1 and Ibaekgrov,nd = 0.243 do not change during a sniff (Li and Hopfield
1988a).
=
Each cell is one unit with its internal state level described by a single variable, and its output a continuous function of the internal state level. The internal states and outputs are respectively X = {Xl' X2, ... , XN} and G z ( X) =
{gz(xd, gz(x2),?.? ,gz(XN)} (Y = {YI' Y2,??? ,YM} and Gy{Y) = {gy(Yd,
gy(Y2), ... ,gy(YM)}) for the mitral (granule) cells, where gz > 0 and gy > 0 are
the neurons' non-linear sigmoid output functions essential for the bulbar oscillation
dynamics (Freeman and Skarda 1985) to be studied.
Receptor inputs
other
brain
... ... ... ... ... ... ... ...
(t)(t)(t)(t)(t)(t)(t)(t)
It' t ~
,
+)IfInt~~:~lion
22222222
It.
It.
It. It. It. It.
Central inputs
It.
...
Fig.l. Left: olfactory system; Right: bulbar structure
Cells marked "+" are mitral cells, "_" are granule cells
The geometry of bulbar structure is simplified to a one dimensional ring. Each
cell is specified by an index, e.g. ith mitral cell, and jth granule cell for all i,
i
404
Li and Hopfield
indicating cell locations on the ring (Fig 1). N X M matrix Ho and M X N
matrix Wo are used respectively to describe the synaptic strengths (postsynaptic
input: presynaptic output) from granule cells to mitral cells and vice versa. The
bulb model system has equations of motion:
X=
Y
-HoGy(Y) -
= WoGz(X) -
O:zX
+ I,
(2.1)
O:yY + Ie.
where O:z = 1/'rz , O:y = 1/'ry, and 'rz = 'ry = 7 msec are the time constants
of the mitral and granule cells respectively (Freeman and Skarda 1985; Shepherd
1988). In simulation, weak random noise is added to I .and Ie to simulate the
fluctuations in the system.
3. SIMULATION RESULT
Computer simulation was done with 10 mitral and granule cells, and show
that the model can capture the major effects of the real bulb. The rise and fall
of oscillations with input and the baseline shift wave phase locked with sniff cycles
are obvious (Fig.2). The simulated EEG (calculated using the approximation by
Freeman (1980)) and the measured EEG are shown for comparison. During a sniff,
all the cells oscillate coherently with the same frequency as physiologically observed.
7\
.'""1
B
GrwIule Output
~J1 .0
~EEGW'"
lOOms
H
EEG Wave
Respiratory Wave
looms
~
Fig.2. A: Simulation result; B: measured result from Freeman and Schneider 1982.
The model also shows the capability of a pattern classifier. During a sniff,
some input patterns induce oscillation, while others do not, and different inputs
induce different oscillation patterns. We showed (Li and Hopfield 1988a) that the
bulb amplifies the differences between the different inputs to give different output
patterns, while the responses to same odor inputs with different noise samples differ
negligibly.
4. MATHEMATICAL ANALYSIS
A (damped) oscillator with frequency w can be described by the equations
X = -wy - o:x
iI
= wx -
o:y
or
(4.1)
Modeling the Olfactory Bulb-Coupled Nonlinear Oscillators
The solution orbit in (x, y) space is a circle if a = 0 (non-damped oscillator),
and spirals into the origin otherwise (damped oscillator). IT a mitral cell and a
granule cell are connected to each other, with inputs i(t) and ie(t) respectively,
then
x = -h . gy(y) - azx + i(t),
y = w . gz(x) - ayy + ie(t).
(4.2)
This is the scalar version of equation (2.1) with each upper case letter representing
a vector or matrix replaced by a lower case letter representing a scalar. It is assumed that i(t) has a much slower time course than X or y (frequency of sniffs ~
characteristic neural oscillation frequency). Use the adiabatic approximation, and
define the equilibrium point (xo, Yo) as
Xo ~ 0
Yo ~ 0
Define x'
=
x -
Xo,
= -h . gy(yo) - azxo + i,
= w . gz(x o) - ayyo + ie'
y' - y - Yo' Then
x' = -h(gy(y) il = w(gz(x) (cf. equation (4.1)). IT a z
=f
gy(yo)) - azx',
gz(x o)) - ayy'.
= ay = 0, then the solution orbit
zo+z'
R
(4.3)
yo+Y'
w(gz(s) - gz(xo))ds +
Zo
f
h(gy(s) - gy(Yo))ds
= constant
Yo
is a closed curve in the original (x, y) space surrounding the point (xo, Yo), i.e.,
(x, y) oscillates around the point (xo, Yo). When the dissipation is included,
dR/ dt < 0, the orbit in (x, y) space will spiral into the point (xo, Yo). Thus
a connected pair of mitral and granule cells behaves as a damped non-linear oscillator, whose oscillation center (xo, Yo) is determined by the external inputs i and ie'
For small oscillation amplitudes, it can be approximated by a sinusoidal oscillator
via linearization around the (xo, Yo):
x = -h . g~(yo)Y - azx
iI = w . g~(xo)x - a 1l y
(4.4)
where (x, y) is the deviation from (xo, Yo . The solution is X
where a
= (az + a y)/2 and w =
hwg~(xo)g~(yo)
= Toe-at sin(wt+<p)
+ (a z - a y)2/4.
IT az =
a y, which is about right in the bulb, w = Jhwg~(xo)g~(yo). For the bulb,
a ~ 0.3w. The oscillation frequency depends on the synaptic strengths hand w,
and is modulated by the receptor and central input via (xo, Yo).
405
406
Li and Hopfield
N such mitral-granule pairs with cell interconnections between the pairs represent a group of N coupled non-linear damped oscillators. This is exactly the
situation in the olfactory bulb. The locality of synaptic connections in the bulb implies that the oscillator coupling is also local. (That there are many more granule
cells than mitral cells only means that there is more than one granule cell in each
oscillator.) Corresponding to equation (4.2) and (4.4), we have equation (2.1) and
,
.
X = -HoGy(Yo)Y - o.zX
y
=
WoG~(Xo)X - ayY
= -HY -
o.zX,
=
(4.5)
WX - o.yY.
where (X, Y) are now deviations from (Xo, Yo) and G~(Xo) and G~(Yo) are
diagonal matrices with elements: [G~(Xo)lii = g~(Xi,o)
g~(Yilo)
>
>
0, [G~(Yo)lii =
0, for all i,j. Eliminating Y,
(4.6)
=
where A
HW
the equation
Xi
= HoG~(Yo)WoG~(Xo).
The ith oscillator (mitral cell) follows
+ (o.z + o.y)Xi + (Aii + o.zo.y)Xi + L
AijXj = 0
(4.7)
jt.i
(cf.equation (4.1)), the the last term describes the coupling between oscillators.
Non-linear effect occurs when the amplitude is large, and make the oscillation wave
form non-sinusoidal.
If X k is one of the eigenvectors of A with eigenvalue Ak, equation (4.6) has
kth oscillation mode
Components of Xk indicate oscillators' relative amplitudes and phases (for each
k = 1,2, ... , N independent mode). For simplicity, we set 0.2: = 0.1/ = 0.,
then X ex: Xke-at?i../X,.t. Each mode has frequency Re~k' where Re means
the real part of a complex number. If Re( -0. ? i~k) > 0 is satisfied for
some k, then the amplitude of the kth mode will increase with time, i.e. growing
oscillation. Starting from an initial condition of arbitrary small amplitudes in linear
analysis, the mode with the fastest growing amplitude will dominate the output,
and the whole bulb will oscillate in the same frequency as observed physiologically
(Freeman 1978; Freeman and Schneider 1982) as well as in the simulation. With the
non-linear effect, the strongest mode will suppress the others, and the final activity
output will be a single "mode" in a non-linear regime.
Modeling the Olfactory Bulb-Coupled Nonlinear Oscillators
Because of the coupling between the (damped) oscillators, the equilibrium point
(Xo, Yo) of a group of oscillators is no longer always stable with the possibility
of growing oscillation modes. Ak must be complex in order to have kth mode
grow. For this, a necessary (but not sufficient) condition is that matrix A is nonsymmetric. Those inputs that make matrix A less symmetric will more likely induce
the oscillatory output and thus presumpably be noticed by the following olfactory
cortex (see Li and Hopfield 1988a for details).
The consequences (also observed physiologically) of our model are (Freeman
1975,1978; Freeman and Schneider 1982; Li and Hopfield 1988a): 1): local mitral
cells' oscillation phase leads that of the local granule cells by a quarter cycle; 2):
oscillations across the bulb have the same dominant frequency whose range possible
should be narrow; 3): there should be a non-zero phase gradient field across the
bulb; 4): the oscillation activity will rise during the inhale and fall at exhale, and
rides on a slow background baseline shift wave phase locked with the sniff cycles.
This model of the olfactory bulb can be generalized to other masses of interacting excitatory and inhibitory cells such as those in olfactory cortex, neocortex
and hippocampus (Shepherd 1979) etc. where there may be connections between
the excitatory cells as well as the inhibitory cells (Li and Hopfield 1988a). Suppose
that Bo and Co are excitatory-to-excitatory and inhibitory-to-inhibitory connection matrices respectively, then equation (4.6) becomes
x + (az -
B
+ a y + C)k + (A + (a z - B)(a y + C))X = 0
(4.9)
5. COMPUTATIONS IN THE OLFACTORY BULB
Receptor input I influences (Xo, Yo) as follows
dXo ~ (a 2 + HW)-l(adI + di)
dYo ~ (a 2 + W H)-l(W dI - aH-1di)
(5.1)
This is how the odor input determines the bulbar output. Increasing Iodor not only
raises the mean activity level (Xo, Yo) (and thus the gain (G~(Xo), G~(Yo))), but
also slowly changes the oscillation modes by structurally changing the oscillation
equation (4.6) through matrix A = HoG~(Yo)WoG~(Xo). If (Xo, Yo) is raised
to such an extent that Re( -a?iv'Ak) > 0 is satisfied for some mode k, the equilibrium point (Xo, Yo) becomes unstable and this mode emerges with oscillatory
bursts. Different oscillation modes that emerge are indicative of the different odor
inputs controlling the system parameters (Xo, Yo), and can be thought of as the
decision states reached for odor information, i.e., the oscillation pattern classifies
odors. When (Xo, Yo) is very low (e.g. before inhale), all modes are damped, and
only small amplitUde oscillations occur, driven by noise and the weak time variation
of the odor input. The absence of oscillation can be interpreted by higher processing
407
408
Li and Hopfield
centers as the absence of an odor (Skarda and Freeman 1987). Detailed analysis
shows how the bulb selectively responds (or not to respond) to certain input patterns (Li and Hopfield 1988a) by choosing the synaptic connections appropriately.
This means the bulb can have non-uniform sensitivities to different odor receptor
inputs and achieve better odor discriminations.
6. PERFORMANCE OPTIMIZATION IN THE BULB
We discussed (Li and Hopfield 1988a) how the olfactory bulb makes the least
information contamination between sniffs and changes the motivation level for odor
discrimination. We further postulate with our model that the bulb, with appropriate
inputs from the higher centers, can enhance or suppress the sensitivity to particular
odors (details in Li and Hopfield 1988b). When the central input Ie is not fixed, it
can control the bulbar output by shifting (Xo, Yo), just as the odor input I can,
equation (5.1) becomes:
dXo ~ (a 2 + HW)-l(adI + di - HdIe + aW-ldic)
dYo ~ (a 2 + W H)-l(W dI - aH-ldi + adIe + die)
(6.1)
=
Suppose that Ie
Ie,ba.ekground + Ie,eontrol where Ie,eontrol is the control signal which changes during a sniff. Olfactory adaptation is achieved by having an
Ie,eontrol = Iga.neel which cancels the effect of Iodor on Xo - cancelling. This
keeps the mitral cells baseline output Gz(Xo) and gain G~(Xo) low, and thus
makes the oscillation output impossible as if no odor exists. We can then expect
that reversing the sign of Iga.neel will cause the bulb to have an enhanced, instead
of reduced (adapted), response to Iodor - anti-cancelling, and achieve the olfactory enhancement. We can derive further phenomena such as recognizing an odor
component in an odor mixture, cross-adaptation and cross-enhancement (Li and
Hopfield 1988b). Computer simulations confirmed the expected results.
7. DISCUSSION
Our model of the olfactory bulb is a simplification of the known anatomy
and physiology. The net of the mitral and granule cells simulates a group of coupled non-linear oscillators which are the sources of the rhythmic activities in the
bulb. The coupling makes the oscillation coherent across the bulb surface for each
sniff. The model suggests, in agreement with Freeman and coworkers, that stability
change bifurcation is used for the bulbar oscillator system to decide primitively on
the relevance of the receptor input information. Different non-damping oscillation
modes emerged are used to distinguish the different odor input information which
is the driving source for the bifurcations, and are approximately thought of as the
(unitary) decision states of the system for the odor information. With the extra
information represented in the oscillation phases of the cells, the bulb emphasizes
the differences between different input patterns (section 4). Both the analysis and
simulation show that the bulb is selectively sensitive to different receptor input patterns. This selectivity as well as the motivation level of the animal could also be
Modeling the Olfactory Bulb-Coupled Nonlinear Oscillators
modulated from higher centers. This model also successfully applies to bulbar ability to use input from higher centers to suppress or enhance sensitivity to particular
target or to mask odors.
This model does not exclude the possibility that the information be coded in
the non-oscillatory slow wave Xo which is also determined by the odor input. The
chief behaviors do not depend on the number of cells in the model. The model can
be generalized to olfactory cortex, hippocampus and neocortex etc. where there are
more varieties of synaptic organizations.
Acknowledgements
This research was supported by ONR contract NOO014-87-K-0377. We would also
like to acknowledge discussions with J.A. Bower.
References
Baird B. Nonlinear dynamics of pattern formation and pattern recognition in rabbit
olfactory bulb. Physica 22D, 150-175 (1986)
Freeman W.J. Mass action in the nervous system. New York: Academic Press 1975
Freeman W.J. Spatial properties of an EEG event in the olfactory bulb and cortex.
Electroencephalogr. Clin. Neurophysiol. 44, 586-605 (1978)
Freeman W.J. Nonlinear Gain mediating cortical stimulus-response relations. BioI.
Cybernetics 33, 237-247 (1979a)
Freeman W.J. Nonlinear dynamics of paleocortex manifested in the olfactory EEG.
BioI. Cybernetics 35, 21-37 (1979b)
Freeman W.J. EEG analysis gives model of neuronal template-matching mechanism
for sensory search with olfactory bulb. BioI. Cybernetics 35, 221-234 (1979c)
Freeman W.J. Use of spatial deconvolution to compensate for distortion of EEG by
volume conduction. IEEE Trans. Biomed. Engineering 21,421-429 (1980)
Freeman W.J., Schneider W.S. Changes in spatial patterns of rabbit olfactory EEG
with conditioning to odors. Psychophysiology 19, 44-56 (1982)
Freeman W.J., Skarda C.A. Spatial EEG patterns, non-linear dynamics and perception: the Neo-Sherringtonian view. Brain Res. Rev. 10, 147-175 (1985)
Getchell T.V., Shepherd G.M. Responses of olfactory receptor cells to step pulses
of odour at different concentrations in the salamender. J. Physiol. 282, 521-540
(1978)
Lancet D. Vertebrate olfactory reception Ann. Rev. Neurosci. 9, 329-355 (1986)
Li Z., Hopfield J.J. Modeling the olfactory bulb. Submitted to Biological Cybernetics (1988a)
Li Z., Hopfield J.J. A model of olfactory adaptation and enhancement in the olfactory bulb. In preparation. (1988b)
Scott J. W. The olfactory bulb and central pathways. Experientia 42,223-232 (1986)
Shepherd G.M. The synaptic organization of the brain. New York: Oxford University Press 1979
Shepherd G.M. Private communications. (1988)
Skarda C.A., Freeman W.J. How brains make chaos in order to make sense of the
world. Behavioral and Brain Sciences 10, 161-195 (1987)
409
| 138 |@word private:1 version:1 eliminating:1 hippocampus:2 nd:2 ayy:3 simulation:9 pulse:3 wog:3 mammal:1 carry:1 initial:1 odour:1 must:1 physiol:1 j1:1 wx:2 discrimination:3 nervous:1 indicative:1 xk:1 ith:2 short:1 provides:1 location:1 simpler:1 mathematical:2 burst:1 pathway:2 behavioral:1 olfactory:35 mask:1 expected:1 behavior:1 examine:1 growing:3 ry:2 brain:6 paleocortex:1 freeman:27 little:1 increasing:1 becomes:3 vertebrate:1 classifies:1 mass:2 interpreted:1 astronomy:1 xd:1 oscillates:1 bulbar:13 classifier:1 exactly:1 control:2 originates:1 unit:1 before:1 engineering:2 local:3 consequence:1 receptor:11 ak:3 oxford:1 firing:1 fluctuation:1 yd:1 approximately:1 reception:1 studied:1 suggests:1 co:1 fastest:1 range:2 locked:3 directed:1 responsible:1 bell:1 thought:4 physiology:1 matching:1 induce:3 suggest:1 influence:1 impossible:1 center:8 starting:1 rabbit:3 simplicity:2 insight:1 dominate:1 population:1 stability:1 variation:1 controlling:1 play:1 suppose:2 enhanced:1 target:1 origin:1 agreement:1 element:1 approximated:1 recognition:1 located:1 observed:4 role:1 negligibly:1 capture:1 cycle:5 connected:2 contamination:1 dynamic:4 raise:1 depend:1 division:3 neurophysiol:1 aii:1 hopfield:16 iga:2 represented:1 surrounding:1 zo:3 describe:1 formation:1 choosing:1 whose:2 emerged:1 distortion:1 otherwise:1 interconnection:1 ability:1 skarda:8 transform:1 itself:1 final:1 eigenvalue:1 net:1 adaptation:3 cancelling:2 achieve:2 az:3 amplifies:1 enhancement:3 produce:1 ring:2 lamina:1 depending:1 coupling:4 derive:1 measured:2 ldi:1 implies:1 indicate:1 differ:1 anatomy:2 biological:1 physica:1 experientia:1 around:2 equilibrium:3 neel:2 driving:1 major:1 early:1 superposition:1 sensitive:1 vice:1 successfully:1 electroencephalogr:1 always:1 rather:1 yo:33 mainly:1 baseline:3 sense:1 stopping:1 pasadena:1 relation:1 biomed:1 animal:2 raised:1 spatial:4 bifurcation:2 field:2 having:1 biology:1 zhaoping:1 lit:1 inhalation:2 eontrol:3 cancel:1 others:2 stimulus:1 replaced:1 phase:7 geometry:1 organization:3 possibility:2 mixture:1 damped:7 ambient:1 necessary:1 damping:1 iv:1 orbit:3 circle:1 re:5 modeling:6 deviation:2 uniform:1 recognizing:1 conduction:1 aw:1 peak:1 sensitivity:5 ie:15 contract:1 physic:1 receiving:1 enhance:4 ym:2 quickly:1 central:6 satisfied:2 postulate:1 slowly:1 dr:1 external:1 lii:2 return:1 li:15 potential:1 sinusoidal:2 exclude:1 chemistry:1 gy:11 includes:1 baird:2 depends:2 view:1 closed:1 reached:1 wave:9 parallel:1 capability:1 il:1 characteristic:1 yield:1 weak:2 emphasizes:1 confirmed:1 zx:3 cybernetics:4 ah:2 submitted:1 oscillatory:5 strongest:1 influenced:1 synaptic:6 frequency:9 obvious:1 toe:1 di:5 gain:3 emerges:1 amplitude:8 higher:7 dt:1 noo014:1 psychophysiology:1 response:4 synapse:1 done:1 just:1 d:2 hand:1 nonlinear:8 mode:15 aj:1 usa:1 effect:4 y2:2 chemical:1 symmetric:1 laboratory:1 sin:1 during:8 die:1 generalized:2 ay:1 motion:1 dissipation:1 neo:1 chaos:1 sigmoid:1 behaves:1 stimulation:1 quarter:1 conditioning:1 exponentially:1 volume:1 nonsymmetric:1 discussed:1 versa:1 mathematics:1 ride:2 stable:3 cortex:7 surface:2 longer:1 etc:2 dominant:1 showed:1 driven:1 selectivity:1 certain:1 manifested:1 onr:1 yi:1 schneider:6 coworkers:1 signal:2 ii:3 academic:1 cross:2 compensate:1 coded:1 neuro:1 represent:1 achieved:1 cell:40 background:3 grow:1 source:2 appropriately:1 extra:1 shepherd:9 hz:3 simulates:1 unitary:1 ideal:1 spiral:2 variety:1 shift:2 linkage:1 effort:1 wo:1 york:2 oscillate:2 cause:1 action:1 detailed:1 eigenvectors:1 neocortex:2 locally:1 reduced:1 inhibitory:6 sign:1 arising:1 yy:2 group:4 changing:1 letter:2 respond:1 decide:1 oscillation:29 decision:4 simplification:1 distinguish:1 mitral:18 activity:9 strength:2 occur:1 adapted:1 x2:2 hy:1 simulate:1 across:5 terminates:2 describes:1 postsynaptic:1 rev:2 xo:33 equation:12 mechanism:1 know:1 appropriate:3 odor:33 ho:1 slower:1 rz:2 original:1 include:1 cf:2 clin:1 granule:20 noticed:1 added:1 coherently:1 occurs:1 concentration:1 sniff:10 diagonal:1 responds:1 unclear:1 exhibit:1 gradient:1 kth:3 simulated:1 presynaptic:1 extent:1 unstable:1 toward:1 index:1 mediating:1 hog:2 rise:2 aijxj:1 suppress:5 ba:1 upper:1 neuron:1 acknowledge:1 anti:1 situation:1 communication:1 interacting:1 arbitrary:1 pair:3 specified:1 connection:4 california:1 coherent:2 narrow:1 trans:1 lion:1 pattern:13 scott:2 wy:1 perception:1 breathing:1 regime:1 shifting:1 event:1 representing:2 loom:2 technology:1 gz:11 coupled:8 dxo:2 acknowledgement:1 relative:1 expect:1 bulb:45 sufficient:1 principle:1 lancet:1 excitatory:6 course:1 supported:1 last:1 jth:1 understand:1 institute:1 fall:2 template:1 emerge:1 rhythmic:2 curve:1 calculated:1 xn:2 cortical:2 world:1 sensory:2 made:1 simplified:1 keep:1 assumed:2 xi:4 continuous:1 physiologically:3 search:1 chief:1 ca:1 eeg:10 interact:1 adi:2 complex:2 neurosci:1 motivation:3 noise:3 whole:1 respiratory:2 fig:5 neuronal:1 slow:3 aid:1 structurally:1 position:1 adiabatic:1 msec:1 xl:1 candidate:1 bower:1 hw:3 jt:1 deconvolution:1 intrinsic:2 essential:1 exists:1 effectively:1 linearization:1 locality:1 electrophysiology:1 likely:1 scalar:2 bo:1 applies:1 determines:1 bioi:3 marked:1 ann:1 oscillator:21 absence:2 change:6 included:1 typical:1 determined:3 reversing:1 wt:1 indicating:1 selectively:2 internal:3 modulated:4 relevance:1 preparation:1 phenomenon:1 ex:1 |
417 | 1,380 | Estimating Dependency Structure as a Hidden
Variable
Marina Meill and Michael I. Jordan
{mmp, jordan}@ai.mit.edu
Center for Biological & Computational Learning
Massachusetts Institute of Technology
45 Carleton St. E25-201
Cambridge, MA 02142
Abstract
This paper introduces a probability model, the mixture of trees that can
account for sparse, dynamically changing dependence relationships. We
present a family of efficient algorithms that use EM and the Minimum
Spanning Tree algorithm to find the ML and MAP mixture of trees for a
variety of priors, including the Dirichlet and the MDL priors.
1 INTRODUCTION
A fundamental feature of a good model is the ability to uncover and exploit independencies
in the data it is presented with. For many commonly used models, such as neural nets and
belief networks, the dependency structure encoded in the model is fixed, in the sense that it
is not allowed to vary depending on actual values of the variables or with the current case.
However, dependency structures that are conditional on values of variables abound in the
world around us. Consider for example bitmaps of handwritten digits. They obviously
contain many dependencies between pixels; however, the pattern of these dependencies
will vary across digits. Imagine a medical database recording the body weight and other
data for each patient. The body weight could be a function of age and height for a healthy
person, but it would depend on other conditions if the patient suffered from a disease or
was an athlete.
Models that are able to represent data conditioned dependencies are decision trees and
mixture models, including the soft counterpart of the decision tree, the mixture of experts.
Decision trees however can only represent certain patterns of dependecy, and in particular
are designed to represent a set of conditional probability tables and not a joint probability
distribution. Mixtures are more flexible and the rest of this paper will be focusing on one
special case called the mixtures of trees.
We will consider domains where the observed variables are related by pairwise dependencies
only and these dependencies are sparse enough to contain no cycles. Therefore they can
Estimating Dependency Structure as a Hidden Variable
585
be represented graphically as a tree. The structure of the dependencies may vary from
one instance to the next. We index the set of possible dependecy structures by a discrete
structure variable z (that can be observed or hidden) thereby obtaining a mixture.
In the framework of graphical probability models, tree distributions enjoy many properties
that make them attractive as modelling tools: they have a flexible topology, are intuitively
appealing, sampling and computing likelihoods are linear time, simple efficient algorithms
for marginalizing and conditioning (O(1V1 2 ) or less) exist. Fitting the best tree to a given
distribution can be done exactly and efficiently (Chow and Liu, 1968). Trees can capture
simple pairwise interactions between variables but they can prove insufficient for more
complex distributions. Mixtures of trees enjoy most of the computational advantages of
trees and, in addition, they are universal approximators over the space of all distributions.
Therefore, they are fit for domains where the dependency patterns become tree like when a
possibly hidden variable is instantiated.
Mixture models have been extensively used in the statistics and neural network literature.
Of relevance to the present work are the mixtures of Gaussians, whose distribution space, in
the case of continuous variables overlaps with the space of mixtures of trees. Work on fitting
a tree to a distribution in a Maximum-Likelihood (ML) framework has been pioneered by
(Chow and Liu, 1968) and was extended to poly trees by (Pearl, 1988) and to mixtures
of trees with observed structure variable by (Geiger, 1992; Friedman and Goldszmidt,
1996). Mixtures of factorial distributions were studied by (Kontkanen et al., 1996) whereas
(Thiesson et aI., 1997) discusses mixtures of general belief nets. Multinets (Geiger, 1996)
which are essentially mixtures of Bayes nets include mixtures of trees as a special case.
It is however worth studying mixtures of trees separately for their special computational
advantages.
This work presents efficient algorithms for learning mixture of trees models with unknown
or hidden structure variable. The following section introduces the model; section 3 develops
the basic algorithm for its estimation from data in the ML framework. Section 4 discusses
the introduction of priors over mixtures of trees models and presents several realistic
factorized priors for which the MAP estimate can be computed by a modified versions of
the basic algorithm. The properties of the model are verified by simulation in section 5 and
section 6 concludes the paper.
2 THE MIXTURE OF TREES MODEL
In this section we will introduce the mixture of trees model and the notation that will be
used throughout the paper. Let V denote the set of variables of interest. According to the
graphical model paradigm, each variable is viewed as a vertex of a graph. Let Tv denote
the number of values of variable v E V, XV a particular value of V, XA an assignment to the
variables in the subset A of V. To simplify notation Xv will be denoted by x.
We use trees as graphical representations for families of probability distributions over V
that satisfy a common set of independence relationships encoded in the tree topology. In
this representation, an edge of the tree shows a direct dependence, or, more precisely, the
absence of an edge between two variables signifies that they are independent, conditioned
on all the other variables in V. We shall call a graph that has no cycles a tree I and shall
denote by E the set of its (undirected) edges. A probability distribution T that is conformal
with the tree (V, E) is a distribution that can be factorized as:
T (X) =
IT(u,v)EE Tuv (xu, xv)
IT vEV T,v (x v )degv-l
(1)
Here deg v denotes the degree of v, e.g. the number of edges incident to node v E V. The
l In the graph theory literature, our definition corresponds to a forest. The connected components
of a forest are called trees.
M. MeillJ and M. I. Jordan
586
factors Tuv and Tv are the marginal distributions under T:
Tuv(xu,xv) =
2: T(xu , xv,XV-{u ,v}),
2: T(xv,xv-{ v}) '
Tv(xv) =
(2)
Xv-tv}
XV-{u.v}
The distribution itself will be called a tree when no confusion is possible. Note that a tree
distribution has for each edge (u, v) E E a factor depending on xu, Xv onlyl If the tree is
connected, e.g. it spans all the nodes in V , it is often called a spanning tree.
An equivalent representation for T in terms of conditional probabilities is
T(x) =
II Tvlpa(v)(xvlxpa(v?)
(3)
vEV
The form (3) can be obtained from (1) by choosing an arbitrary root in each connected
component and recursively substituting Tv
t';V) by Tvlpa(v) starting from the root. pa(v)
represents the parent of v in the thus directed tree or the empty set if v is the root of
a connected component. The directed tree representation has the advantage of having
independent parameters. The total number of free parameters in either representation is
E(u,v)EET rurv - EVEv(degv - l)rv .
Now we define a mixture of trees to be a distribution of the form
m
Q(X)
= 2: AkTk(x);
Ak 2: 0, k
= 1, . .. , m;
(4)
k=1
From the graphical models perspecti ve, a mixture of trees can be viewed as a containing an
unobserved choice variable z, taking value k E {I, ... m} with probability Ak. Conditioned
on the value of z the distribution of the visible variables X is a tree. The m trees may have
different structures and different parameters. Note that because of the structure variable,
a mixture of trees is not properly a belief network, but most of the results here owe to the
belief network perspective.
3 THE BASIC ALGORITHM: ML FITIING OF MIXTURES OF
TREES
This section will show how a mixture of trees can be fit to an observed dataset in the Maximum Likelihood paradigm via the EM algorithm (Dempster et al., 1977). The observations
are denoted by {xl , x 2 , ... , xN}; the corresponding values of the structure variable are
{ zi,i=I, ... N}.
Following a usual EM procedure for mixtures, the Expectation (E) step consists in estimating
the posterior probability of each tree to generate datapoint xi
Pr[zi = klxl ,.. .,N, model] = 'Yk(i) =
AkTk(x:).
Lkl AklTk (x')
(5)
Then the expected complete log-likelihood to be maximized by the M step of the algorithm
is
m
E[Ic Ixl ,...N , model]
N
L rk[log Ak + L
k=1
pk(x i ) 10gTk(xi )]
(6)
i=1
N
rk =
2: 'Yk(X
i ),
(7)
i=1
The maximizing values for the parameters A are Akew = rk/ N. To obtain the new
distributions Tk, we have to maximize for each k the expression that is the negative of the
587
Estimating Dependency Structure as a Hidden Variable
Figure 1: The Basic Algorithm: ML Fitting of a Mixture of Trees
Input:Dataset {xl, ... xN}
Initial model m, Tk, ,\k, k = I, . .. m
Procedure MST( weights) that fits a maximum weight spanning tree over V
Iterate until convergence
Estep:
compute'Y~, pk(X') fork = I, . .. m, i= 1, . . . Nby(5),(7)
Mstep:
Ml. '\k +- rk/N, k = I, ... m
M2. compute marginals P:, p!v, U, v E V, k = I , ... m
MJ. compute mutual information I!v u, v E V , k
I , ... m
M4. call MST( { I!v }) to generate ETk for k = I, ... m
M5. T!v +- p!v, ; T: +- P: for (u, v) E ETk, k = I, ... m
=
crossentropy between pk and Tk.
N
L
pk(x i ) 10gTk(xi)
(8)
i=l
This problem can be solved exactly as shown in (Chow and Liu, 1968). Here we will
give a brief description of the procedure. First, one has to compute the mutual information
between each pair of variables in V under the target distribution P
JUti
'""
(
)
PUtl(x u , Xtl)
= Jvu = L.J P uv Xu, Xv log Pu(xu)PtI(x v )'
u, v E V, u
f=v .
(9)
X",X v
Second, the optimal tree structure is found by a Maximum Spanning Tree (MST) algorithm
using JUti as the weight for edge (u, v), \lu, v E V.Once the tree is found, its marginals Tutl
(or Tul v ), (u, v) E ET are exactly equal to the corresponding marginals P Uti of the target
distribution P. They are already computed as an intermediate step in the computation of
the mutual informations JUti (9).
In our case, the target distribution for Tk is represented by the posterior sample distribution
pk. Note that although each tree fit to pk is optimal, for the encompassing problem of
fitting a mixture of trees to a sample distribution only a local optimum is guaranteed to be
reached. The algorithm is summarized in figure 1.
This procedure is based on one important assumption that should be made explicit now. It
is the Parameter independence assumption: The distribution T:1pa( tI) for any k, v and
value of pa( v) is a multinomial with rv - 1 free parameters that are independent of any
other parameters of the mixture.
It is possible to constrain the m trees to share the same structure, thus constructing a truly
Bayesian network. To achieve this, it is sufficient to replace the weights in step M4 by
Lk J~tI and run the MST algorithm only once to obtain the common structure ET. The
tree stuctures obtained by the basic algorithm are connected. The following section will
give reasons and ways to obtain disconnected tree structures.
4
MAP MIXTURES OF TREES
In this section we extend the basic algorithm to the problem of finding the Maximum a
Posteriori (MAP) probability mixture of trees for a given dataset. In other words, we will
consider a nonuniform prior P[mode/] and will be searching for the mixture of trees that
maximizes
log P[model\x1 ,.. .N] = 10gP[xl, ... N\model]
+ log P[model] + constant.
(10)
Factorized priors The present maximization problem differs from the ML problem solved
in the previous section only by the addition of the term log P[model]. We can as well
M. Meilii and M. l. Jordan
588
approach it from the EM point of view, by iteratively maximizing
E [logP[modelJxl ,...N,
ZI , ...NJ]
= E[lc{xl ,...N, zl ,... NJmodel)]
+ 10gP[model]
(11)
It is easy to see that the added term does not have any influence on the E step,which
will proceed exactly as before. However, in the M step, we must be able to successfully
maximize the r.h.s. of (11). Therefore, we look for priors of the form
m
P[model] = P[AI, .. .m]
II P[Tkl
(12)
k=1
This class of priors is in agreement with the parameter independence assumption and
includes the conjugate prior for the multinomial distribution which is the Dirichlet prior. A
Dirichlet prior over a tree can be represented as a table of fictitious marginal probabilities
P~~ for each pair u , v of variables plus an equivalent sample size Nt that gives the strength
of the prior (Heckerman et al., 1995). However, for Dirichlet priors, the maximization over
tree structures (corresponding to step M4) can only be performed iteratively (Meilli et al.,
1997).
MDL (Minimum Description Length) priors are less informative priors. They attempt
to balance the number of parameters that are estimated with the amount of data available,
usually by introducing a penalty on model complexity. For the experiments in section 5
we used edge pruning. More smoothing methods are presented in (Meilli et al., 1997). To
penalize the number of parameters in each component we introduce a prior that penalizes
each edge that is added to a tree, thus encouraging the algorithm to produce disconnected
trees. The edge pruning prior is P [T] <X exp [-(3 L:utJ EET L\utJ] . We choose a uniform
penalty L\utJ = 1. Another possible choice is L\utJ = (r u - 1)( rtJ - 1) which is the number
of parameters introduced by the presence of edge (u , v) w.r.t. a factorized distribution.
Using this prior is equivalent to maximizing the following expression in step M4 of the
Basic Algorithm (the index k being dropped for simplicity)
argmax
L:
ET
utJEET
max[O, r1utJ
-
(3~tJ] = argmax
ET
L:
WutJ
(13)
utJEET
5 EMPIRICAL RESULTS
We have tested our model and algorithms for their ability to retrieve the dependency
structure in the data, as classifiers and as density estimators.
For the first objective, we sampled 30,000 datapoints from a mixture of 5 trees over 30
variables with rtJ = 4 for all vertices. All the other parameters of the generating model
and the initial points for the algorithm were picked at random. The results on retrieving
the original trees were excellent: out of 10 trials, the algorithm failed to retrieve correctly
only 1 tree in 1 trial. This bad result can be accounted for by sampling noise. The tree that
wasn't recovered had a A of only 0.02. The difference between the log likelihood of the
samples of the generating tree and the approximating tree was 0.41 bits per example.
For classification, we investigated the performance of mixtures of trees on a the Australian
Credit dataset from the UCI repository2. The data set has 690 instances of 14-dimensional
attribute vectors. Nine attributes are discrete ( 2 - 14 values) and 5 are continuous. The
class variable has 6 values. The continuous variables were discretized in 3 - 5 uniform bins
each. We tested mixtures with different values for m and for the edge pruning parameter (3.
For comparison we tried also mixtures of factorial distributions of different sizes. One tenth
of the data, picked randomly at each trial, was used for testing and the rest for training. In
the training phase, we learned a MT model of the joint distribution of all the 15 variables.
2
http://www.ics . uci.edu/-mlearn/MLRepository.html
Estimating Dependency Structure as a Hidden Variable
589
Figure 2: Performance of different algorithms on the Australian Credit dataset. - is mixture
of trees with j3 = 10, - - is mixture of trees with beta = 11m, -.- is mixture of factorial
distributions.
88
87
86
C1)
584
u
81
,
,,
~'i
.-
~83
82
,
-,,
(385
I
~
'?
t7
..,
I
...
I
I'
I '
800
5
10
15
m
20
25
30
Table 1: a) Mixture of trees compression rates [log lte&t! N te &t1. b) Compression rates
(bits/digit) for the single digit (Digit) and double digit (Pairs) datasets. MST is mixtures of
trees, MF is a mixture of factorial distributions, BR is base rate model, H-WS is Helmholtz
Machine trained with the wake-sleep algorithm (Frey et aI., 1996), H-MF is Helmholtz
Machine trained with the Mean Field approximation, FV is a fully visible bayes net.
(*=best)
(a)
(b)
Algorithm Digits Pairs
gzip
44.3
89.2
m
Digits
Pairs
BR
59.2 118.4
16 *34.72
79.25
MF
37.5
92.7
32
34.48 *78.99
H-MF
80.7
39.5
64
79.70
34.84
H-WS
39.1
80.4
128
34.88
81 .26
FV
35.9 *72.9
*34.7
79.0
MT
In the testing phase, the output of our classifier was chosen to be the class value with the
largest posterior probability given the inputs. Figure 2 shows that the results obtained for
mixtures of trees are superior to those obtained for mixtures of factorial distributions.For
comparison, correct classification rates obtained and cited in (Kontkanen et aI., 1996) on
training/test sets of the same size are: 87.2next best model (a decision tree called CaI50).
We also tested the basic algorithm as a density estimator by running it on a subset of
binary vector representations of handwritten digits and measuring the compression rate.
One dataset contained images of single digits in 64 dimensions, the second contained 128
dimensional vectors representing randomly paired digit images. The training, validation and
test set contained 6000, 2000, and 5000 exemplars respectively. The data sets, the training
conditions and the algorithms we compared with are described in (Frey et aI., 1996). We
tried mixtures of 16, 32, 64 and 128 trees, fitted by the basic algorithm. The results (shown
in table1 averaged over 3 runs) are very encouraging: the mixture of trees is the absolute
winner for compressing the simple digits and comes in second as a model for pairs of digits.
This suggests that our model (just like the mixture of factorized distributions) is able to
perform good compression of the digit data but is unable to discover the independency in
the double digit set.
590
M. Meilli and M. I. Jordan
6 CONCLUSIONS
This paper has shown a method of modeling and exploiting sparse dependency structure
that is conditioned on values of the data. By using trees, our method avoids the exponential
computation demands that plague both inference and structure finding in wider classes of
belief nets. The algorithms presented here are linear in m and N and quadratic in IV I.
Each M step is performing exact maximization over the space of all the tree structures and
parameters. The possibility of pruning the edges of the components of a mixture of trees
can playa role in classification, as a means of automatically selecting the variables that are
relevant for the task.
The importance of using the right priors in constructing models for real-world problems
can hardly be understated. In this context, the present paper has presented a broad class of
priors that are efficiently handled in the framework of our algorithm and it has shown that
this class includes important priors like the MDL prior and the Dirichlet prior.
Acknowledgements
Thanks to Quaid Morris for running the digits and structure finding experiments and to
Brendan Frey for providing the digits datasets.
References
Chow, C. K. and Liu, C. N. (1968). Approximating discrete probability distributions with dependence
trees. "IEEE Transactions on Information Theory ", IT-14(3 ):462-467.
Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977). Maximum likelihood from incomplete data
via the EM algorithm. Journal of the Royal Statistical Society, B, 39: 1-38.
Frey, B. J., Hinton, G. E., and Dayan, P. (1996). Does the wake-sleep algorithm produce good density
estimators? In Touretsky, D., Mozer, M., and Hasselmo, M., editors, Neural Information
Processing Systems, number 8, pages 661-667. MIT Press.
Friedman, N. and Goldszmidt, M. (1996). Building classifiers using Bayesian networks. In Proceedings of the National Conference on Artificial Intelligence (AAAI 96), pages 1277-1284, Menlo
Park, CA. AAAI Press.
Geiger, D. (1992). An entropy-based learning algorithm of bayesian conditional trees. In Proceedings
of the 8th Conferenceon Uncertainty in AI, pages 92-97. Morgan Kaufmann Publishers.
Geiger, D. (1996). Knowledge representation and inference in similarity networks and bayesian
multinets. "Artificial Intelligence", 82:45-74.
Heckerman, D., Geiger, D., and Chickering, D. M. (1995). Learning Bayesian networks: the
combination of knowledge and statistical data. Machine Learining, 20(3): 197-243.
Kontkanen, P., Myllymaki, P., and Tirri, H. (1996). Constructing bayesian finite mixture models by the
EM algorithm. Technical Report C-1996-9, Univeristy of Helsinky. Department of Computer
Science.
Meilli, M., Jordan, M. I., and Morris, Q. D. (1997). Estimating dependency structure as a hidden
variable. Technical Report AIM-1611 ,CBCL-151, Massachusetts Institute of Technology,
Artificial Intelligence Laboratory.
Pearl, 1. (1988). Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference.
Morgan Kaufman Publishers, San Mateo, CA.
Thiesson, B., Meek, C., Chickering, D. M., and Heckerman, D. (1997). Learning mixtures of Bayes
networks . Technical Report MSR-POR-97-30, Microsoft Research.
| 1380 |@word trial:3 msr:1 version:1 compression:4 simulation:1 tried:2 thereby:1 recursively:1 initial:2 liu:4 selecting:1 t7:1 bitmap:1 current:1 recovered:1 nt:1 must:1 mst:5 visible:2 realistic:1 informative:1 mstep:1 meilii:1 designed:1 intelligence:3 node:2 height:1 direct:1 become:1 beta:1 tirri:1 retrieving:1 prove:1 consists:1 xtl:1 fitting:4 introduce:2 pairwise:2 expected:1 discretized:1 automatically:1 actual:1 encouraging:2 abound:1 estimating:6 notation:2 discover:1 maximizes:1 factorized:5 kaufman:1 unobserved:1 finding:3 nj:1 ti:2 exactly:4 classifier:3 zl:1 medical:1 enjoy:2 before:1 t1:1 dropped:1 local:1 frey:4 xv:13 ak:3 plus:1 studied:1 mateo:1 dynamically:1 suggests:1 averaged:1 lte:1 directed:2 testing:2 differs:1 digit:17 procedure:4 empirical:1 universal:1 word:1 context:1 influence:1 www:1 equivalent:3 map:4 center:1 maximizing:3 graphically:1 starting:1 simplicity:1 m2:1 estimator:3 datapoints:1 retrieve:2 searching:1 imagine:1 target:3 pioneered:1 exact:1 agreement:1 pa:3 helmholtz:2 database:1 observed:4 fork:1 role:1 solved:2 capture:1 compressing:1 cycle:2 connected:5 yk:2 disease:1 mozer:1 dempster:2 complexity:1 trained:2 depend:1 joint:2 represented:3 instantiated:1 artificial:3 choosing:1 whose:1 encoded:2 plausible:1 ability:2 statistic:1 gp:2 itself:1 laird:1 obviously:1 advantage:3 net:5 interaction:1 uci:2 relevant:1 achieve:1 description:2 exploiting:1 parent:1 empty:1 convergence:1 optimum:1 double:2 produce:2 generating:2 table1:1 tk:4 wider:1 depending:2 exemplar:1 come:1 australian:2 correct:1 attribute:2 bin:1 tul:1 biological:1 around:1 lkl:1 ic:2 credit:2 exp:1 cbcl:1 substituting:1 vary:3 understated:1 estimation:1 healthy:1 largest:1 hasselmo:1 myllymaki:1 successfully:1 tool:1 mit:2 aim:1 modified:1 crossentropy:1 properly:1 modelling:1 likelihood:6 brendan:1 sense:1 posteriori:1 inference:3 dayan:1 chow:4 hidden:8 w:2 pixel:1 classification:3 flexible:2 html:1 denoted:2 smoothing:1 special:3 univeristy:1 mutual:3 marginal:2 equal:1 once:2 tkl:1 having:1 field:1 sampling:2 represents:1 broad:1 look:1 park:1 report:3 develops:1 simplify:1 intelligent:1 randomly:2 ve:1 national:1 m4:4 quaid:1 argmax:2 phase:2 owe:1 microsoft:1 friedman:2 attempt:1 interest:1 possibility:1 mdl:3 introduces:2 mixture:51 truly:1 tj:1 edge:12 tree:81 iv:1 incomplete:1 penalizes:1 fitted:1 instance:2 soft:1 modeling:1 measuring:1 logp:1 assignment:1 maximization:3 signifies:1 introducing:1 vertex:2 subset:2 uniform:2 dependency:16 st:1 person:1 fundamental:1 density:3 cited:1 thanks:1 probabilistic:1 michael:1 e25:1 aaai:2 containing:1 choose:1 possibly:1 por:1 expert:1 account:1 summarized:1 includes:2 satisfy:1 tuv:3 performed:1 root:3 view:1 picked:2 reached:1 bayes:3 kaufmann:1 efficiently:2 maximized:1 handwritten:2 bayesian:6 lu:1 worth:1 mlearn:1 datapoint:1 definition:1 sampled:1 dataset:6 massachusetts:2 knowledge:2 uncover:1 focusing:1 aktk:2 done:1 xa:1 just:1 until:1 mode:1 building:1 contain:2 counterpart:1 iteratively:2 laboratory:1 attractive:1 mlrepository:1 m5:1 complete:1 gzip:1 confusion:1 reasoning:1 image:2 common:2 superior:1 thiesson:2 multinomial:2 mt:2 conditioning:1 winner:1 extend:1 marginals:3 cambridge:1 ai:7 uv:1 had:1 similarity:1 base:1 pu:1 playa:1 posterior:3 perspective:1 certain:1 binary:1 approximators:1 morgan:2 minimum:2 paradigm:2 maximize:2 ii:2 rv:2 kontkanen:3 technical:3 marina:1 paired:1 j3:1 basic:9 patient:2 essentially:1 expectation:1 represent:3 penalize:1 c1:1 addition:2 whereas:1 separately:1 wake:2 suffered:1 carleton:1 publisher:2 rest:2 recording:1 undirected:1 jordan:6 call:2 ee:1 presence:1 intermediate:1 enough:1 easy:1 variety:1 independence:3 fit:4 zi:3 iterate:1 topology:2 wasn:1 br:2 expression:2 handled:1 penalty:2 proceed:1 nine:1 hardly:1 etk:2 factorial:5 amount:1 extensively:1 morris:2 generate:2 http:1 exist:1 estimated:1 correctly:1 per:1 discrete:3 shall:2 independency:2 changing:1 verified:1 tenth:1 v1:1 graph:3 run:2 uncertainty:1 family:2 throughout:1 uti:1 geiger:5 decision:4 bit:2 guaranteed:1 meek:1 sleep:2 quadratic:1 strength:1 precisely:1 constrain:1 multinets:2 athlete:1 span:1 performing:1 utj:4 estep:1 department:1 tv:5 according:1 combination:1 disconnected:2 conjugate:1 across:1 heckerman:3 em:6 pti:1 appealing:1 intuitively:1 pr:1 discus:2 conformal:1 studying:1 available:1 gaussians:1 original:1 denotes:1 dirichlet:5 include:1 running:2 graphical:4 exploit:1 approximating:2 society:1 objective:1 already:1 added:2 dependence:3 usual:1 unable:1 spanning:4 reason:1 length:1 touretsky:1 index:2 relationship:2 insufficient:1 providing:1 balance:1 negative:1 unknown:1 perform:1 observation:1 datasets:2 finite:1 extended:1 hinton:1 nonuniform:1 arbitrary:1 introduced:1 pair:6 plague:1 fv:2 learned:1 pearl:2 able:3 usually:1 pattern:3 including:2 rtj:2 max:1 belief:5 royal:1 overlap:1 representing:1 technology:2 brief:1 lk:1 concludes:1 prior:23 literature:2 acknowledgement:1 marginalizing:1 encompassing:1 fully:1 fictitious:1 ixl:1 age:1 validation:1 degree:1 incident:1 sufficient:1 rubin:1 editor:1 share:1 accounted:1 free:2 institute:2 taking:1 absolute:1 sparse:3 dimension:1 xn:2 world:2 avoids:1 commonly:1 made:1 san:1 transaction:1 pruning:4 eet:2 ml:7 deg:1 xi:3 continuous:3 table:3 mj:1 ca:2 menlo:1 obtaining:1 forest:2 excellent:1 complex:1 poly:1 constructing:3 domain:2 investigated:1 pk:6 noise:1 gtk:2 allowed:1 body:2 xu:6 x1:1 lc:1 explicit:1 mmp:1 exponential:1 xl:4 chickering:2 rk:4 bad:1 importance:1 te:1 conditioned:4 demand:1 mf:4 entropy:1 vev:2 failed:1 contained:3 corresponds:1 ma:1 conditional:4 viewed:2 replace:1 absence:1 called:5 total:1 repository2:1 goldszmidt:2 relevance:1 tested:3 |
418 | 1,381 | Hippocampal Model of Rat Spatial Abilities
Using Temporal Difference Learning
David J Foster*
Centre for Neuroscience
Edinburgh University
Richard GM Morris
Centre for Neuroscience
Edinburgh University
Peter Dayan
E25-210, MIT
Cambridge, MA 02139
Abstract
We provide a model of the standard watermaze task, and of a more
challenging task involving novel platform locations, in which rats
exhibit one-trial learning after a few days of training. The model
uses hippocampal place cells to support reinforcement learning,
and also, in an integrated manner, to build and use allocentric
coordinates.
1 INTRODUCTION
Whilst it has long been known both that the hippocampus of the rat is needed for
normal performance on spatial tasks l3 , 11 and that certain cells in the hippocampus
exhibit place-related firing,12 it has not been clear how place cells are actually used
for navigation. One of the principal conceptual problems has been understanding
how the hippocampus could specify or learn paths to goals when spatially tuned
cells in the hippocampus respond only on the basis of the rat's current location.
This work uses recent ideas from reinforcement learning to solve this problem in
the context of two rodent spatial learning results.
Reference memory in the watermaze l l (RMW) has been a key task demonstrating
the importance of the hippocampus for spatial learning. On each trial, the rat is
placed in a circular pool of cloudy water, the only escape from which is a platform
which is hidden (below the water surface) but which remains in a constant position.
A random choice of starting pOSition is used for each trial. Rats take asymptotically
short paths after approximately 10 trials (see figure 1 a). Delayed match-to-place
(DMP) learning is a refined version in which the platform'S location is changed on
each day. Figure 1b shows escape latencies for rats given four trials per day for nine
days, with the platform in a novel position on each day. On early days, acquisition
?Crichton Street, Edinburgh EH8 9LE, United Kingdom. Funded by Edin. Univ.
Holdsworth Scholarship, the McDonnell-Pew foundation and NSF grant IBN-9634339.
Email: djf@cfn.ed.ac.uk
D. J Foster; R. G. M. Mo"is and P. Dayan
146
100
100
a
90
90
b
80
_ 70
'"
i
oo
'" 50
-'
~40
~30
20
10
13
17
21
2S
Figure 1: a) Latencies for rats on the reference memory in the watermaze (RMW)
task (N=8). b) Latencies for rats on the Delayed Match-to-Place (DMP) task (N=62).
is gradual but on later days, rats show one-trial learning, that is, near asymptotic
performance on the second trial to a novel platform position.
The RMW task has been extensively modelled. 6,4,5,20 By contrast, the DMP task
is new and computationally more challenging. It is solved here by integrating a
standard actor-critic reinforcement learning system2 ,7 which guarantees that the
rat will be competent to perform well in arbitrary mazes, with a system that learns
spatial coordinates in the maze. Temporal difference learning1 7 (TO) is used for actor,
critic and coordinate learning. TO learning is attractive because of its generality for
arbitrary Markov decision problems and the fact that reward systems in vertebrates
appear to instantiate it. 14
2 THEMODEL
The model comprises two distinct networks (figure 2): the actor-critic network and
a coordinate learning network. The contribution of the hippocampus, for both
networks, is to provide a state-space representation in the form of place cell basis
functions. Note that only the activities of place cells are required, by contrast with
decoding schemes which require detailed information about each place cell. 4
COORDINATE SYSTEM
ACTOR-CRITIC
SYSTEM
r------------Remembered
1
Goal coordinates
1
1
1
VECTOR COMPUTA nONI
~
Coordinate
Representation
1
1______ -------1
Figure 2: Model diagram showing the interaction between actor-critic and coordinate system components.
Hippocampal Model of Rat Spatial Abilities Using TD Learning
147
2.1 Actor-Critic Learning
Place cells are modelled as being tuned to location. At position p, place cell
i has an output given by h(p) = exp{ -lip - sdI2/2(12}, where Si is the place
field centre, and (1 = 0.1 for all place fields. The critic learns a value function
V(p)
wih(p) which comes to represent the distance of p from the goal, using
= L:i
the TO rule 6.w~ ex: 8t h(pt), where
(1)
is the TD error, pt is position at time t, and the reward r(pt, pt+I) is 1 for any
move onto the platform, and 0 otherwise. In a slight alteration of the original rule,
the value V (p) is set to zero when p is at the goal, thus ensuring that the total
future rewards for moving onto the goal will be exactly 1. Such a modification
improves stability in the case of TD learning with overlapping basis functions.
The discount factor, I' was set to 0.99. Simultaneously the rat refines a policy,
which is represented by eight action cells. Each action cell (aj in figure 2) receives
a parameterised input at any position p: aj (p) = L:i qjdi (p). An action is chosen
stochastically with probabilities given by P(aj) = exp{2aj}/ L:k exp{2ak}. Action
weights are reinforced according to: 2
(2)
where 9j((Jt) is a gaussian function of the difference between the head direction
(Jt at time t and the preferred direction of the jth action cell. Figure 3 shows the
development of a policy over a few trials.
V(p)l
Triall
V(p)
1
0.5
0.5
0.
0.5
01
0.5
TrialS
Triall3
V(P)l
0.5
I
0
0.5
0:
0.5
0.5
-0.5 -------0.5
.---'--- 0
0.5
Figure 3: The RMW task: the value function gradually disseminates information
about reward proximity to all regions of the environment. Policies and paths are
also shown.
There is no analytical guarantee for the convergence of TD learning with policy
adaptation. However our simulations show that the algorithm always converges
for the RMW task. In a simulated arena of diameter 1m and with swimming speeds
of 20cm/s, the simulation matched the performance of the real rats very closely (see
figure S). This demonstrates that TD-based reinforcement learning is adequately
fast to account for the learning performance of real animals.
D. 1. Foster, R. G. M Morris and P. Dayan
148
2.2
Coordinate Learning
Although the learning of a value function and policy is appropriate for finding
a fixed platform, the actor-critic model does not allow the transfer of knowledge
from the task defined by one goal position to that defined by any other; thus it
could not generate the sort of one-trial learning that is shown by rats on the DMP
task (see figure 1b). This requires acquisition of some goal-independent know ledge
about s~ace. A natural mechanism for this is the path integration or self-motion
system. 0,10 However, path integration presents two problems. First, since the rat
is put into the maze in a different position for each trial, how can it learn consistent
coordinates across the whole maze? Second, how can a general, powerful, but slow,
behavioral learning mechanism such as TO be integrated with a specific, limited,
but fast learning mechanism involving spatial coordinates?
Since TO critic learning is based on enforcing consistency in estimates of future
reward, we can also use it to learn spatially consistent coordinates on the basis
of samples of self-motion. It is assumed that the rat has an allocentric frame of
reference. 1s The model learns parameterised estimates of the x and y coordinates
of all positions p: x(p) = Li w[ fi(P) and y(p) = Li wY h(p), Importantly, while
place cells were again critical in supporting spatial representation, they do not embody
a map of space. The coordinate functions, like the value function previously, have to
be learned.
As the simulated rat moves around, the coordinate weights {w[} are adjusted
according to:
t
Llwi ()( (Llxt
+ X(pt+l ) -
X(pt))
LA
t-
k h(pk)
(3)
k=1
where Llxt is the self-motion estimate in the x direction. A similar update is applied
to {wn. In this case, the full TO(A) algorithm was used (with A = 0.9); however
TD(O) could also have been used, taking slightly longer. Figure 4a shows the x and
y coordinates at early and late phases of learning. It is apparent that they rapidly
become quite accurate - this is an extremely easy task in an open field maze.
An important issue in the learning of coordinates is drift, since the coordinate
system receives no direct information about the location of the origin. It turns out
that the three controlling factors over the implicit origin are: the boundary of the
arena, the prior setting of the coordinate weights (in this case all were zero) and
the position and prior value of any absorbing area (in this case the platform). If the
coordinate system as a whole were to drift once coordinates have been established,
this would invalidate coordinates that have been remembered by the rat over long
periods. However, since the expected value of the prediction error at time steps
should be zero for any self-consistent coordinate mapping, such a mapping should
remain stable. This is demonstrated for a single run: figure 4b shows the mean
value of coordinates x evolving over trials, with little drift after the first few trials.
We modeled the coordinate system as influencing the choice of swimming direction
in the manner of an abstract action. I5 The (internally specified) coordinates of the
most recent goal position are stored in short term memory and used, along with the
current coordinates, to calculate a vector heading. This vector heading is thrown
into the stochastic competition with the other possible actions, governed by a
single weight which changes in a similar manner to the other action weights (as in
equation 2, see also fig 4d), depending on the TO error, and on the angular proximity
of the current head direction to the coordinate direction. Thus, whether the the
coordinate-based direction is likely to be used depends upon its past performance.
One simplification in the model is the treatment of extinction. In the DMP task,
Hippocampal Model ofRat Spatial Abilities Using 1D Learning
149
,.
"
TJUAL
d
III
.1
TRIAL
26
16
i:
~Ol
~o
!"
~o
Figure 4: The evolution of the coordinate system for a typical simulation run: a.)
coordinate outputs at early and late phases of learning, b.) the extent of drift in the
coordinates, as shown by the mean coordinate value for a single run, c.) a measure
A2
~ ~ {X (Pr.)-X -X(pr.)}2
?
o f coord mate error for the same run (7E =
r
r. (Np-l)N
' where k
r
r
r
indexes measurement points (max N p ) and r indexes runs (max N r ), Xr(Pk) is the
model estimate of X at position Pk, X(Pk) is the ideal estimate for a coordinate
system centred on zero, and Xr is the mean value over all the model coordinates,
d.) the increase during training of the probability of choosing the abstract action.
This demonstrates the integration of the coordinates into the control system.
real rats extinguish to a platform that has moved fairly quickly whereas the actorcritic model extinguishes far more slowly. To get around this, when a simulated
rat reaches a goal that has just been moved, the value and action weights are
reinitialised, but the coordinate weights wf and wf, and the weights for the abstract
action, are not.
3 RESULTS
The main results of this paper are the replication by simulation of rat performance
on the RMW and DMP tasks. Figures la and b show the course of learning for
the rats; figures Sa and b for the model. For the DMP task, one-shot acquisition is
apparent by the end of training.
4
DISCUSSION
We have built a model for one-trial spatial learning in the watermaze which uses
a single TD learning algorithm in two separate systems. One system is based on a
reinforcement learning that can solve general Markovian decision problems, and
the other is based on coordinate learning and is specialised for an open-field water
maze. Place cells in the hippocampus offer an excellent substrate for learning the
actor, the critic and the coordinates.
The model is explicit about the relationship between the general and specific learning systems, and the learning behavior shows that they integrate seamlessly. As
currently constituted, the coordinate system would fail if there were a barrier in
the maze. We plan to extend the model to allow the coordinate system to specify
abstract targets other than the most recent platform position - this could allow
it fast navigation around a larger class of environments. It is also important to
improve the model of learning 'set' behavior - the information about the nature of
D. 1. Foster; R. G. M. Mo"is and P. Dayan
150
14
a
b
12
10
.
.
?.
z
j:10\
~
~
~
12
~
S>
..
'"~ ............................................
0~D.~yl~~y~2~D~.y~3~D~.y~47~~yS~~~.~~~y~7~~~y~.7D.~y9~
Figure 5: a.) Performance of the actor-critic model on the RMW task, and b.)
performance of the full model on the DMP task. The data for comparison is shown
in figures la and b.
the DMP task that the rats acquire over the course of the first few days of training.
Interestingly, learning set is incomplete - on the first trial of each day, the rats
still aim for the platform position on the previous day, even though this is never
correct. 16 The significant differences in the path lengths on the first trial of each
day (evidence in figure Ib and figure 5b) come from the relative placements of the
platforms. However, the model did not use the same positions as the empirical
data, and, in any case, the model of exploration behavior is rather simplistic.
The model demonstrates that reinforcement learning methods are perfectly fast
enough to match empirical learning curves. This is fortunate, since, unlike most
models specifically designed for open-field navigation,6,4,5,2o RL methods can
provably cope with substantially more complicated tasks with arbitrary barriers,
etc, since they solve the temporal credit assignment problem in its full generality.
The model also addresses the problem that coordinates in different parts of the
same environment need to be mutually consistent, even if the animal only experiences some parts on separate trials. An important property of the model is that
there is no requirement for the animal to have any explicit knowledge of the relationship between different place cells or place field position, size or shape. Such a
requirement is imposed in various models. 9 ,4,6,2o
Experiments that are suggested by this model (as well as by certain others) concern the relationship between hippocampally dependent and independent spatial
learning. First, once the coordinate system has been acquired, we predict that
merely placing the rat at a new location would be enough to let it find the platform
in one shot, though it might be necessary to reinforce the placement e.g. by first
placing the rat in a bucket of cold water. Second, we know that the establishment
of place fields in an environment happens substantiallr faster than establishment
of one-shot or even ordinary learning to a platform. 2 We predict that blocking
plasticity in the hippocampus following the establishment of place cells (possibly
achieved without a platform) would not block learning of a platform. In fact, new
experiments show that after extensive pre-training, rats can perform one-trial learning in the same environment to new platform positions on the DMP task without
hippocampal synaptic plasticity. 16 This is in contrast to the effects of hippocampal
lesion, which completely disrupts performance. According to the model, coordinates will have been learned during pre-training. The full prediction remains
untested: that once place fields have been established, coordinates could be learned
in the absence of hippocampal synaptic plasticity. A third prediction follows from
evidence that rats with restricted hippocampal lesions can learn the fixed platform
Hippocampal Model of Rat Spatial Abilities Using TD Learning
151
task, but much more slowly, based on a gradual "shaping" procedure. 22 In our
model, they may also be able to learn coordinates. However, a lengthy training
procedure could be required, and testing might be complicated if expressing the
knowledge required the use of hippocampus dependent short-term memory for
the last platform location. I6
One way of expressing the contribution of the hippocampus in the model is to say
that its function is to provide a behavioural state space for the solution of complex
tasks. Hence the contribution of the hippocampus to navigation is to provide
place cells whose firing properties remain consistent in a given environment. It
follows that in different behavioural situations, hippocampal cells should provide
a representation based on something other than locations - and, indeed, there
is evidence for this. 8 With regard to the role of the hippocampus in spatial tasks,
the model demonstrates that the hippocampus may be fundamentally necessary
without embodying a map.
References
[1] Barto, AG & Sutton, RS (1981) BioI. Cyber., 43:1-8.
[2] Barto, AG, Sutton, RS & Anderson, CW (1983) IEEE Trans. on Systems, Man
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
and Cybernetics 13:834-846.
Barto, AG, Sutton, RS & Watkins, CJCH (1989) Tech Report 89-95, CAIS, Univ.
Mass., Amherst, MA.
Blum, KI & Abbott, LF (1996) Neural Computation, 8:85-93.
Brown, MA & Sharp, PE (1995) Hippocampus 5:171-188.
Burgess, N, Reece, M & O'Keefe, J (1994) Neural Networks, 7:1065-1081.
Dayan, P (1991) NIPS 3, RP Lippmann et aI, eds., 464-470.
Eichenbaum, HB (1996) CurroOpin. Neurobiol., 6:187-195.
Gerstner, W & Abbott, LF (1996) J. Computational Neurosci. 4:79-94.
McNaughton, BL et a1 (1996) J. Exp. BioI., 199:173-185.
Morris, RGM et al (1982) Nature, 297:681-683.
O'Keefe, J & Dostrovsky, J (1971) Brain Res., 34(171).
Olton, OS & Samuelson, RJ (1976) J. Exp. Psych: A.B.P., 2:97-116.
Rudy, JW & Sutherland, RW (1995) Hippocampus, 5:375-389.
SchUltz, W, Dayan, P & Montague, PR (1997) Science, 275, 1593-1599.
Singh, SP Reinforcement learning with a hierarchy of abstract models.
Steele, RJ & Morris, RGM in preparation.
Sutton, RS (1988) Machine Learning, 3:9-44.
Taube, JS (1995) J. Neurosci. 15(1):70-86.
Tsitsiklis, IN & Van Roy, B (1996) Tech Report LIDS-P-2322, M.LT.
Wan, HS, Touretzky, OS & Redish, AD (1993) Proc. 1993 Connectionist Models
Summer School, Lawrence Erlbaum, 11-19.
Watkins, CJCH (1989) PhD Thesis, Cambridge.
Whishaw, IQ & Jarrard, LF (1996) Hippocampus
Wilson, MA & McNaughton, BL (1993) Science 261:1055-1058.
| 1381 |@word h:1 trial:19 version:1 hippocampus:16 extinction:1 open:3 gradual:2 simulation:4 r:4 shot:3 united:1 tuned:2 interestingly:1 past:1 current:3 si:1 refines:1 plasticity:3 shape:1 designed:1 update:1 instantiate:1 short:3 rgm:2 location:8 along:1 direct:1 become:1 replication:1 behavioral:1 manner:3 acquired:1 indeed:1 expected:1 behavior:3 disrupts:1 embody:1 ol:1 brain:1 td:8 little:1 vertebrate:1 matched:1 mass:1 cm:1 neurobiol:1 substantially:1 psych:1 whilst:1 finding:1 ag:3 cjch:2 guarantee:2 temporal:3 computa:1 exactly:1 edin:1 demonstrates:4 uk:1 control:1 grant:1 internally:1 appear:1 sutherland:1 influencing:1 sutton:4 ak:1 firing:2 path:6 approximately:1 might:2 extinguishes:1 coord:1 challenging:2 limited:1 testing:1 block:1 lf:3 xr:2 cold:1 procedure:2 area:1 empirical:2 evolving:1 pre:2 integrating:1 get:1 onto:2 put:1 context:1 map:2 demonstrated:1 imposed:1 starting:1 rule:2 importantly:1 stability:1 coordinate:46 mcnaughton:2 rmw:7 pt:6 gm:1 controlling:1 target:1 hierarchy:1 substrate:1 us:3 origin:2 roy:1 blocking:1 role:1 solved:1 calculate:1 region:1 environment:6 reward:5 singh:1 upon:1 basis:4 completely:1 dmp:10 montague:1 represented:1 various:1 univ:2 distinct:1 fast:4 reece:1 choosing:1 refined:1 apparent:2 ace:1 quite:1 solve:3 larger:1 say:1 whose:1 otherwise:1 ability:4 analytical:1 wih:1 interaction:1 adaptation:1 rapidly:1 moved:2 competition:1 convergence:1 requirement:2 converges:1 oo:1 depending:1 ac:1 iq:1 school:1 sa:1 ibn:1 come:2 direction:7 untested:1 closely:1 correct:1 stochastic:1 exploration:1 require:1 learning1:1 adjusted:1 proximity:2 around:3 credit:1 normal:1 exp:5 lawrence:1 mapping:2 predict:2 mo:2 early:3 a2:1 proc:1 currently:1 mit:1 gaussian:1 always:1 aim:1 establishment:3 rather:1 barto:3 wilson:1 seamlessly:1 tech:2 contrast:3 wf:2 dayan:6 dependent:2 integrated:2 hidden:1 provably:1 issue:1 development:1 plan:1 spatial:13 platform:19 animal:3 integration:3 fairly:1 field:8 once:3 never:1 placing:2 hippocampally:1 future:2 report:2 cfn:1 np:1 others:1 richard:1 few:4 fundamentally:1 connectionist:1 escape:2 simultaneously:1 delayed:2 phase:2 thrown:1 circular:1 arena:2 navigation:4 accurate:1 necessary:2 experience:1 incomplete:1 re:1 dostrovsky:1 markovian:1 assignment:1 ordinary:1 erlbaum:1 stored:1 rudy:1 amherst:1 yl:1 decoding:1 pool:1 e25:1 quickly:1 again:1 thesis:1 wan:1 slowly:2 possibly:1 stochastically:1 li:2 account:1 centred:1 alteration:1 redish:1 depends:1 ad:1 later:1 sort:1 complicated:2 djf:1 actorcritic:1 contribution:3 reinforced:1 samuelson:1 modelled:2 disseminates:1 cybernetics:1 reach:1 touretzky:1 ed:2 lengthy:1 email:1 synaptic:2 acquisition:3 treatment:1 holdsworth:1 knowledge:3 improves:1 shaping:1 actually:1 day:11 specify:2 jw:1 though:2 generality:2 anderson:1 parameterised:2 implicit:1 angular:1 just:1 receives:2 y9:1 o:2 overlapping:1 llxt:2 aj:4 effect:1 steele:1 brown:1 adequately:1 evolution:1 hence:1 spatially:2 attractive:1 llwi:1 during:2 self:4 rat:30 hippocampal:10 motion:3 novel:3 fi:1 absorbing:1 rl:1 extend:1 slight:1 measurement:1 significant:1 expressing:2 cambridge:2 ai:1 pew:1 consistency:1 i6:1 centre:3 funded:1 l3:1 moving:1 stable:1 actor:9 longer:1 invalidate:1 surface:1 etc:1 something:1 j:1 recent:3 certain:2 remembered:2 period:1 taube:1 full:4 rj:2 match:3 faster:1 offer:1 long:2 y:1 a1:1 ensuring:1 prediction:3 involving:2 simplistic:1 represent:1 achieved:1 cell:18 whereas:1 diagram:1 unlike:1 cyber:1 near:1 ideal:1 crichton:1 easy:1 wn:1 iii:1 enough:2 hb:1 burgess:1 perfectly:1 idea:1 whether:1 peter:1 nine:1 action:11 latency:3 clear:1 detailed:1 discount:1 extensively:1 morris:4 embodying:1 diameter:1 rw:1 generate:1 nsf:1 neuroscience:2 per:1 key:1 four:1 demonstrating:1 blum:1 abbott:2 asymptotically:1 swimming:2 merely:1 run:5 powerful:1 respond:1 i5:1 place:20 decision:2 ki:1 summer:1 simplification:1 activity:1 placement:2 speed:1 cloudy:1 extremely:1 eichenbaum:1 according:3 mcdonnell:1 across:1 slightly:1 remain:2 lid:1 modification:1 happens:1 gradually:1 pr:3 restricted:1 bucket:1 computationally:1 equation:1 mutually:1 remains:2 previously:1 turn:1 behavioural:2 mechanism:3 fail:1 needed:1 know:2 end:1 eight:1 appropriate:1 allocentric:2 rp:1 ledge:1 original:1 scholarship:1 build:1 bl:2 move:2 exhibit:2 distance:1 separate:2 reinforce:1 simulated:3 cw:1 street:1 extent:1 water:4 enforcing:1 length:1 modeled:1 index:2 relationship:3 acquire:1 kingdom:1 noni:1 policy:5 perform:2 markov:1 mate:1 supporting:1 situation:1 head:2 frame:1 arbitrary:3 sharp:1 watermaze:4 drift:4 david:1 required:3 specified:1 extensive:1 whishaw:1 learned:3 eh8:1 established:2 nip:1 trans:1 address:1 able:1 suggested:1 below:1 wy:1 built:1 max:2 memory:4 critical:1 natural:1 scheme:1 improve:1 prior:2 understanding:1 asymptotic:1 relative:1 foundation:1 integrate:1 consistent:5 foster:4 critic:11 course:2 changed:1 placed:1 last:1 jth:1 heading:2 tsitsiklis:1 allow:3 taking:1 barrier:2 edinburgh:3 regard:1 boundary:1 curve:1 van:1 maze:7 reinforcement:7 schultz:1 far:1 cope:1 lippmann:1 preferred:1 conceptual:1 assumed:1 themodel:1 lip:1 learn:5 transfer:1 nature:2 excellent:1 complex:1 gerstner:1 did:1 pk:4 main:1 constituted:1 neurosci:2 sp:1 whole:2 lesion:2 competent:1 fig:1 slow:1 position:18 comprises:1 explicit:2 fortunate:1 governed:1 pe:1 ib:1 watkins:2 late:2 third:1 learns:3 specific:2 jt:2 showing:1 system2:1 evidence:3 concern:1 extinguish:1 importance:1 keefe:2 phd:1 rodent:1 specialised:1 lt:1 likely:1 ma:4 bioi:2 goal:9 absence:1 man:1 change:1 typical:1 specifically:1 principal:1 total:1 la:3 support:1 preparation:1 ex:1 |
419 | 1,382 | Hippocampal Model of Rat Spatial Abilities
Using Temporal Difference Learning
David J Foster*
Centre for Neuroscience
Edinburgh University
Richard GM Morris
Centre for Neuroscience
Edinburgh University
Peter Dayan
E25-210, MIT
Cambridge, MA 02139
Abstract
We provide a model of the standard watermaze task, and of a more
challenging task involving novel platform locations, in which rats
exhibit one-trial learning after a few days of training. The model
uses hippocampal place cells to support reinforcement learning,
and also, in an integrated manner, to build and use allocentric
coordinates.
1 INTRODUCTION
Whilst it has long been known both that the hippocampus of the rat is needed for
normal performance on spatial tasks l3 , 11 and that certain cells in the hippocampus
exhibit place-related firing,12 it has not been clear how place cells are actually used
for navigation. One of the principal conceptual problems has been understanding
how the hippocampus could specify or learn paths to goals when spatially tuned
cells in the hippocampus respond only on the basis of the rat's current location.
This work uses recent ideas from reinforcement learning to solve this problem in
the context of two rodent spatial learning results.
Reference memory in the watermaze l l (RMW) has been a key task demonstrating
the importance of the hippocampus for spatial learning. On each trial, the rat is
placed in a circular pool of cloudy water, the only escape from which is a platform
which is hidden (below the water surface) but which remains in a constant position.
A random choice of starting pOSition is used for each trial. Rats take asymptotically
short paths after approximately 10 trials (see figure 1 a). Delayed match-to-place
(DMP) learning is a refined version in which the platform'S location is changed on
each day. Figure 1b shows escape latencies for rats given four trials per day for nine
days, with the platform in a novel position on each day. On early days, acquisition
?Crichton Street, Edinburgh EH8 9LE, United Kingdom. Funded by Edin. Univ.
Holdsworth Scholarship, the McDonnell-Pew foundation and NSF grant IBN-9634339.
Email: djf@cfn.ed.ac.uk
D. J Foster; R. G. M. Mo"is and P. Dayan
146
100
100
a
90
90
b
80
_ 70
'"
i
oo
'" 50
-'
~40
~30
20
10
13
17
21
2S
Figure 1: a) Latencies for rats on the reference memory in the watermaze (RMW)
task (N=8). b) Latencies for rats on the Delayed Match-to-Place (DMP) task (N=62).
is gradual but on later days, rats show one-trial learning, that is, near asymptotic
performance on the second trial to a novel platform position.
The RMW task has been extensively modelled. 6,4,5,20 By contrast, the DMP task
is new and computationally more challenging. It is solved here by integrating a
standard actor-critic reinforcement learning system2 ,7 which guarantees that the
rat will be competent to perform well in arbitrary mazes, with a system that learns
spatial coordinates in the maze. Temporal difference learning1 7 (TO) is used for actor,
critic and coordinate learning. TO learning is attractive because of its generality for
arbitrary Markov decision problems and the fact that reward systems in vertebrates
appear to instantiate it. 14
2 THEMODEL
The model comprises two distinct networks (figure 2): the actor-critic network and
a coordinate learning network. The contribution of the hippocampus, for both
networks, is to provide a state-space representation in the form of place cell basis
functions. Note that only the activities of place cells are required, by contrast with
decoding schemes which require detailed information about each place cell. 4
COORDINATE SYSTEM
ACTOR-CRITIC
SYSTEM
r------------Remembered
1
Goal coordinates
1
1
1
VECTOR COMPUTA nONI
~
Coordinate
Representation
1
1______ -------1
Figure 2: Model diagram showing the interaction between actor-critic and coordinate system components.
Hippocampal Model of Rat Spatial Abilities Using TD Learning
147
2.1 Actor-Critic Learning
Place cells are modelled as being tuned to location. At position p, place cell
i has an output given by h(p) = exp{ -lip - sdI2/2(12}, where Si is the place
field centre, and (1 = 0.1 for all place fields. The critic learns a value function
V(p)
wih(p) which comes to represent the distance of p from the goal, using
= L:i
the TO rule 6.w~ ex: 8t h(pt), where
(1)
is the TD error, pt is position at time t, and the reward r(pt, pt+I) is 1 for any
move onto the platform, and 0 otherwise. In a slight alteration of the original rule,
the value V (p) is set to zero when p is at the goal, thus ensuring that the total
future rewards for moving onto the goal will be exactly 1. Such a modification
improves stability in the case of TD learning with overlapping basis functions.
The discount factor, I' was set to 0.99. Simultaneously the rat refines a policy,
which is represented by eight action cells. Each action cell (aj in figure 2) receives
a parameterised input at any position p: aj (p) = L:i qjdi (p). An action is chosen
stochastically with probabilities given by P(aj) = exp{2aj}/ L:k exp{2ak}. Action
weights are reinforced according to: 2
(2)
where 9j((Jt) is a gaussian function of the difference between the head direction
(Jt at time t and the preferred direction of the jth action cell. Figure 3 shows the
development of a policy over a few trials.
V(p)l
Triall
V(p)
1
0.5
0.5
0.
0.5
01
0.5
TrialS
Triall3
V(P)l
0.5
I
0
0.5
0:
0.5
0.5
-0.5 -------0.5
.---'--- 0
0.5
Figure 3: The RMW task: the value function gradually disseminates information
about reward proximity to all regions of the environment. Policies and paths are
also shown.
There is no analytical guarantee for the convergence of TD learning with policy
adaptation. However our simulations show that the algorithm always converges
for the RMW task. In a simulated arena of diameter 1m and with swimming speeds
of 20cm/s, the simulation matched the performance of the real rats very closely (see
figure S). This demonstrates that TD-based reinforcement learning is adequately
fast to account for the learning performance of real animals.
D. 1. Foster, R. G. M Morris and P. Dayan
148
2.2
Coordinate Learning
Although the learning of a value function and policy is appropriate for finding
a fixed platform, the actor-critic model does not allow the transfer of knowledge
from the task defined by one goal position to that defined by any other; thus it
could not generate the sort of one-trial learning that is shown by rats on the DMP
task (see figure 1b). This requires acquisition of some goal-independent know ledge
about s~ace. A natural mechanism for this is the path integration or self-motion
system. 0,10 However, path integration presents two problems. First, since the rat
is put into the maze in a different position for each trial, how can it learn consistent
coordinates across the whole maze? Second, how can a general, powerful, but slow,
behavioral learning mechanism such as TO be integrated with a specific, limited,
but fast learning mechanism involving spatial coordinates?
Since TO critic learning is based on enforcing consistency in estimates of future
reward, we can also use it to learn spatially consistent coordinates on the basis
of samples of self-motion. It is assumed that the rat has an allocentric frame of
reference. 1s The model learns parameterised estimates of the x and y coordinates
of all positions p: x(p) = Li w[ fi(P) and y(p) = Li wY h(p), Importantly, while
place cells were again critical in supporting spatial representation, they do not embody
a map of space. The coordinate functions, like the value function previously, have to
be learned.
As the simulated rat moves around, the coordinate weights {w[} are adjusted
according to:
t
Llwi ()( (Llxt
+ X(pt+l ) -
X(pt))
LA
t-
k h(pk)
(3)
k=1
where Llxt is the self-motion estimate in the x direction. A similar update is applied
to {wn. In this case, the full TO(A) algorithm was used (with A = 0.9); however
TD(O) could also have been used, taking slightly longer. Figure 4a shows the x and
y coordinates at early and late phases of learning. It is apparent that they rapidly
become quite accurate - this is an extremely easy task in an open field maze.
An important issue in the learning of coordinates is drift, since the coordinate
system receives no direct information about the location of the origin. It turns out
that the three controlling factors over the implicit origin are: the boundary of the
arena, the prior setting of the coordinate weights (in this case all were zero) and
the position and prior value of any absorbing area (in this case the platform). If the
coordinate system as a whole were to drift once coordinates have been established,
this would invalidate coordinates that have been remembered by the rat over long
periods. However, since the expected value of the prediction error at time steps
should be zero for any self-consistent coordinate mapping, such a mapping should
remain stable. This is demonstrated for a single run: figure 4b shows the mean
value of coordinates x evolving over trials, with little drift after the first few trials.
We modeled the coordinate system as influencing the choice of swimming direction
in the manner of an abstract action. I5 The (internally specified) coordinates of the
most recent goal position are stored in short term memory and used, along with the
current coordinates, to calculate a vector heading. This vector heading is thrown
into the stochastic competition with the other possible actions, governed by a
single weight which changes in a similar manner to the other action weights (as in
equation 2, see also fig 4d), depending on the TO error, and on the angular proximity
of the current head direction to the coordinate direction. Thus, whether the the
coordinate-based direction is likely to be used depends upon its past performance.
One simplification in the model is the treatment of extinction. In the DMP task,
Hippocampal Model ofRat Spatial Abilities Using 1D Learning
149
,.
"
TJUAL
d
III
.1
TRIAL
26
16
i:
~Ol
~o
!"
~o
Figure 4: The evolution of the coordinate system for a typical simulation run: a.)
coordinate outputs at early and late phases of learning, b.) the extent of drift in the
coordinates, as shown by the mean coordinate value for a single run, c.) a measure
A2
~ ~ {X (Pr.)-X -X(pr.)}2
?
o f coord mate error for the same run (7E =
r
r. (Np-l)N
' where k
r
r
r
indexes measurement points (max N p ) and r indexes runs (max N r ), Xr(Pk) is the
model estimate of X at position Pk, X(Pk) is the ideal estimate for a coordinate
system centred on zero, and Xr is the mean value over all the model coordinates,
d.) the increase during training of the probability of choosing the abstract action.
This demonstrates the integration of the coordinates into the control system.
real rats extinguish to a platform that has moved fairly quickly whereas the actorcritic model extinguishes far more slowly. To get around this, when a simulated
rat reaches a goal that has just been moved, the value and action weights are
reinitialised, but the coordinate weights wf and wf, and the weights for the abstract
action, are not.
3 RESULTS
The main results of this paper are the replication by simulation of rat performance
on the RMW and DMP tasks. Figures la and b show the course of learning for
the rats; figures Sa and b for the model. For the DMP task, one-shot acquisition is
apparent by the end of training.
4
DISCUSSION
We have built a model for one-trial spatial learning in the watermaze which uses
a single TD learning algorithm in two separate systems. One system is based on a
reinforcement learning that can solve general Markovian decision problems, and
the other is based on coordinate learning and is specialised for an open-field water
maze. Place cells in the hippocampus offer an excellent substrate for learning the
actor, the critic and the coordinates.
The model is explicit about the relationship between the general and specific learning systems, and the learning behavior shows that they integrate seamlessly. As
currently constituted, the coordinate system would fail if there were a barrier in
the maze. We plan to extend the model to allow the coordinate system to specify
abstract targets other than the most recent platform position - this could allow
it fast navigation around a larger class of environments. It is also important to
improve the model of learning 'set' behavior - the information about the nature of
D. 1. Foster; R. G. M. Mo"is and P. Dayan
150
14
a
b
12
10
.
.
?.
z
j:10\
~
~
~
12
~
S>
..
'"~ ............................................
0~D.~yl~~y~2~D~.y~3~D~.y~47~~yS~~~.~~~y~7~~~y~.7D.~y9~
Figure 5: a.) Performance of the actor-critic model on the RMW task, and b.)
performance of the full model on the DMP task. The data for comparison is shown
in figures la and b.
the DMP task that the rats acquire over the course of the first few days of training.
Interestingly, learning set is incomplete - on the first trial of each day, the rats
still aim for the platform position on the previous day, even though this is never
correct. 16 The significant differences in the path lengths on the first trial of each
day (evidence in figure Ib and figure 5b) come from the relative placements of the
platforms. However, the model did not use the same positions as the empirical
data, and, in any case, the model of exploration behavior is rather simplistic.
The model demonstrates that reinforcement learning methods are perfectly fast
enough to match empirical learning curves. This is fortunate, since, unlike most
models specifically designed for open-field navigation,6,4,5,2o RL methods can
provably cope with substantially more complicated tasks with arbitrary barriers,
etc, since they solve the temporal credit assignment problem in its full generality.
The model also addresses the problem that coordinates in different parts of the
same environment need to be mutually consistent, even if the animal only experiences some parts on separate trials. An important property of the model is that
there is no requirement for the animal to have any explicit knowledge of the relationship between different place cells or place field position, size or shape. Such a
requirement is imposed in various models. 9 ,4,6,2o
Experiments that are suggested by this model (as well as by certain others) concern the relationship between hippocampally dependent and independent spatial
learning. First, once the coordinate system has been acquired, we predict that
merely placing the rat at a new location would be enough to let it find the platform
in one shot, though it might be necessary to reinforce the placement e.g. by first
placing the rat in a bucket of cold water. Second, we know that the establishment
of place fields in an environment happens substantiallr faster than establishment
of one-shot or even ordinary learning to a platform. 2 We predict that blocking
plasticity in the hippocampus following the establishment of place cells (possibly
achieved without a platform) would not block learning of a platform. In fact, new
experiments show that after extensive pre-training, rats can perform one-trial learning in the same environment to new platform positions on the DMP task without
hippocampal synaptic plasticity. 16 This is in contrast to the effects of hippocampal
lesion, which completely disrupts performance. According to the model, coordinates will have been learned during pre-training. The full prediction remains
untested: that once place fields have been established, coordinates could be learned
in the absence of hippocampal synaptic plasticity. A third prediction follows from
evidence that rats with restricted hippocampal lesions can learn the fixed platform
Hippocampal Model of Rat Spatial Abilities Using TD Learning
151
task, but much more slowly, based on a gradual "shaping" procedure. 22 In our
model, they may also be able to learn coordinates. However, a lengthy training
procedure could be required, and testing might be complicated if expressing the
knowledge required the use of hippocampus dependent short-term memory for
the last platform location. I6
One way of expressing the contribution of the hippocampus in the model is to say
that its function is to provide a behavioural state space for the solution of complex
tasks. Hence the contribution of the hippocampus to navigation is to provide
place cells whose firing properties remain consistent in a given environment. It
follows that in different behavioural situations, hippocampal cells should provide
a representation based on something other than locations - and, indeed, there
is evidence for this. 8 With regard to the role of the hippocampus in spatial tasks,
the model demonstrates that the hippocampus may be fundamentally necessary
without embodying a map.
References
[1] Barto, AG & Sutton, RS (1981) BioI. Cyber., 43:1-8.
[2] Barto, AG, Sutton, RS & Anderson, CW (1983) IEEE Trans. on Systems, Man
[3]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
[21]
[22]
[23]
and Cybernetics 13:834-846.
Barto, AG, Sutton, RS & Watkins, CJCH (1989) Tech Report 89-95, CAIS, Univ.
Mass., Amherst, MA.
Blum, KI & Abbott, LF (1996) Neural Computation, 8:85-93.
Brown, MA & Sharp, PE (1995) Hippocampus 5:171-188.
Burgess, N, Reece, M & O'Keefe, J (1994) Neural Networks, 7:1065-1081.
Dayan, P (1991) NIPS 3, RP Lippmann et aI, eds., 464-470.
Eichenbaum, HB (1996) CurroOpin. Neurobiol., 6:187-195.
Gerstner, W & Abbott, LF (1996) J. Computational Neurosci. 4:79-94.
McNaughton, BL et a1 (1996) J. Exp. BioI., 199:173-185.
Morris, RGM et al (1982) Nature, 297:681-683.
O'Keefe, J & Dostrovsky, J (1971) Brain Res., 34(171).
Olton, OS & Samuelson, RJ (1976) J. Exp. Psych: A.B.P., 2:97-116.
Rudy, JW & Sutherland, RW (1995) Hippocampus, 5:375-389.
SchUltz, W, Dayan, P & Montague, PR (1997) Science, 275, 1593-1599.
Singh, SP Reinforcement learning with a hierarchy of abstract models.
Steele, RJ & Morris, RGM in preparation.
Sutton, RS (1988) Machine Learning, 3:9-44.
Taube, JS (1995) J. Neurosci. 15(1):70-86.
Tsitsiklis, IN & Van Roy, B (1996) Tech Report LIDS-P-2322, M.LT.
Wan, HS, Touretzky, OS & Redish, AD (1993) Proc. 1993 Connectionist Models
Summer School, Lawrence Erlbaum, 11-19.
Watkins, CJCH (1989) PhD Thesis, Cambridge.
Whishaw, IQ & Jarrard, LF (1996) Hippocampus
Wilson, MA & McNaughton, BL (1993) Science 261:1055-1058.
Reinforcement Learning for Call Admission
Control and Routing in Integrated Service
Networks
Oliver Mihatsch
Siemens AG
Corporate Technology, ZT IK 4
0-81730 Munich, Germany
email:oliver.mihatsch@
mchp.siemens.de
Peter Marbach"
LIDS
MIT
Cambridge, MA, 02139
email: marbach@mi t . edu
Miriam Schulte
Zentrum Mathematik
Technische UniversWit Miinchen
D-80290 Munich
Germany
John N. Tsitsiklis
LIDS
MIT
Cambridge, MA, 02139
email: jnt@mit. edu
Abstract
In integrated service communication networks, an important problem is
to exercise call admission control and routing so as to optimally use the
network resources. This problem is naturally formulated as a dynamic
programming problem, which, however, is too complex to be solved exactly. We use methods of reinforcement learning (RL), together with a
decomposition approach, to find call admission control and routing policies. The performance of our policy for a network with approximately
10 45 different feature configurations is compared with a commonly used
heuristic policy.
1 Introduction
The call admission control and routing problem arises in the context where a telecommunication provider wants to sell its network resources to customers in order to maximize
long term revenue. Customers are divided into different classes, called service types. Each
service type is characterized by its bandwidth demand, its average call holding time and
the immediate reward the network provider obtains, whenever a call of that service type is
? Author to whom correspondence should be addressed.
Reinforcement Learning for Call Admission Control and Routing
923
accepted. The control actions for maximizing the long term revenue are to accept or reject
new calls (Call Admission Control) and, if a call is accepted, to route the call appropriately through the network (Routing). The problem is naturally formulated as a dynamic
programming problem, which, however, is too complex to be solved exactly. We use the
methodology of reinforcement learning (RL) to approximate the value function of dynamic
programming. Furthermore, we pursue a decomposition approach, where the network is
viewed as consisting of link processes, each having its own value function. This has the
advantage, that it allows a decentralized implementation of the training methods of RL
and a decentralized implementation of the call admission control and routing policies. Our
method learns call admission control and routing policies which outperform the commonly
used heuristic "Open-Shortest-Path-First" (OSPF) policy.
In some earlier related work, we applied RL to the call admission problem for a single
communication link in an integrated service environment. We found that in this case, RL
methods performed as well, but no better than, well-designed heuristics. Compared with
the single link problem, the addition of routing decisions makes the network problem more
complex and good heuristics are not easy to derive.
2 Call Admission Control and Routing
We are given a telecommunication network consisting of a set of nodes N = {I, ... , N} and
a set of Iinks .c = {I, ... , L}, where link I has a a total capacity of B(l) units of bandwidth.
We support a set M = {I, "', M} of different service types, where a service type m is
characterized by its bandwidth demand b(m), its average call holding time I/v(m) (here
we assume that the call holding times are exponentially distributed) and the immediate
reward c( m) we obtain, whenever we accept a call of that service type. A link can carry
simultaneously any combination of calls, as long as the bandwidth used by these calls does
not exceed the total bandwidth of the link (Capacity Constraint). When a new call of
service type m requests a connection between a node i and a node j, we can either reject
or accept that request (Call Admission Control). If we accept the call, we choose a route
out of a list of predefined routes (Routing). The call then uses b(m) units of bandwidth
on each link along that route for the duration of the call. We can, therefore, only choose
a route, which does not violate the capacity constraints of its links, if the call is accepted.
Furthermore, if we accept the call, we obtain an immediate reward c( m). The objective
is to exercise call admission control and routing in such a way that the long term revenue
obtained by accepting calls is maximized.
We can formulate the call admission control and routing problem using dynamic programming (e. g. Bertsekas, 1995). Events w which incur state transitions, are arrivals of new
calls and call terminations. The state Xt at time t consists of a list for each route, indicating
how many calls of each service type are currently using that route. The decision/control Ut
applied at the time t of an arrival of a new call is to decide, whether to reject or accept the
call, and, if the call is accepted, how to route it through the network. The objective is to
learn a policy that assigns decisions to each state so as to
where E{?} is the expectation operator, tk is the time when the kth event happens,
g( Xtk' Wk, Ut,,) is the immediate reward associated with the kth event, and f3 is a discount
factor that makes immediate rewards more valuable than future ones.
924
P Marbach, O. Mihatsch, M. Schulte and 1. N. Tsitsiklis
3 Reinforcement Learning Solution
RL methods solve optimal control (or dynamic programming) problems by learning good
approximations to the optimal value function r, given by the solution to the Bellman optimality equation which takes the following form for the caB admission control and routing
problem
J*(x) = Er {e- th } Ew { max [g(x,w, u)
ueU(x)
+ J*(X
I )]}
where U ( x) is the set of control actions available in the current state x, T is the time when
the first event w occurs and x' is the successor state. Note that x' is a deterministic function
of the current state x, the control u and the event w.
RL uses a compact representation j (', 0) to learn and store an estimate of J" (.). On each
event, i(., 0) is both used to make decisions and to update the parameter vector e. In the
caB admission control and routing problem, one has only to choose a control action when
a new call requests a connection. In such a case, J(,,0) is used to choose a control action
according to the formula
u=arg max [g(x,w,u)
ueU(x)
+ J(X', e)]
(1)
This can be expressed in words as follows.
Decision Making: When a new call requests a connection, use J (', e) to evaluate, for each
permissible route, the successor state x' we transit to, when we choose that route, and pick
a route which maximizes that value. If the sum of the immediate reward and the value
associated with this route is higher than the value of the current state, route the call over
that route; otherwise reject the call.
Usually, RL uses a global feature extractor f(x) to form an approximate compact representation of the state of the system, which forms the input to a function approximator
i(., e). Sutton's temporal difference (TO()'? algorithms (Sutton, 1988) can then be used
to train i(., 0) to learn an estimate of J*. Using ID(O), the update at the kth event takes
the following form
where
dk
e-/J(t/c-t/c-d (g(Xt/c, Wk, Ut/c)
+ J(!(Xt/c), ek-I))
-J(I(Xt/C_l)' Ok-I)
and where 'Yk is a small step size parameter and Utk is the control action chosen according
to the decision making rule described above.
Here we pursue an approach where we view the network as being composed of link processes. Furthermore, we decompose immediate rewards g( Xtk' Wk, Ut/c) associated with the
kth event, into link rewards g(l) (Xt/c, Wk, Ut/c) such that
L
g(Xtk' Wk. Ut/c)
= L gil) (Xtl:' Wk, UtI:)
1=1
We then define, for each link I, a value function J(I) (I(l) (x), e( I?), which is interpreted as
an estimate of the discounted long term revenue associated with that link. Here, f(l) defines
a local feature, which forms the input to the value function associated with link I. To obtain
925
Reinforcement Learning for Call Admission Control and Routing
an approximation of J* (x), the functions
](1) (J(l)
(x), 0(1)) are combined as follows
L
L
](1) (J(I)
(x), (J(l)).
1=1
At each event, we update the parameter vector (J(l) of link 1, only if the event is associated
with the link. Events associated with a link 1 are arrivals of new calls which are potentially
routed over link 1 and termination of calls which were routed over the link I. The update
rule of the parameter vector 0(1) is very similar to the TD(O) algorithm described above
(J(l) k -
(J(l)
k-l
+ "V(I)d(I)V
()](l) (/(1) (x
) (J(l) )
Ik
k
9 I
tk_1 , k-l
(2)
where
e - i3(t~') -t~/~ I) (g(l) (x t(l), Wkl ) , Ut(l))
k
_](1) (J(l)
(x t (/)
),
+ ](l) (J(l) (xt(l?), (Jkl~ 1))
k
(3)
k
(Jkl~l)
k-I
and where Ill) is a small step size parameter and tr) is the time when the kth event will
associated with link 1 occurs. Whenever a new call of a service of type m is routed over
a route r which contains the link i, the immediate reward g(l) associated with the link i is
equal to c( m) / #r, where #r is the number of links along the route r. For all other events,
the immediate reward associated with link 1 is equal to O.
The advantage of this decomposition approach is that it allows decentralized training and
decentralized decision making. Furthermore, we observed that this decomposition approach leads to much shorter training times for obtaining an approximation for J* than the
approach without decomposition. All these features become very important if one considers applying methods of RL to large integrated service networks supporting a fair number
of different service types.
We use exploration to obtain the states at which we update the parameter vector O. At each
state, with probability p == 0.5, we apply a random action, instead of the action recommended by the current value function, to generate the next state in our training trajectory.
However, the action Ut(I), that is used in the update rule (3), is still the one chosen ack
cording to the rule given in (1). Exploration during the training significantly improved the
performance of the policy.
Table I: Service Types.
SERVICE TYPE m
BANDWIDTH DEMAND b( m)
AVERAGE HOLDING TIME l/v(m)
IMMEDIATE REWARD c( m)
1
2
3
1
10
3
10
2
5
2
50
1
4 Experimental Results
In this section, we present experimental results obtained for the case of an integrated service
network consisting of 4 nodes and 12 unidirectional links. There are two different classes
of links with a total capacity of 60 and 120 units of bandwidth, respectively (indicated by
thick and thin arrows in Figure 1). We assume a set M == {I, 2, 3} of three different service types. The corresponding bandwidth demands, average holding times and immediate
926
P. Marbach, O. Mihatsch, M Schulte and 1. N. Tsitsiklis
Figure 1: Telecommunication Network Consisting of 4 Nodes and 12 Unidirectional Links.
PERFORMANCE DURING LEAANING
Figure 2: Average Reward per TIme Unit During the Whole Training Phase of 10 7 Steps
(Solid) and During Shorter Time Windows of 10 5 Steps (Dashed).
rewards are given in Table 1. Call arrivals are modeled as independent Poisson processes,
with a separate mean for each pair of source and destination nodes and each service type.
Furthermore, for each source and destination node pair, the list of possible routes consists
of three entries: the direct path and the two alternative 2-hop-routes.
We compare the policy obtained through RL with the commonly used heuristic OSPF
(Open Shortest Path First). For every pair of source and destination nodes, OSPF orders
the list of predefined routes. When a new call arrives, it is routed along the first route in the
corresponding list, that does not violate the capacity constraint; if no such a route exists,
the call is rejected. We use the average reward per unit time as performance measure to
compare the two policies.
For the RL approach, we use a quadratic approximator, which is linear with respect to the
parameters ()(I), as a compact representation of ](1). Other approximation architectures
were tried, but we found that the quadratic gave the best results with respect to both the
speed of convergence and the final performance. As inputs to the compact representation
Reinforcement Learning for Call Admission Control and Routing
927
AVERAGE REWARD
potential reward
reward obtained by RL
reward obtained by OSPF
o
50
150
100
reward per time
200
250
un~
COMPARISON OF REJECTION RATES
o
5
10
15
20
25
30
35
40
45
50
percentage of calls rejected
Figure 3: Comparison of the Average Rewards and Rejection Rates of the RL and OSPF
Policies.
ROUTING (OSPF)
direct link
alternative route no. 1
alternative route no. 2
o
10
20
30
~
..
----~~-----.0:-=
50
~
ro
~
90
100
80
90
100
percentage of calls routed on direct and alternative paths
ROUTING (RL)
direct link
alternative route no. 1
alternative route no. 2
o
10
20
30
40
50
60
70
percentage of calls routed on direct and alternative paths
Figure 4: Comparison of the Routing Behaviour of the RL and OSPF Policies.
](/), we use a set of local features, which we chose to be the number of ongoing calls
of each service type on link l. For the 4-node network, there are approximately 1.6. 1045
different feature configurations. Note that the total number of possible states is even higher.
The results of the case studies are given in in Figure 2 (Training Phase), Figure 3 (Performance) and Figure 4 (Routing Behaviour). We give here a summary of the results.
Training Phase: Figure 2 shows the average reward of the RL policy as a function of
the training steps. Although the average reward increases during the training, it does not
exceed 141, the average reward of the heuristic OSPF. This is due to the high amount of
exploration in the training phase.
Performance Comparison: The policy obtained through RL gives an average reward of
212, which as about 50% higher than the one of 141 achieved by OSPF. Furthermore, the
RL policy reduces the number of rejected calls for all service types. The most significant
reduction is achieved for calls of service type 3, the service type, which has the highest
P. Marbach, O. Mihatsch, M Schulte and I. N. Tsitsiklis
928
immediate reward. Figure 3 also shows that the average reward of the RL policy is close
to the potential average reward of 242, which is the average reward we would obtain if
all calls were accepted. This leaves us to believe that the RL policy is close to optimal.
Figure 4 compares the routing behaviour of the RL control policy and OSPF. While OSPF
routes about 15% - 20% of all calls along one of the alternative 2-hop-routes, the RL
policy almost uses alternative routes for calls of type 3 (about 25%) and routes calls of the
other two service types almost exclusively over the direct route. This indicates, that the RL
policy uses a routing scheme, which avoids 2-hop-routes for calls of service type 1 and 2,
and which allows us to use network resources more efficiently.
5
Conclusion
The call admission control and routing problem for integrated service networks is naturally formulated as a dynamic programming problem, albeit one with a very large state
space. Traditional dynamic programming methods are computationally infeasible for such
large scale problems. We use reinforcement learning, based on Sutton's (1988) T D(O),
combined with a decomposition approach, which views the network as consisting of link
processes. This decomposition has the advantage that it allows decentralized decision making and decentralized training, which reduces significantly the time of the training phase.
We presented a solution for an example network with about 10 45 different feature configurations. Our RL policy clearly outperforms the commonly used heuristic OSPF. Besides
the game of backgammon (Tesauro, 1992), the elevator scheduling (Crites & Barto, 1996),
the jop-shop scheduling (Zhang & Dietterich, 1996) and the dynamic channel allocation
(Singh & Bertsekas, 1997), this is another successful application of RL to a large-scale
dynamic programming problem for which a good heuristic is hard to find.
References
Bertsekas, D. P; (1995) Dynamic Programming and Optimal Control. Athena Scientific,
Belmont, MA.
Crites, R. H., Barto, A. G. (1996) Improving elevator performance using reinforcement
learning. In D. S. Touretzky, M. C. Mozer and M. E. Hasselmo (eds.), Advances in Neural
Information Processing Systems 8, pp. 1017-1023. Cambridge, MA: MIT Press.
Singh, S., Bertsekas, D. P. (1997) Reinforcement learning for dynamic channel allocation
in cellular telephone systems. To appear in Advances in Neural Information Processing
Systems 9, Cambridge, MA: MIT Press.
Sutton, R. S. (1988) Learning to predict by the method of temporal differences. Machine
Learning, 3:9-44.
Tesauro, G. J. (1992) Practical issues in temporal difference learning. Machine Learning,
8(3/4):257-277.
Zhang, W., Dietterich, T. G. (1996) High performance job-shop scheduling with a timedelay TD(>.) network. In D. S. Touretzky, M. C. Mozer and M. E. Hasselmo (eds.), Advances in Neural Information Processing Systems 8, pp. 1024-1030. Cambridge. MA: MIT
Press.
| 1382 |@word h:1 trial:19 version:1 hippocampus:16 extinction:1 open:5 termination:2 gradual:2 simulation:4 r:4 decomposition:7 tried:1 pick:1 tr:1 solid:1 shot:3 carry:1 reduction:1 configuration:3 contains:1 exclusively:1 united:1 tuned:2 interestingly:1 past:1 outperforms:1 current:7 si:1 jkl:2 john:1 refines:1 belmont:1 plasticity:3 shape:1 designed:2 update:7 instantiate:1 leaf:1 short:3 accepting:1 rgm:2 location:8 miinchen:1 node:9 zhang:2 admission:18 along:5 direct:7 become:2 ik:2 replication:1 consists:2 xtl:1 behavioral:1 manner:3 acquired:1 indeed:1 expected:1 behavior:3 disrupts:1 embody:1 ol:1 brain:1 bellman:1 discounted:1 td:10 little:1 window:1 vertebrate:1 matched:1 maximizes:1 mass:1 cm:1 neurobiol:1 substantially:1 psych:1 interpreted:1 pursue:2 whilst:1 finding:1 ag:4 cjch:2 guarantee:2 temporal:6 every:1 computa:1 exactly:3 edin:1 demonstrates:4 ro:1 uk:1 control:28 unit:5 grant:1 internally:1 appear:2 bertsekas:4 sutherland:1 influencing:1 service:26 local:2 sutton:8 ak:1 id:1 firing:2 path:11 approximately:3 might:2 chose:1 extinguishes:1 coord:1 challenging:2 limited:1 practical:1 testing:1 block:1 lf:3 xr:2 cold:1 procedure:2 area:1 empirical:2 evolving:1 reject:4 significantly:2 pre:2 integrating:1 word:1 get:1 onto:2 close:2 operator:1 scheduling:3 put:1 context:2 applying:1 map:2 demonstrated:1 imposed:1 customer:2 maximizing:1 deterministic:1 starting:1 duration:1 formulate:1 assigns:1 rule:6 importantly:1 stability:1 coordinate:46 mcnaughton:2 rmw:7 pt:6 gm:1 controlling:1 target:1 hierarchy:1 substrate:1 programming:9 us:8 origin:2 roy:1 blocking:1 observed:1 role:1 solved:3 calculate:1 region:1 highest:1 valuable:1 yk:1 mozer:2 environment:7 reward:33 dynamic:11 singh:3 incur:1 upon:1 basis:4 completely:1 dmp:10 montague:1 represented:1 various:1 train:1 univ:2 distinct:1 fast:4 reece:1 universwit:1 choosing:1 refined:1 apparent:2 ace:1 quite:1 solve:4 larger:1 say:1 whose:1 otherwise:2 heuristic:8 ability:4 final:1 advantage:3 analytical:1 wih:1 interaction:1 adaptation:1 rapidly:1 moved:2 competition:1 convergence:2 requirement:2 converges:1 tk:1 oo:1 depending:1 ac:1 iq:1 derive:1 school:1 miriam:1 job:1 sa:1 ibn:1 come:2 direction:7 untested:1 thick:1 closely:1 correct:1 stochastic:1 exploration:4 routing:24 successor:2 require:1 behaviour:3 decompose:1 learning1:1 adjusted:1 proximity:2 around:3 credit:1 normal:1 exp:5 lawrence:1 mapping:2 predict:3 mo:2 early:3 a2:1 proc:1 currently:2 hasselmo:2 mit:7 clearly:1 gaussian:1 always:1 aim:1 establishment:3 rather:1 i3:1 barto:5 wilson:1 jnt:1 backgammon:1 indicates:1 seamlessly:1 tech:2 contrast:3 wf:2 dayan:6 dependent:2 integrated:8 accept:6 hidden:1 germany:2 provably:1 issue:2 arg:1 ill:1 development:1 plan:1 spatial:13 platform:19 integration:3 animal:3 fairly:1 field:8 once:3 never:1 having:1 schulte:4 f3:1 hop:3 equal:2 placing:2 sell:1 thin:1 hippocampally:1 future:3 report:2 cfn:1 np:1 richard:1 few:4 fundamentally:1 others:1 connectionist:1 escape:2 composed:1 simultaneously:2 zentrum:1 delayed:2 elevator:2 phase:7 consisting:5 thrown:1 circular:1 arena:2 navigation:4 arrives:1 predefined:2 accurate:1 oliver:2 necessary:2 experience:1 shorter:2 incomplete:1 mihatsch:5 re:1 dostrovsky:1 earlier:1 markovian:1 assignment:1 ordinary:1 technische:1 entry:1 successful:1 erlbaum:1 too:2 optimally:1 stored:1 combined:2 rudy:1 amherst:1 destination:3 yl:1 decoding:1 pool:1 together:1 e25:1 quickly:1 again:1 thesis:1 wan:1 slowly:2 possibly:1 choose:5 stochastically:1 ek:1 li:2 account:1 potential:2 de:1 centred:1 alteration:1 redish:1 wk:6 depends:1 ad:1 later:1 performed:1 view:2 sort:1 complicated:2 unidirectional:2 djf:1 actorcritic:1 contribution:3 efficiently:1 reinforced:1 maximized:1 samuelson:1 modelled:2 disseminates:1 provider:2 trajectory:1 cybernetics:1 reach:1 touretzky:3 whenever:3 ed:4 email:4 synaptic:2 lengthy:1 acquisition:3 pp:2 naturally:3 associated:10 mi:1 treatment:1 holdsworth:1 knowledge:3 ut:8 improves:1 shaping:1 actually:1 ok:1 higher:3 day:11 methodology:1 specify:2 improved:1 jw:1 though:2 generality:2 anderson:1 parameterised:2 implicit:1 angular:1 just:1 furthermore:6 rejected:3 receives:2 y9:1 o:2 overlapping:1 defines:1 llxt:2 aj:4 indicated:1 scientific:1 believe:1 effect:1 steele:1 brown:1 dietterich:2 adequately:1 evolution:1 hence:1 spatially:2 attractive:1 llwi:1 during:7 self:4 game:1 timedelay:1 rat:30 hippocampal:10 ack:1 motion:3 novel:3 fi:1 absorbing:1 rl:26 exponentially:1 extend:1 slight:1 measurement:1 significant:2 expressing:2 cambridge:7 ai:1 pew:1 consistency:1 i6:1 marbach:5 centre:3 funded:1 l3:1 moving:1 stable:1 actor:9 longer:1 surface:1 invalidate:1 etc:1 something:1 j:1 own:1 recent:3 tesauro:2 route:31 certain:2 store:1 remembered:2 utk:1 maximize:1 period:1 taube:1 shortest:2 recommended:1 dashed:1 full:4 corporate:1 rj:2 violate:2 reduces:2 match:3 faster:1 characterized:2 offer:1 long:7 divided:1 y:1 a1:1 ensuring:1 prediction:3 involving:2 simplistic:1 expectation:1 poisson:1 represent:1 achieved:3 cell:18 whereas:1 want:1 addition:1 addressed:1 diagram:1 source:3 appropriately:1 permissible:1 unlike:1 cyber:1 call:60 near:1 ideal:1 exceed:2 crichton:1 easy:2 wn:1 iii:1 enough:2 hb:1 burgess:1 gave:1 architecture:1 perfectly:1 bandwidth:9 idea:1 whether:2 routed:6 peter:2 nine:1 action:19 latency:3 clear:1 detailed:1 amount:1 discount:2 extensively:1 morris:4 embodying:1 diameter:1 rw:1 generate:2 outperform:1 percentage:3 nsf:1 gil:1 neuroscience:2 per:4 key:1 four:1 demonstrating:1 blum:1 abbott:2 asymptotically:1 swimming:2 merely:1 sum:1 run:5 powerful:1 respond:1 i5:1 telecommunication:3 place:20 almost:2 decide:1 uti:1 decision:10 ki:1 summer:1 simplification:1 correspondence:1 quadratic:2 activity:1 placement:2 constraint:3 speed:2 cloudy:1 extremely:1 optimality:1 eichenbaum:1 xtk:3 munich:2 according:5 mcdonnell:1 combination:1 request:4 across:1 slightly:1 remain:2 lid:3 modification:1 happens:2 making:4 gradually:1 pr:3 restricted:1 bucket:1 computationally:2 resource:3 equation:2 mutually:1 remains:2 previously:1 turn:1 behavioural:2 mechanism:3 fail:1 needed:1 know:2 mathematik:1 end:1 available:1 decentralized:6 eight:1 apply:1 appropriate:1 alternative:9 allocentric:2 rp:1 ledge:1 original:1 scholarship:1 build:1 bl:2 move:2 objective:2 occurs:2 traditional:1 exhibit:2 kth:5 distance:1 separate:3 reinforce:1 simulated:3 cw:1 street:1 link:30 capacity:5 athena:1 transit:1 whom:1 extent:1 considers:1 cellular:1 water:4 enforcing:1 length:1 besides:1 modeled:2 index:2 relationship:3 acquire:1 kingdom:1 potentially:1 holding:5 noni:1 implementation:2 zt:1 policy:26 perform:2 markov:1 mate:1 supporting:2 immediate:12 situation:1 communication:2 head:2 frame:1 wkl:1 arbitrary:3 sharp:1 watermaze:4 drift:4 david:1 pair:3 required:3 specified:1 extensive:1 whishaw:1 connection:3 learned:3 eh8:1 established:2 nip:1 trans:1 address:1 able:1 suggested:1 below:1 wy:1 usually:1 built:1 max:4 memory:4 critical:1 event:13 natural:1 scheme:2 improve:1 shop:2 technology:1 prior:2 understanding:1 asymptotic:1 relative:1 allocation:2 approximator:2 revenue:4 foundation:1 integrate:1 consistent:5 foster:4 critic:11 cording:1 course:2 changed:1 summary:1 placed:1 last:1 jth:1 heading:2 tsitsiklis:5 infeasible:1 allow:3 taking:1 barrier:2 mchp:1 edinburgh:3 regard:1 boundary:1 curve:1 van:1 distributed:1 transition:1 maze:7 avoids:1 author:1 commonly:4 reinforcement:17 schultz:1 far:1 cope:1 approximate:2 lippmann:1 obtains:1 preferred:1 compact:4 global:1 conceptual:1 assumed:1 themodel:1 un:1 table:2 lip:1 learn:8 transfer:1 nature:2 channel:2 obtaining:1 improving:1 excellent:1 complex:4 gerstner:1 did:1 pk:4 main:1 constituted:1 neurosci:2 sp:1 whole:3 arrow:1 crites:2 arrival:4 lesion:2 competent:1 fair:1 fig:1 slow:1 position:18 comprises:1 explicit:2 ueu:2 fortunate:1 exercise:2 governed:1 pe:1 ib:1 watkins:2 late:2 third:1 extractor:1 learns:4 formula:1 specific:2 xt:6 jt:2 showing:1 er:1 system2:1 list:5 dk:1 evidence:3 concern:1 exists:1 albeit:1 extinguish:1 importance:1 keefe:2 cab:2 phd:1 demand:4 rodent:1 rejection:2 specialised:1 lt:1 likely:1 expressed:1 ma:10 bioi:2 goal:9 formulated:3 viewed:1 absence:1 man:1 change:1 hard:1 typical:1 specifically:1 telephone:1 principal:1 total:5 called:1 accepted:5 experimental:2 la:3 siemens:2 ew:1 indicating:1 support:2 arises:1 preparation:1 ongoing:1 evaluate:1 ex:1 |
420 | 1,383 | The Asymptotic Convergence-Rate of
Q-Iearning
es. Szepesvari*
Research Group on Artificial Intelligence, "Jozsef Attila" University,
Szeged, Aradi vrt. tere 1, Hungary, H-6720
szepes@math.u-szeged.hu
Abstract
In this paper we show that for discounted MDPs with discount
factor, > 1/2 the asymptotic rate of convergence of Q-Iearning
is O(1/t R (1-1') if R(1 - ,) < 1/2 and O( Jlog log tit) otherwise
provided that the state-action pairs are sampled from a fixed probability distribution. Here R = Pmin/Pmax is the ratio of the minimum and maximum state-action occupation frequencies. The results extend to convergent on-line learning provided that Pmin > 0,
where Pmin and Pmax now become the minimum and maximum
state-action occupation frequencies corresponding to the stationary distribution.
1
INTRODUCTION
Q-Iearning is a popular reinforcement learning (RL) algorithm whose convergence is
well demonstrated in the literature (Jaakkola et al., 1994; Tsitsiklis, 1994; Littman
and Szepesvari, 1996; Szepesvari and Littman, 1996). Our aim in this paper is to
provide an upper bound for the convergence rate of (lookup-table based) Q-Iearning
algorithms. Although, this upper bound is not strict, computer experiments (to be
presented elsewhere) and the form of the lemma underlying the proof indicate that
the obtained upper bound can be made strict by a slightly more complicated definition for R. Our results extend to learning on aggregated states (see (Singh et al.,
1995? and other related algorithms which admit a certain form of asynchronous
stochastic approximation (see (Szepesv<iri and Littman, 1996?.
Present address: Associative Computing, Inc., Budapest, Konkoly Thege M. u. 29-33,
HUNGARY-1121
The Asymptotic Convergence-Rate of Q-leaming
2
1065
Q-LEARNING
Watkins introduced the following algorithm to estimate the value of state-action
pairs in discounted Markovian Decision Processes (MDPs) (Watkins, 1990):
Here x E X and a E A are states and actions, respectively, X and A are finite. It is
assumed that some random sampling mechanism (e.g. simulation or interaction with
a real Markovian environment) generates random samples of form (Xt, at, Yt, rt),
where the probability of Yt given (xt,at) is fixed and is denoted by P(xt,at,Yt),
E[rt I Xt, at] = R(x, a) is the immediate average reward which is received when
executing action a from state x, Yt and rt are assumed to be independent given the
history of the learning-process, and also it is assumed that Var[rt I Xt, at] < C for
some C > O. The values 0 ~ at(x,a) ~ 1 are called the learning rate associated
with the state-action pair (x, a) at time t. This value is assumed to be zero if
(x,a) =J (xt,at), i.e. only the value of the actual state and action is reestimated in
each step. If
00
L at(x, a) =
00
(2)
00
(3)
t=l
and
00
L a;(x, a) <
t=l
then Q-Iearning is guaranteed to converge to the only fixed point Q* of the operator
T : lR X x A ~ lR XxA defined by
(TQ)(x,a)
= R(x, a)
+, L
P(x,a,y)mFQ(y,b)
yEX
(convergence proofs can be found in (Jaakkola et al., 1994; TSitsiklis, 1994; Littman
and Szepesv.hi, 1996; Szepesvari and Littman, 1996)). Once Q* is identified the
learning agent can act optimally in the underlying MDP simply by choosing the
action which maximizes Q* (x, a) when the agent is in state x (Ross, 1970; Puterman,
1994).
3
THE MAIN RESULT
Condition (2) on the learning rate at(x, a) requires only that every state-action pair
is visited infinitely often, which is a rather mild condition. In this article we take the
stronger assumption that {(Xt, at) h is a sequence of independent random variables
with common underlying probability distribution. Although this assumption is not
essential it simplifies the presentation of the proofs greatly. A relaxation will be
discussed later. We further assume that the learning rates take the special form
at ( x, a )
=
{
l
,
if (x,a) = (xt,a);
~
Ol,x,a,
otherwise,
0,
where St (x, a) is the number of times the state-action pair was visited by the process
(xs, as) before time step t plus one, i.e. St(x, a) = 1 + #{ (xs, as) = (x, a), 1 ~ s ~
C. Szepesvari
1066
t }. This assumption could be relaxed too as it will be discussed later. For technical
reasons we further assume that the absolute value of the random reinforcement
signals Tt admit a common upper bound. Our main result is the following:
3.1 Under the above conditions the following relations hold asymptotically and with probability one:
THEOREM
IQt(x, a) - Q*(x, a)1
~ tR(~-'Y)
(4)
and
*
JIOg log t
IQt(x,a) - Q (x,a)1 ~ B
t'
(5)
for some suitable constant B > O. Here R = Pmin/Pmax, where Pmin =
min(z,a) p(x, a) and Pmax = max(z,a) p(x, a), where p(x, a) is the sampling probability of (x, a).
Note that if'Y 2: 1 - Pmax/2pmin then (4) is the slower, while if'Y < 1 - Pmax/2Pmin
then (5) is the slower. The proof will be presented in several steps.
Step 1. Just like in (Littman and Szepesvari, 1996) (see also the extended version
(Szepesvciri and Littman, 1996)) the main idea is to compare Qt with the simpler
process
Note that the only (but rather essential) difference between the definition of Qt and
that of Qt is the appearance of Q* in the defining equation of Qt. Firstly, notice
that as a consequence of this change the process Qt clearly converges to Q* and
this convergence may be investigated along each component (x, a) separately using
standard stochastic-approximation techniques (see e.g. (Was an , 1969; Poljak and
Tsypkin, 1973)).
Using simple devices one can show that the difference process At(x, a)
= IQt(x, a)-
at(x, a)1 satisfies the following inequality:
A t + 1 (x, a) ~ (1 - Ot(x, a))At(x, a)
+ 'Y0t(x, a)(IIAtll + lIat -
Q*II).
(7)
Here 11?11 stands for the maximum norm. That is the task of showing the convergence
rate of Qt to Q* is reduced to that of showing the convergence rate of At to zero.
Step 2. We simplify the notation by introducing the abstract process whose update
equation is
(8)
where i E 1,2, ... , n can be identified with the state-action pairs, Xt with At,
with Qt - Q*, etc. We analyze this process in two steps. First we consider
processes when the "perturbation-term" f.t is missing. For such processes we have
the following lemma:
f.t
3.2 Assume that 771,1]2, ... ,'TIt, . .. are independent random variables with
a common underlying distribution P{TJt = i) = Pi > O. Then the process Xt defined
LEMMA
The Asymptotic Convergence-Rate of Q-leaming
1067
by
(9)
satisfies
IIxtil =
wi~h
OCR(~--Y?)
probability one (w.p.1), where R = mini Pi/ maxi Pi.
Proof. (Outline) Let To
Tk+l
= 0 and
= min{ t ~ Tk IVi = 1 . . . n, 3s = s(i) : 1]8 = i},
i.e. Tk+1 is the smallest time after time Tk such that during the time interval
[Tk + 1, Tk+d all the components of XtO are "updated" in Equation (9) at least
once. Then
(10)
where Sk = maxi Sk(i) . This inequality holds because if tk(i) is the last time in
[Tk + 1, T k+1] when the ith component is updated then
XT"+l+1(i)
Xtk(i)+l(i) = (1-1/St/o(i?Xt,,(i)(i) + ,/St,,(i) II X t,,(i) 011
< (l-l/S t,,(i?lI x t/o(i)OIl +,/St,,(i)lIxt,,(i)OIl
=
=
<
(1 -1St,,(i)
-,) IIXt,,(i) 011
(1- 1;k') IIXT,,+1011,
where it was exploited that
time yields
Ilxtll
is decreasing. Now, iterating (10) backwards in
X7Hl(-):::
IIxolin (1- 1~ 'Y).
Now, consider the following approximations: Tk ~ Ck, where C ~ 1/Pmin (C can be
computed explicitly from {Pi}), Sk ~ PmaxTk+1 ~ Pmax/Pmin(k + 1) ~ (k + 1)/ Ro,
where Ro = 1/CPmax' Then, using Large Deviation's Theory,
IT
k-l (
j=O
1_,) ~II
k-l (
1-Sj
1-
j=O
holds w.p.1. Now, by defining s = Tk
Ro(1 _
.
J
,?) ~(1) Ro(l--Y)
+1
+ 1 so that
(11)
k
siC ~ k we get
which holds due to the monotonicity of Xt and l/k Ro (l--y) and because R
Pmin/Pmax ~ Ro.
0
Step 3. Assume that, > 1/2. Fortunately, we know by an extension of the Law of
the Iterated Logarithm to stochastic approximation processes that the convergence
C. Szepesvari
1068
rate of IIOt -Q*II is 0 (y'loglogt/t) (the uniform boundedness ofthe random reinforcement signals must be exploited in this step) (Major, 1973). Thus it is sufficient
to provide a convergence rate estimate for the perturbed process, Xt, defined by (8),
when f.t = Cy'loglogt/t for some C > O. We state that the convergence rate of f.t
is faster than that of Xt. Define the process
ZHI (i)
= { (1 - ~~l)) Zt(i),
Zt (i),
if 7Jt = i;
if 7Jt f. i.
(12)
This process clearly lower bounds the perturbed process, Xt. Obviously, the convergence rate of Zt is O(l/tl-'Y) which is slower than the convergence rate of f.t
provided that, > 1/2, proving that f.t must be faster than Xt. Thus, asymptotically f.t ~ (1/, - l)xt, and so Ilxtll is decreasing for large enough t. Then, by an
argument similar to that of used in the derivation of (10), we get
XTIo+1+1(i)
~ (1- 1~k') II XTk +1 II ~ ~ f.Tk,
(13)
where Sk = mini Sk(i). By some approximation arguments similar to that of Step 2,
together with the bound (l/n71)
s71- 3 / 2Jloglogs ~ s-1/2Jloglogs, 1 > 7J > 0,
which follows from the mean-value theorem for integrals and the law of integration
by parts, we get that Xt ~ O(l/t R (l-'Y?). The case when , ~ 1/2 can be treated
similarly.
2::
Step 5. Putting the pieces together and applying them for At = Ot - Qt yields
Theorem 3.1.
4
DISCUSSION AND CONCLUSIONS
The most restrictive of our conditions is the assumption concerning the sampling
of (Xt, at). However, note that under a fixed learning policy the process (Xt, at)
is a (non-stationary) Markovian process and if the learning policy converges in
the sense that limt-+oo peat 1Ft) = peat I Xt) (here Ft stands for the history of the
learning process) then the process (Xt, at) becomes eventually stationary Markovian
and the sampling distribution could be replaced by the stationary distribution of
the underlying stationary Markovian process. If actions become asymptotically
optimal during the course of learning then the support of this stationary process
will exclude the state-action pairs whose action is sub-optimal, i.e. the conditions
of Theorem 3.1 will no longer be satisfied. Notice that the proof of convergence of
such processes still follows very similar lines to that of the proof presented here (see
the forthcoming paper (Singh et al., 1997)), so we expect that the same convergence
rates hold and can be proved using nearly identical techniques in this case as well.
A further step would be to find explicit expressions for the constant B of Theorem 3.1. Clearly, B depends heavily on the sampling of (Xt, at), as well as the
transition probabilities and rewards of the underlying MDP. Also the choice of harmonic learning rates is arbitrary. If a general sequence at were employed then the
artificial "time" Tt (x, a) = 1/IT}=o (1 - at (x, a)) should be used (note that for the
harmonic sequence Tt(x, a) ~ t). Note that although the developed bounds are
asymptotic in their present forms, the proper usage of Large Deviation's Theory
would enable us to develope non-asymptotic bounds.
The Asymptotic Convergence-Rate ofQ-learning
1069
Other possible ways to extend the results of this paper may include Q-Iearning
when learning on aggregated states (Singh et al., 1995), Q-Iearning for alternating/simultaneous Markov games (Littman, 1994; Szepesvari and Littman, 1996)
and any other algorithms whose corresponding difference process At satisfies an
inequality similar to (7).
Yet another application of the convergence-rate estimate might be the convergence
proof of some average reward reinforcement learning algorithms. The idea of those
algorithms follows from a kind of Tauberian theorem, Le. that discounted sums
converge to the average value if the discount rate converges to one (see e.g. Lemma 1
of (Mahadevan, 1994; Mahadevan, 1996) or for a value-iteration scheme relying on
this idea (Hordjik and Tijms, 1975)). Using the methods developed here the proof
of convergence of the corresponding Q-learning algorithms seems quite possible. We
would like to note here that related results were obtained by Bertsekas et al. et. al
(see e.g. (Bertsekas and Tsitsiklis, 1996)).
Finally, note that as an application of this result we immediately get that the convergence rate of the model-based RL algorithm, where the transition probabilities
and rewards are estimated by their respective averages, is clearly better than that of
for Q-Iearning. Indeed, simple calculations show that the law of iterated logarithm
holds for the learning process underlying model-based RL. Moreover, the exact expression for the convergence rate depends explicitly on how much computational
effort we spend on obtaining the next estimate of the optimal value function, the
more effort we spend the faster is the convergence. This .bound thus provides a direct
way to control the tradeoff between the computational effort and the convergence
rate.
Acknowledgements
This research was supported by aTKA Grant No. F20132 and by a grant provided
by the Hungarian Educational Ministry under contract no. FKFP 1354/1997. I
would like to thank Andras Kramli and Michael L. Littman for numerous helpful
and thought-provoking discussions.
References
Bertsekas, D. and Tsitsiklis, J. (1996).
Scientific, Belmont, MA.
Neuro-Dynamic Programming.
Athena
Hordjik, A. ~nd Tijms, H. (1975). A modified form of the iterative method of
dynamic programming. Annals of Statistics, 3:203-208.
Jaakkola, T., Jordan, M., and Singh, S. (1994). On the convergence of stochastic
iterative dynamic programming algorithms. Neural Computation, 6(6):11851201.
Littman, M. (1994). Markov games as a framework for multi-agent reinforcement
learning. In Proc. of the Eleventh International Conference on Machine Learning, pages 157-163, San Francisco, CA. Morgan Kauffman.
Littman, M. and Szepesvciri, C. (1996). A Generalized Reinforcement Learning
Model: Convergence and applications. In Int. Con/. on Machine Learning.
http://iserv.ikLkfki.hu/ asl-publs.html.
1070
C. Szepesvari
Mahadevan, S. (1994). To discount or not to discount in reinforcement learning: A
case study comparing R learning and Q learning. In Proceedings of the Eleventh
International Conference on Machine Learning, pages 164-172, San Francisco,
CA. Morgan Kaufmann.
Mahadevan, S. (1996). Average reward reinforcement learning: Foundations, algorithms, and empirical results. Machine Learning, 22(1,2,3):124-158.
Major, P. (1973). A law of the iterated logarithm for the Robbins-Monro method.
Studia Scientiarum Mathematicarum Hungarica, 8:95-102.
Poljak, B. and Tsypkin, Y. (1973). Pseudogradient adaption and training algorithms. Automation and Remote Control, 12:83-94.
Puterman, M. L. (1994). Markov Decision Processes - Discrete Stochastic Dynamic
Programming. John Wiley & Sons, Inc., New York, NY.
Ross, S. (1970). Applied Probability Models with Optimization Applications. Holden
Day, San Francisco, California.
Singh, S., Jaakkola, T., and Jordan, M. (1995). Reinforcement learning with soft
state aggregation. In Proceedings of Neural Information Processing Systems.
Singh, S., Jaakkola, T., Littman, M., and Csaba Szepesva ri (1997). On the convergence of single-step on-policy reinforcement-learning al gorithms. Machine
Learning. in preparation.
Szepesvari, C. and Littman, M. (1996). Generalized Markov Decision Processes: Dynamic programming and reinforcement learning algorithms. Machine Learning.
in preparation, available as TR CS96-10, Brown Univ.
Tsitsiklis, J. (1994). Asynchronous stochastic approximation and q-learning. Machine Learning, 8(3-4):257-277.
Wasan, T. (1969). Stochastic Approximation. Cambridge University Press, London.
Watkins, C. (1990). Learning /rom Delayed Rewards. PhD thesis, King's College,
Cambridge. QLEARNING.
| 1383 |@word mild:1 version:1 seems:1 stronger:1 norm:1 nd:1 hu:2 simulation:1 tr:2 boundedness:1 comparing:1 yet:1 must:2 john:1 belmont:1 update:1 stationary:6 intelligence:1 device:1 ith:1 lr:2 provides:1 math:1 firstly:1 simpler:1 along:1 direct:1 become:2 eleventh:2 indeed:1 multi:1 ol:1 discounted:3 decreasing:2 relying:1 zhi:1 actual:1 becomes:1 provided:4 underlying:7 notation:1 maximizes:1 moreover:1 kind:1 developed:2 csaba:1 every:1 act:1 iearning:8 ro:6 control:2 grant:2 bertsekas:3 before:1 consequence:1 might:1 plus:1 iqt:3 empirical:1 thought:1 peat:2 get:4 operator:1 applying:1 demonstrated:1 yt:4 missing:1 educational:1 iri:1 immediately:1 proving:1 updated:2 annals:1 heavily:1 exact:1 programming:5 ft:2 cy:1 remote:1 environment:1 reward:6 littman:14 dynamic:5 singh:6 tit:2 derivation:1 univ:1 london:1 artificial:2 y0t:1 choosing:1 whose:4 quite:1 spend:2 otherwise:2 statistic:1 associative:1 obviously:1 sequence:3 interaction:1 hungary:2 budapest:1 convergence:28 executing:1 converges:3 tk:11 oo:1 qt:8 received:1 tijms:2 hungarian:1 indicate:1 stochastic:7 enable:1 ofq:1 tjt:1 mathematicarum:1 extension:1 hold:6 major:2 smallest:1 proc:1 visited:2 ross:2 robbins:1 clearly:4 aim:1 modified:1 rather:2 ck:1 jaakkola:5 greatly:1 sense:1 helpful:1 holden:1 relation:1 html:1 denoted:1 special:1 integration:1 once:2 sampling:5 identical:1 nearly:1 simplify:1 delayed:1 replaced:1 tq:1 integral:1 respective:1 poljak:2 logarithm:3 soft:1 markovian:5 introducing:1 deviation:2 uniform:1 too:1 optimally:1 perturbed:2 st:6 international:2 iixt:2 contract:1 michael:1 together:2 reestimated:1 thesis:1 satisfied:1 admit:2 pmin:10 li:1 exclude:1 lookup:1 automation:1 int:1 inc:2 explicitly:2 depends:2 piece:1 later:2 analyze:1 aggregation:1 complicated:1 monro:1 kaufmann:1 yield:2 ofthe:1 iterated:3 history:2 simultaneous:1 definition:2 frequency:2 proof:9 associated:1 con:1 sampled:1 proved:1 popular:1 studia:1 day:1 just:1 szeged:2 ilxtll:2 scientiarum:1 sic:1 scientific:1 mdp:2 usage:1 oil:2 brown:1 asl:1 alternating:1 puterman:2 during:2 game:2 szepesva:1 generalized:2 attila:1 outline:1 tt:3 harmonic:2 common:3 rl:3 extend:3 discussed:2 cambridge:2 similarly:1 longer:1 etc:1 certain:1 inequality:3 exploited:2 morgan:2 ministry:1 minimum:2 fortunately:1 relaxed:1 employed:1 converge:2 aggregated:2 signal:2 ii:6 technical:1 faster:3 calculation:1 concerning:1 neuro:1 iteration:1 limt:1 szepesv:2 separately:1 interval:1 ot:2 ivi:1 strict:2 jordan:2 backwards:1 mahadevan:4 enough:1 forthcoming:1 identified:2 simplifies:1 idea:3 pseudogradient:1 tradeoff:1 expression:2 effort:3 york:1 action:15 iterating:1 discount:4 reduced:1 http:1 lixt:1 notice:2 estimated:1 discrete:1 group:1 putting:1 asymptotically:3 relaxation:1 sum:1 decision:3 bound:9 hi:1 guaranteed:1 convergent:1 ri:1 generates:1 argument:2 min:2 xtk:2 tauberian:1 slightly:1 son:1 wi:1 equation:3 eventually:1 mechanism:1 know:1 tsypkin:2 available:1 ocr:1 slower:3 include:1 jlog:1 restrictive:1 rt:4 thank:1 athena:1 xto:1 reason:1 rom:1 mini:2 ratio:1 pmax:8 proper:1 policy:3 zt:3 upper:4 xxa:1 markov:4 finite:1 immediate:1 defining:2 extended:1 perturbation:1 arbitrary:1 introduced:1 pair:7 california:1 address:1 kauffman:1 provoking:1 max:1 suitable:1 treated:1 scheme:1 mdps:2 numerous:1 hungarica:1 literature:1 acknowledgement:1 asymptotic:7 occupation:2 law:4 expect:1 var:1 foundation:1 agent:3 sufficient:1 article:1 pi:4 elsewhere:1 course:1 supported:1 last:1 asynchronous:2 tsitsiklis:5 absolute:1 stand:2 transition:2 made:1 reinforcement:11 san:3 sj:1 qlearning:1 monotonicity:1 assumed:4 francisco:3 iterative:2 sk:5 table:1 szepesvari:10 ca:2 obtaining:1 investigated:1 yex:1 main:3 tl:1 ny:1 wiley:1 gorithms:1 sub:1 andras:1 explicit:1 watkins:3 theorem:6 xt:24 jt:2 showing:2 maxi:2 x:2 essential:2 phd:1 simply:1 appearance:1 infinitely:1 jozsef:1 satisfies:3 adaption:1 ma:1 presentation:1 king:1 leaming:2 change:1 lemma:4 called:1 vrt:1 e:1 college:1 support:1 preparation:2 |
421 | 1,384 | Reinforcement Learning with
Hierarchies of Machines *
Ronald Parr and Stuart Russell
Computer Science Division, UC Berkeley, CA 94720
{parr,russell}@cs.berkeley.edu
Abstract
We present a new approach to reinforcement learning in which the policies considered by the learning process are constrained by hierarchies of
partially specified machines. This allows for the use of prior knowledge
to reduce the search space and provides a framework in which knowledge
can be transferred across problems and in which component solutions
can be recombined to solve larger and more complicated problems. Our
approach can be seen as providing a link between reinforcement learning and "behavior-based" or "teleo-reactive" approaches to control. We
present provably convergent algorithms for problem-solving and learning with hierarchical machines and demonstrate their effectiveness on a
problem with several thousand states.
1 Introduction
Optimal decision making in virtually all spheres of human activity is rendered intractable
by the complexity of the task environment. Generally speaking, the only way around intractability has been to provide a hierarchical organization for complex activities. Although
it can yield suboptimal policies, top-down hierarchical control often reduces the complexity
of decision making from exponential to linear in the size of the problem. For example, hierarchical task network (HTN) planners can generate solutions containing tens of thousands
of steps [5], whereas "fiat" planners can manage only tens of steps.
HTN planners are successful because they use a plan library that describes the decomposition
of high-level activities into lower-level activities. This paper describes an approach to
learning and decision making in uncertain environments (Markov decision processes) that
uses a roughly analogous form of prior knowledge. We use hierarchical abstract machines
(HAMs), which impose constraints on the policies considered by our learning algorithms.
HAMs consist of nondeterministic finite state machines whose transitions may invoke
lower-level machines. Nondeterminism is represented by choice states where the optimal
action is yet to be decided or learned. The language allows a variety of prior constraints
to be expressed, ranging from no constraint all the way to a fully specified solution. One
*This research was supported in part by ARO under the MURI program "Integrated Approach to
Intelligent Systems," grant number DAAH04-96-1-0341.
R. Parr and S. Russell
1044
0_1.
(a)
(b)
(c)
Figure 1: (a) An MOP with ~ 3600 states. The initial state is in the top left. (b) Closeup showing a typical obstacle. (c) Nondetenninistic finite-state controller for negotiating
obstacles.
useful intennediate point is the specification of just the general organization of behavior
into a layered hierarchy, leaving it up to the learning algorithm to discover exactly which
lower-level activities should be invoked by higher levels at each point.
The paper begins with a brief review of Markov decision processes (MOPs) and a description of hierarchical abstract machines. We then present, in abbreviated fonn, the following
results: 1) Given any HAM and any MOP, there exists a new MOP such that the optimal
policy in the new MOP is optimal in the original MOP among those policies that satisfy the
constraints specified by the HAM. This means that even with complex machine specifications we can still apply standard decision-making and learning methods. 2) An algorithm
exists that detennines this optimal policy, given an MOP and a HAM. 3) On an illustrative
problem with 3600 states, this algorithm yields dramatic perfonnance improvements over
standard algorithms applied to the original MOP. 4) A reinforcement learning algorithm
exists that converges to the optimal policy, subject to the HAM constraints, with no need
to construct explicitly a new MOP. 5) On the sample problem, this algorithm learns dramatically faster than standard RL algorithms. We conclude with a discussion of related
approaches and ongoing work.
2 Markov Decision Processes
We assume the reader is familiar with the basic concepts of MOPs. To review, an MOP is
a 4-tuple, (5, A, T, R) where 5 is a set of states, A is a set of actions, T is a transition
model mapping 5 x A x 5 into probabilities in [0, I J, and R is a reward function mapping
5 x A x 5 into real-valued rewards. Algorithms for solving MOPs can return a policy 7r that
maps from 5 to A, a real-valued value function V on states, or a real-valued Q-function on
state-action pairs. In this paper, we focus on infinite-horizon MOPs with a discount factor
/3. The aim is to find an optimal policy 7r* (or, equivalently, V* or Q*) that maximizes the
expected discounted total reward of the agent.
Throughout the paper, we will use as an example the MOP shown in Figure l(a). Here A
contains four primitive actions (up, down, left, right). The transition model, T, specifies that
each action succeeds 80% of time, while 20% of the time the agent moves in an unintended
perpendicular direction. The agent begins in a start state in the upper left corner. A reward
of 5.0 is given for reaching the goal state and the discount factor /3 is 0.999.
3 Hierarchical abstract machines
A HAM is a program which, when executed by an agent in an environment, constrains the
actions that the agent can take in each state. For example, a very simple machine might
dictate, "repeatedly choose right or down," which would eliminate from consideration all
policies that go up or left. HAMs extend this simple idea of constraining policies by
providing a hierarchical means of expressing constraints at varying levels of detail and
Reinforcement Learning with Hierarchies of Machines
1045
specificity. Machines for HAMs are defined by a set of states, a transition function, and a
start function that detennines the initial state of the machine. Machine states are of four
types: Action states execute an action in the environment. Call states execute another
machine as a subroutine. Choice states nondetenninistically select a next machine state.
Stop states halt execution of the machine and return control to the previous call state.
The transition function detennines the next machine state after an action Or call state
as a stochastic function of the current machine state and some features of the resulting
environment state. Machines will typically use a partial description of the environment to
detennine the next state. Although machines can function in partially observable domains,
for the purposes of this paper we make the standard assumption that the agent has access to
a complete description as well.
A HAM is defined by an initial machine in which execution begins and the closure of all
machines reachable from the initial machine. Figure I(c) shows a simplified version of
one element of the HAM we used for the MDP in Figure I. This element is used for
traversing a hallway while negotiating obstacles of the kind shown in Figure 1(b). It runs
until the end of the hallway or an intersection is reached. When it encounters an obstacle, a
choice point is created to choose between two possible next machine states. One calls the
backoff machine to back away from the obstacle and then move forward until the next one.
The other calls the follow-wall machine to try to get around the obstacle. The follow-wall
machine is very simple and will be tricked by obstacles that are concave in the direction of
intended movement; the backoff machine, on the other hand, can move around any obstacle
in this world but could waste time backing away from some obstacles unnecessarily and
should be used sparingly.
Our complete "navigation HAM" involves a three-level hierarchy, somewhat reminiscent
of a Brooks-style architecture but with hard-wired decisions replaced by choice states. The
top level of the hierarchy is basically just a choice state for choosing a hallway navigation
direction from the four coordinate directions. This machine has control initially and regains
control at intersections or corners. The second level of the hierarchy contains four machines
for moving along hallways, one for each direction. Each machine at this level has a choice
state with four basic strategies for handling obstacles. Two back away from obstacles
and two attempt to follow walls to get around obstacles. The third level of the hierarchy
implements these strategies using the primitive actions.
The transition function for this HAM assumes that an agent executing the HAM has access
to a short-range, low-directed sonar that detects obstacles in any of the four axis-parallel
adjacent squares and a long-range, high-directed sonar that detects larger objects such as
the intersections and the ends of hallways. The HAM uses these partial state descriptions
to identify feasible choices. For example, the machine to traverse a hallway northwards
would not be called from the start state because the high-directed sonar would detect a wall
to the north.
Our navigation HAM represents an abstract plan to move about the environment by repeatedly selecting a direction and pursuing this direction until an intersection is reached.
Each machine for navigating in the chosen direction represents an abstract plan for moving
in a particular direction while avoiding obstacles. The next section defines how a HAM
interacts with a specific MDP and how to find an optimal policy that respects the HAM
constraints.
4 Defining and solving the HAM-induced MDP
A policy for a model, M, that is HAM-consistent with HAM H is a scheme for making
choices whenever an agent executing H in M, enters a choice state. To find the optimal
HAM-consistent policy we apply H to M to yield an induced MDP, HoM. A somewhat
simplified description of the construction of HoM is as follows: 1) The set of states in
HoM is the cross-product of the states of H with the states of M. 2) For each state in
HoM where the machine component is an action state, the model and machine transition
R. Parr and S. Russell
1046
WllItoIIlHAM - -
WillloutHAM WittlHAM .......
WtthHAM .... --
r-20
~
1
I~ :
10
o ____~________________~~.
o
SOO
1(0)
lSOO
2:C)X)
l500
lOOO
3SOO
.4QOO
..soD
~
RuntiJne(.ca::ondI'
(a)
(b)
Figure 2: Experimental results showing policy value (at the initial state) as a function of
runtime on the domain shown in Figure 1. (a) Policy iteration with and without the HAM.
(b) Q-learning with and without the HAM (averaged over 10 runs).
functions are combined. 3) For each state where the machine component is a choice state,
actions that change only the machine component of the state are introduced. 4) The reward
is taken from M for primitive actions, otherwise it is zero. With this construction, we have
the following (proof omitted):
Lemma 1 For any Markov decision process M and any! HAM H, the induced process
HoM is a Markov decision process.
Lemma 2 If 7r is an optimal policy for HoM , then the primitive actions specified by
constitute the optimal policy for M that is HAM-consistent with H.
7r
Of course, HoM may be quite large. Fortunately, there are two things that will make
the problem much easier in most cases. The first is that not all pairs of HAM states and
environment states will be possible, i.e., reachable from an initial state. The second is that
the actual complexity of the induced MOP is determined by the number of choice points,
i.e., states of HoM in which the HAM component is a choice state. This leads to the
following:
Theorem 1 For any MOP, M, and HAM, H, let C be the set of choice points in HoM .
There exists a decision process, reduce(H 0 M), with states C such that the optimal policy
for reduce(H 0 M) corresponds to the optimal policy for M that is HAM-consistent with
H.
Proof sketch We begin by applying Lemma 1 and then observing that in states of HoM
where the HAM component is not a choice state, only one action is permitted. These
states can be removed to produce an equivalent Semi-Markov decision process (SMOP).
(SMOPs are a generalization of Markov decision processes that permit different discount
rates for different transitions.) The optimal policy for this SMOP will be the same as the
optimal policy for HoM and by Lemma 2, this will be the optimal policy for M that is
HAM-consistent with H. 0
This theorem formally establishes the mechanism by which the constraints embodied in a
HAM can be used to simplify an MDP. As an example of the power of this theorem, and
to demonstrate that this transformation can be done efficiently, we applied our navigation
HAM to the problem described in the previous section. Figure 2(a) shows the results of
applying policy iteration to the original model and to the transformed model. Even when
we add in the cost of transformation (which, with our rather underoptimized code, takes
ITo preserve the Markov property, we require that if a machine has more than one possible caller
in the hierarchy, that each appearance is treated as a distinct machine. This is equivalent to requiring
that the call graph for the HAM is a tree. It follows from this that circular calling sequences are also
forbidden.
Reinforcement Learning with Hierarchies of Machines
1047
866 seconds), the HAM method produces a good policy in less than a quarter of the time
required to find the optimal policy in the original model. The actual solution time is 185
seconds versus 4544 seconds.
An important property of the HAM approach is that model transformation produces an
MDP that is an accurate model of the application of the HAM to the original MDP. Unlike
typical approximation methods for MDPs, the HAM method can give strict performance
guarantees. The solution to the transformed model Teduce(H 0 M) is the optimal solution
from within a well-defined class of policies and the value assigned to this solution is the
true expected value of applying the concrete HAM policy to the original MDP.
5
Reinforcement learning with HAMs
HAMs can be of even greater advantage in a reinforcement learning context, where the
effort required to obtain a solution typically scales very badly with the size of the problem.
HAM contraints can focus exploration of the state space, reducing the "blind search" phase
that reinforcement learning agents must endure while learning about a new environment.
Learning will also be fasterfor the same reason policy iteration is faster in the HAM-induced
model; the agent is effectively operating in a reduced state space.
We now introduce a variation of Q-learning called HAMQ-1earning that learns directly
in the reduced state space without performing the model transformation described in the
previous section. This is significant because the the environment model is not usually
known a priori in reinforcement learning contexts.
A HAMQ-learning agent keeps track of the following quantities: t, the current environment
state; n, the current machine state; Se and me, the environment state and machine state at
the previous choice point; a, the choice made at the previous choice point; and Te and 13e,
the total accumulated reward and discount since the previous choice point. It also maintains
an extended Q-table, Q([s, m], a), which is indexed by an environment-state/machine-state
pair and by an action taken at a choice point.
For every environment transition from state s to state t with observed reward T and discount
~ Te + 13eT and 13e ~ 13l3e. For each transition
to a choice point, the agent does
13, the HAMQ-Iearning agent updates: Te
Q([se, me], a)
and then Te
~
0, 13e
~
~
Q([se, mc], a)
+ a[Te + 13e V([t, n]) - Q([Se, mc], a)],
1.
Theorem 2 For any finite-state MDP, M, and any HAM, H, HAMQ-Iearning will converge
to the optimal choice for every choice point in Teduce(H 0 M) with probability l.
Proof sketch We note that the expected reinforcement signal in HAMQ-Iearning is the
same as the expected reinforcement signal that would be received if the agent were acting
directly in the transformed model of Theorem 1 above. Thus, Theorem 1 of [11] can be
applied to prove the convergence of the HAMQ-learning agent, provided that we enforce
suitable constraints on the exploration strategy and the update parameter decay rate. 0
We ran some experiments to measure the performance of HAMQ-learning on our sample
problem. Exploration was achieved by selecting actions according to the Boltzman distribution with a temperature parameter for each state. We also used an inverse decay for the
update parameter a. Figure 2(b) compares the learning curves for Q-Iearning and HAMQlearning. HAMQ-Iearning appears to learn much faster: Q-Iearning required 9,000,000
iterations to reach the level achieved by HAMQ-learning after 270,000 iterations. Even
after 20,000,000 iterations, Q-Iearning did not do as well as HAMQ-learning.2
2Speedup techniques such as eligibility traces could be applied to get better Q-Ieaming results;
such methods apply equally well to HAMQ-Iearning.
1048
6
R. Parr and S. Russell
Related work
State aggregation (see, e.g., [18] and [7]) clusters "similar" states together and assigns them
the same value, effectively reducing the state space. This is orthogonal to our approach
and could be combined with HAMs. However, aggregation should be used with caution
as it treats distinct states as a single state and can violate the Markov property leading to
the loss of performance guarantees and oscillation or divergence in reinforcement learning.
Moreover, state aggregation may be hard to apply effectively in many cases.
Dean and Lin [8] and Bertsekas and Tsitsiklis [2], showed that some MDPs are loosely
coupled and hence amenable to divide-and-conquer algorithms. A machine-like language
was used in [13] to partition an MDP into decoupled subproblems. In problems that are
amenable to decoupling, this could approaches could be used in combinated with HAMs.
Dayan and Hinton [6] have proposedJeudal RL which specifies an explicit subgoal structure,
with fixed values for each sub goal achieved, in order to achieve a hierarchical decomposition
of the state space. Dietterich extends and generalizes this approach in [9]. Singh has
investigated a number of approaches to subgoal based decomposition in reinforcement
learning (e.g. [17] and [16]). Subgoals seem natural in some domains, but they may
require a significant amount of outside knowledge about the domain and establishing the
relationship between the value of subgoals with respect to the overall problem can be
difficult.
Bradtke and Duff [3] proposed an RL algorithm for SMDPs. Sutton [19] proposes temporal
abstractions, which concatenate sequences of state transitions together to permit reasoning
about temporally extended events, and which can thereby form a behavioral hierarchy as
in [14] and [15]. Lin's somewhat informal scheme [12] also allows agents to treat entire
policies as single actions. These approaches can be emcompassed within our framework
by encoding the events or behaviors as machines.
The design of hierarchically organized, "layered" controllers was popularized by Brooks [4].
His designs use a somewhat different means of passing control, but our analysis and theorems
apply equally well to his machine description language. The "teleo-reactive" agent designs
of Benson and Nilsson [I] are even closer to our HAM language. Both of these approaches
assume that the agent is completely specified, albeit self-modifiable. The idea of partial
behavior descriptions can be traced at least to Hsu's partial programs [10], which were
used with a deterministic logical planner.
7
Conclusions and future work
We have presented HAMs as a principled means of constraining the set of policies that are
considered for a Markov decision process and we have demonstrated the efficacy of this
approach in a simple example for both policy iteration and reinforcement learning. Our
results show very significant speedup for decision-making and learning-but of course, this
reflects the provision of knowledge in the form of the HAM. The HAM language provides
a very general method of transferring knowledge to an agent and we only have scratched
the surface of what can be done with this approach.
We believe that if desired, subgoal information can be incorporated into the HAM structure,
unifying subgoal-based approaches with the HAM approach. Moreover, the HAM structure
provides a natural decomposition of the HAM-induced model, making it amenable to the
divide-and-conquer approaches of [8] and [2].
There are opportunities for generalization across all levels of the HAM paradigm. Value
function approximation can be used for the HAM induced model and inductive learning
methods can be used to produce HAMs or to generalize their effects upon different regions
of the state space. Gradient-following methods also can be used to adjust the transition
probabilities of a stochastic HAM.
HAMs also lend themselves naturally to partially observable domains. They can be applied
directly when the choice points induced by the HAM are states where no confusion about
Reinforcement Learning with Hierarchies of Machines
1049
the true state of the environment is possible. The application of HAMs to more general
partially observable domains is more complicated and is a topic of ongoing research. We
also believe that the HAM approach can be extended to cover the average-reward optimality
criterion.
We expect that successful pursuit of these lines of research will provide a formal basis for
understanding and unifying several seemingly disparate approaches to control, including
behavior-based methods. It should also enable the use of the MDP framework in real-world
applications of much greater complexity than hitherto attacked, much as HTN planning has
extended the reach of classical planning methods.
References
[1] S. Benson and N. Nilsson. Reacting, planning and learning in an autonomous agent. In
K. Furukawa, D. Michie, and S. Muggleton, editors, Machine Intelligence 14. Oxford University
Press, Oxford, 1995.
[2] D. C. Bertsekas and J. N. Tsitsiklis. Parallel and Distributed Computation: Numerical Metlwds.
Prentice-Hall, Englewood Cliffs, New Jersey, 1989.
[3] S. J. Bradtke and M. O. Duff. Reinforcement learning methods for continuous-time Markov
decision problems. In Advances in Neurallnfonnation Processing Systems 7: Proc. of the 1994
Conference, Denver, Colorado, December 1995. MIT Press.
[4] R. A. Brooks. A robust layered control system for a mobile robot. IEEE Journal of Robotics
and Automation, 2, 1986.
[5] K. W. Currie and A. Tate. O-Plan: the Open Planning Architecture. Artificial Intelligence,
52(1), November 1991.
[6] P. Dayan and G. E. Hinton. Feudal reinforcement learning. In Stephen Jose Hanson, Jack D.
Cowan, and C. Lee Giles, editors, Neural Information Processing Systems 5, San Mateo,
California, 1993. Morgan Kaufman.
[7] T. Dean, R. Givan, and S. Leach. Model reduction techniques for computing approximately
optimal solutions for markov decision processes. In Proc. of the Thirteenth Conference on Uncertainty in Artificial Intelligence , Providence, Rhode Island, August 1997. Morgan Kaufmann.
[8] T. Dean and S.-H. Lin. Decomposition techniques for planning in stochastic domains. In Proc.
of the Fourteenth Int. Joint Conference on Artificial Intelligence, Montreal, Canada, August
1995. Morgan Kaufmann.
[9] Thomas G. Dietterich. Hierarchical reinforcement learning with the MAXQ value function
decomposition. Technical report, Department of Computer Science, Oregon State University,
Corvallis, Oregon, 1997.
[10] Y.-J. Hsu. Synthesizing efficient agents from partial programs. In Metlwdologiesfor Intelligent
Systems: 6th Int. Symposium, ISMIS '91, Proc., Charlotte, North Carolina, October 1991.
Springer-Verlag.
[11] T. Jaakkola, M.l. Jordan, and S.P. Singh. On the convergence of stochastic iterative dynamic
programming algorithms. Neural Computation, 6(6), 1994.
[12] L.-J. Lin. Reinforcement Learning for Robots Using Neural Networks. PhD thesis, Computer
Science Department, Carnegie-Mellon University, Pittsburgh, Pennsylvania, 1993.
[13] Shieu-Hong Lin. Exploiting Structure for Planning and Control. PhD thesis, Computer Science
Department, Brown University, Providence, Rhode Island, 1997.
[14] A. McGovern, R. S. Sutton, and A. H. Fagg. Roles of macro-actions in accelerating reinforcement
learning. In 1997 Grace Hopper Celebration of Women in Computing, 1997.
[15] D. Precup and R. S. Sutton. Multi-time models fortemporally abstract planning. In This Volume .
[16] S. P. Singh. Scaling reinforcement learning algorithms by learning variable temporal resolution
models. In Proceedings of the Ninth International Conference on Machine Learning, Aberdeen,
July 1992. Morgan Kaufmann.
[17] S. P. Singh. Transfer of learning by composing solutions of elemental sequential tasks. Machine
Learning, 8(3), May 1992.
[18] S. P. Singh, T. Jaakola, and M. I. Jordan. Reinforcement learning with soft state aggregation. In
G. Tesauro, D. S. Touretzky, and T. K. Leen, editors, Neural Information Processing Systems 7,
Cambridge, Massachusetts, 1995. MIT Press.
[19] R. S. Sutton. Temporal abstraction in reinforcement learning. In Proc. of the Twelfth Int.
Conference on Machine Learning, Tahoe City, CA, July 1995. Morgan Kaufmann.
| 1384 |@word version:1 twelfth:1 open:1 closure:1 carolina:1 decomposition:6 fonn:1 dramatic:1 thereby:1 reduction:1 initial:6 contains:2 efficacy:1 selecting:2 unintended:1 current:3 yet:1 reminiscent:1 must:1 ronald:1 concatenate:1 partition:1 numerical:1 update:3 intelligence:4 hallway:6 short:1 provides:3 traverse:1 tahoe:1 along:1 symposium:1 prove:1 nondeterministic:1 behavioral:1 introduce:1 expected:4 roughly:1 themselves:1 planning:7 behavior:5 multi:1 discounted:1 detects:2 actual:2 begin:4 discover:1 provided:1 moreover:2 intennediate:1 maximizes:1 what:1 hitherto:1 kind:1 kaufman:1 caution:1 transformation:4 guarantee:2 temporal:3 berkeley:2 every:2 concave:1 iearning:8 runtime:1 exactly:1 control:9 grant:1 bertsekas:2 treat:2 sutton:4 encoding:1 oxford:2 cliff:1 establishing:1 reacting:1 approximately:1 rhode:2 might:1 mateo:1 perpendicular:1 range:2 averaged:1 jaakola:1 decided:1 directed:3 implement:1 dictate:1 specificity:1 get:3 layered:3 closeup:1 context:2 applying:3 prentice:1 equivalent:2 map:1 dean:3 deterministic:1 demonstrated:1 primitive:4 go:1 resolution:1 assigns:1 his:2 coordinate:1 caller:1 analogous:1 variation:1 autonomous:1 hierarchy:12 construction:2 colorado:1 programming:1 us:2 element:2 michie:1 muri:1 observed:1 role:1 enters:1 thousand:2 region:1 russell:5 movement:1 removed:1 ran:1 principled:1 environment:15 ham:63 complexity:4 constrains:1 reward:8 dynamic:1 singh:5 solving:3 upon:1 division:1 completely:1 basis:1 joint:1 represented:1 jersey:1 distinct:2 artificial:3 mcgovern:1 choosing:1 outside:1 whose:1 quite:1 larger:2 solve:1 valued:3 otherwise:1 seemingly:1 sequence:2 advantage:1 aro:1 product:1 macro:1 achieve:1 description:7 exploiting:1 convergence:2 cluster:1 elemental:1 wired:1 produce:4 converges:1 executing:2 object:1 montreal:1 received:1 c:1 involves:1 direction:9 detennines:3 stochastic:4 exploration:3 human:1 enable:1 require:2 generalization:2 wall:4 givan:1 recombined:1 around:4 considered:3 hall:1 mapping:2 mop:16 parr:5 omitted:1 purpose:1 proc:5 establishes:1 city:1 reflects:1 mit:2 aim:1 reaching:1 rather:1 qoo:1 varying:1 mobile:1 jaakkola:1 focus:2 daah04:1 improvement:1 detect:1 tate:1 dayan:2 abstraction:2 accumulated:1 integrated:1 transferring:1 eliminate:1 typically:2 initially:1 entire:1 transformed:3 subroutine:1 provably:1 backing:1 endure:1 among:1 overall:1 l3e:1 priori:1 proposes:1 plan:4 constrained:1 uc:1 construct:1 represents:2 stuart:1 unnecessarily:1 future:1 report:1 intelligent:2 simplify:1 preserve:1 divergence:1 familiar:1 replaced:1 intended:1 teleo:2 phase:1 attempt:1 organization:2 englewood:1 circular:1 adjust:1 navigation:4 amenable:3 accurate:1 ieaming:1 tuple:1 closer:1 partial:5 perfonnance:1 traversing:1 tree:1 indexed:1 orthogonal:1 loosely:1 divide:2 decoupled:1 desired:1 uncertain:1 negotiating:2 soft:1 obstacle:14 giles:1 cover:1 cost:1 sod:1 successful:2 providence:2 sparingly:1 combined:2 international:1 lee:1 invoke:1 together:2 precup:1 concrete:1 thesis:2 manage:1 containing:1 choose:2 woman:1 charlotte:1 corner:2 style:1 return:2 leading:1 waste:1 north:2 automation:1 int:3 oregon:2 satisfy:1 explicitly:1 scratched:1 blind:1 try:1 observing:1 reached:2 start:3 aggregation:4 maintains:1 complicated:2 parallel:2 square:1 kaufmann:4 efficiently:1 yield:3 identify:1 generalize:1 basically:1 tricked:1 mc:2 reach:2 touretzky:1 whenever:1 celebration:1 naturally:1 proof:3 stop:1 hsu:2 massachusetts:1 logical:1 knowledge:6 fiat:1 hamq:11 organized:1 provision:1 back:2 appears:1 higher:1 follow:3 permitted:1 subgoal:4 execute:2 done:2 leen:1 just:2 until:3 hand:1 sketch:2 defines:1 mdp:11 believe:2 dietterich:2 effect:1 concept:1 requiring:1 true:2 brown:1 inductive:1 hence:1 assigned:1 adjacent:1 self:1 eligibility:1 illustrative:1 criterion:1 hong:1 complete:2 demonstrate:2 confusion:1 bradtke:2 temperature:1 reasoning:1 ranging:1 invoked:1 consideration:1 jack:1 hopper:1 quarter:1 rl:3 denver:1 volume:1 subgoals:2 extend:1 expressing:1 significant:3 corvallis:1 mellon:1 cambridge:1 language:5 reachable:2 moving:2 specification:2 access:2 robot:2 operating:1 surface:1 add:1 neurallnfonnation:1 forbidden:1 showed:1 tesauro:1 verlag:1 leach:1 furukawa:1 seen:1 morgan:5 fortunately:1 somewhat:4 impose:1 greater:2 converge:1 paradigm:1 signal:2 semi:1 stephen:1 violate:1 july:2 reduces:1 technical:1 faster:3 cross:1 sphere:1 long:1 lin:5 muggleton:1 equally:2 halt:1 basic:2 controller:2 iteration:7 achieved:3 robotics:1 whereas:1 thirteenth:1 htn:3 leaving:1 unlike:1 strict:1 subject:1 induced:8 virtually:1 thing:1 december:1 cowan:1 effectiveness:1 seem:1 call:6 jordan:2 constraining:2 variety:1 architecture:2 pennsylvania:1 suboptimal:1 reduce:3 idea:2 accelerating:1 effort:1 speaking:1 passing:1 constitute:1 action:19 repeatedly:2 dramatically:1 generally:1 useful:1 backoff:2 se:4 amount:1 discount:5 ten:2 reduced:2 generate:1 specifies:2 track:1 modifiable:1 carnegie:1 four:6 traced:1 graph:1 regains:1 fagg:1 run:2 inverse:1 jose:1 uncertainty:1 fourteenth:1 extends:1 planner:4 reader:1 throughout:1 pursuing:1 earning:1 oscillation:1 decision:17 scaling:1 hom:11 convergent:1 activity:5 badly:1 constraint:9 feudal:1 calling:1 optimality:1 performing:1 rendered:1 transferred:1 speedup:2 department:3 according:1 popularized:1 across:2 describes:2 island:2 making:7 nilsson:2 benson:2 taken:2 abbreviated:1 mechanism:1 end:2 informal:1 generalizes:1 detennine:1 pursuit:1 permit:2 apply:5 hierarchical:10 away:3 enforce:1 encounter:1 original:6 thomas:1 top:3 assumes:1 opportunity:1 unifying:2 conquer:2 classical:1 move:4 quantity:1 strategy:3 interacts:1 grace:1 navigating:1 gradient:1 link:1 me:2 topic:1 reason:1 code:1 relationship:1 providing:2 equivalently:1 difficult:1 executed:1 october:1 subproblems:1 trace:1 disparate:1 synthesizing:1 design:3 contraints:1 policy:32 upper:1 markov:12 finite:3 november:1 attacked:1 defining:1 extended:4 hinton:2 incorporated:1 duff:2 ninth:1 august:2 canada:1 introduced:1 pair:3 required:3 specified:5 hanson:1 california:1 learned:1 maxq:1 brook:3 usually:1 program:4 including:1 soo:2 lend:1 power:1 suitable:1 event:2 treated:1 natural:2 scheme:2 brief:1 library:1 mdps:2 temporally:1 axis:1 created:1 coupled:1 embodied:1 prior:3 review:2 understanding:1 fully:1 loss:1 expect:1 versus:1 agent:21 consistent:5 editor:3 intractability:1 course:2 looo:1 supported:1 tsitsiklis:2 formal:1 nondeterminism:1 distributed:1 curve:1 transition:12 world:2 forward:1 made:1 reinforcement:24 san:1 simplified:2 boltzman:1 observable:3 keep:1 pittsburgh:1 conclude:1 search:2 continuous:1 iterative:1 sonar:3 table:1 learn:1 transfer:1 robust:1 ca:3 decoupling:1 composing:1 investigated:1 complex:2 domain:7 did:1 hierarchically:1 sub:1 explicit:1 exponential:1 third:1 ito:1 learns:2 down:3 theorem:7 specific:1 showing:2 decay:2 intractable:1 consist:1 exists:4 albeit:1 sequential:1 effectively:3 phd:2 execution:2 te:5 horizon:1 easier:1 intersection:4 aberdeen:1 appearance:1 expressed:1 partially:4 springer:1 corresponds:1 goal:2 feasible:1 hard:2 change:1 typical:2 infinite:1 determined:1 reducing:2 acting:1 smdps:1 lemma:4 total:2 called:2 experimental:1 succeeds:1 select:1 formally:1 reactive:2 ongoing:2 avoiding:1 handling:1 |
422 | 1,385 | The Observer-Observation Dilemma
in Neuro-Forecasting
Hans Georg Zimmermann
Ralph Neuneier
SiemensAG
Corporate Technology
D-81730 Munchen, Germany
Georg.Zimmermann@mchp.siemens.de
Siemens AG
Corporate Technology
D-81730 Munchen, Germany
Ralph.Neuneier@mchp.siemens.de
Abstract
We explain how the training data can be separated into clean information and unexplainable noise. Analogous to the data, the neural network
is separated into a time invariant structure used for forecasting, and a
noisy part. We propose a unified theory connecting the optimization algorithms for cleaning and learning together with algorithms that control
the data noise and the parameter noise. The combined algorithm allows
a data-driven local control of the liability of the network parameters and
therefore an improvement in generalization. The approach is proven to
be very useful at the task of forecasting the German bond market.
1 Introduction: The Observer-Observation Dilemma
Human beings believe that they are able to solve a psychological version of the ObserverObservation Dilemma. On the one hand, they use their observations to constitute an understanding of the laws of the world, on the other hand, they use this understanding to evaluate
the correctness of the incoming pieces of information. Of course, as everybody knows,
human beings are not free from making mistakes in this psychological dilemma. We encounter a similar situation when we try to build a mathematical model using data. Learning
relationships from the data is only one part of the model building process. Overrating this
part often leads to the phenomenon of overfitting in many applications (especially in economic forecasting). In practice, evaluation of the data is often done by external knowledge,
i. e. by optimizing the model under constraints of smoothness and regularization [7]. If we
assume, that our model summerizes the best knowledge of the system to be identified, why
should we not use the model itself to evaluate the correctness of the data? One approach to
do this is called Clearning [11]. In this paper, we present a unified approach of the interaction between the data and a neural network (see also [8]). It includes a new symmetric view
on the optimization algorithms, here learning and cleaning, and their control by parameter
and data noise.
The Observer-Observation Dilemma in Neuro-Forecasting
993
2 Learning
2.1
Learning reviewed
We are especially interested in using the output of a neural network y( x, w), given the
input pattern, x, and the weight vector, w, as a forecast of financial time series. In the
context of neural networks learning nonnally means the minimization of an error function
E by changing the weight vector w in order to achieve good generalization performance.
Typical error functions can be written as a sum of individual terms over all T training
patterns, E = ~ 'L,;=1 E t . For example, the maximum-likelihood principle leads to
E t = 1/2
(y(x, w) - yt)2 ,
(1)
yt
with
as the given target pattern. If the error function is a nonlinear function of the parameters, learning has to be done iteratively by a search through the weight space, changing
the weights from step T to T + 1 according to:
(2)
There are several algorithms for choosing the weight increment ~W(T) , the most easiest
being gradient descent. After each presentation of an input pattern, the gradient gt :=
VEt Iw of the error function with respect to the weights is computed. In the batch version
of gradient descent the increments are based on all training patterns
1
~W(T) = -"1g = -"1 T
T
L gt,
(3)
t=l
whereas the pattern-by-pattern version changes the weights after each presentation of a
pattern Xt (often randomly chosen from the training set):
(4)
The learning rate "1 is typically held constant or follows an annealing procedure during
training to assure convergence. Our experiments have shown that small batches are most
useful, especially in combination with Vario-Eta, a stochastic approximation of a QuasiNewton method [3]:
~W(T) = -
N
"1
J+'L,(gt - g)2
.
~L
gt,
(5)
N t=l
with and N ~ 20. Learning pattern-by-pattern or with small batches can be viewed as a
stochastic search process because we can write the weight increments as:
(6)
These increments consist of the terms 9 with a drift to a local minimum and of noise terms
'L,~ 1 gt - g) disturbing this drift.
(k
2.2 Parameter Noise as an Implicit Penalty Function
Consider the Taylor expansion of E (w) around some point w in the weight space
E(w + ~w) = E(w)
1
+ VE ~w + 2~W'H ~w
(7)
H. G. Zimmermann and R. Neuneier
994
with H as the Hessian of the error function. Assume a given sequence of T disturbance
vectors ~ Wt, whose elements ~ Wt ( i) are identically, independently distributed (i .i.d.) with
zero mean and variance (row-)vector var(~wi) to approximate the expectation (E( w) by
1",
(E(w) ~ T L...J E(w
1",
.
+ ~Wt) = E(w) + :2 L...J var(~w(z))Hii'
t
(8)
i
with Hii as the diagonal elements of H. In eq. 8, noise on the weights acts implicitly as a
penalty term to the error function given by the second derivatives H ii . The noise variances
var( ~ w( i)) operate as penalty parameters. As a result of this flat minima solutions which
may be important for achieving good generalization performance are favored [5].
Learning pattern-by-pattem introduces such noise in the training procedure i.e., ~ Wt =
-1] ? gt? Close to convergence, we can assume that gt is i.i.d. with zero mean and variance
vector var(gi) so that the expected value can be approximated by
(E(w) ~ E(w)
82E
Lvar(gd8Wi2'
2 I.
TJ2
+-
(9)
This type of learning introduces to a local penalty parameter var( ~ w (i) ), characterizing
the stability of the weights w = [Wdi=l, ... ,k.
The noise effects due to Vario-Eta learning ~wt(i)
1]2
(E(w)
~ E(w) + 2
= -R .gti leads to
E
:L 88w~'
2
i
(10)
I
By canceling the term var(gi) in eq. 9, Vario-Eta achieves a simplified uniform penalty
parameter, which depends only on the learning rate 1]. Whereas pattern-by-pattern learning
is a slow algorithm with a locally adjusted penalty control, Vario-Eta is fast only at the cost
of a simplified uniform penalty term. We summarize this section by giving some advice on
how to learn to flat minima solutions:
? Train the network to a minimal training error solution with Vario-Eta, which is a
stochastic approximation of a Newton method and therefore very fast.
? Add a final phase of pattem-by-pattern learning with uniform learning rate to fine
tune the local curvature structure by the local penalty parameters (eq. 9).
? Use a learning rate 1] as high as possible to keep the penalty effective. The training
error may vary a bit, but the inclusion of the implicit penalty is more important.
3
3.1
Cleaning
Cleaning reviewed
When training neural networks, one typically assumes that the data is noise-free and one
forces the network to fit the data exactly. Even the control procedures to minimize overfitting effects (i.e., pruning) consider the inputs as exact values. However, this assumption
is often violated, especially in the field of financial analysis, and we are taught by the
phenomenon of overfitting not to follow the data exactly. Clearning, as a combination of
cleaning and learning, has been introduced in the paper of [11]. The motivation was to
minimize overfitting effects by considering the input data as corrupted by noise whose distribution has also to be learned. The Cleaning error function for the pattern t is given by
the sum of two terms
yx
E t'
="21[(Yt - Ytd)2 + (Xt - x td)2] = E ty + E tx
(11)
The Observer-Observatio?n Dilemma in Neuro-Forecasting
995
with xf, yt as the observed data point. In the pattem-by-pattem learning, the network
output y( x t, w) determines the adaptation as usual,
(12)
We have also to memorize correction vectors
present the cleaned input Xt to the network,
Xt
=
&t
for all input data of the training set to
xf + &t
(13)
= 0 can be described as
The update rule for the corrections, initialized with ~x~O)
~X~T+I)
(1 - 1])~X!T)
- 1](Yt -
yt) ~
(14)
All the necessary quantIties, i. e. (Yt - yt) &Y~;w) are computed by typical backpropagation algorithms, anyway. We experienced, that the algorithms work well, if the
same learning rate 1] is used for both, the weight and cleaning updates. For regression,
cleaning forces the acceptance of a small error in x. which can in turn decrease the error in
Y dramatically, especially in the case of outliers. Successful applications of Cleaning are
reported in [11] and [9].
Although the network may learn an optimal model for the cleaned input data, there is
no easy way to work with cleaned data on the test set. As a consequence, the model is
evaluated on a test set with a different noise characteristic compared to the training set. We
will later propose a combination of learning with noise and cleaning to work around this
serious disadvantage.
3.2 Data Noise reviewed
Artificial noise on the input data is often used during training because it creates an infinite
number of training examples and expands the data to empty parts of the input space. As
a result, the tendency of learning by heart may be limited because smoother regression
functions are produced.
Now, we are considering again the Taylor expansion, this time applied to E (x) around
some point x in the input space. The expected value (E (x)) is approximated by
(E(x)}
~ ~L
E(x
+ &t} =
E(x)
+~L
t
var(&(j))Hjj,
(15)
j
with Hjj as the diagonal elements of the Hessian Hxx of the error function with respect to
the inputs x. Again, in eq. 15, noise on the inputs acts implicitly as a penalty term to the
error function with the noise variances var( & (j)) operating as penalty parameters. Noise
on the input improve generalization behavior by favoring smooth models [1].
The noise levels can be set to a constant value, e. g. given by a priori knowledge, or adaptive
as described now. We will concentrate on a uniform or normal noise distribution. Then,
the adaptive noise level ~j is estimated for each input j individually. Suppressing pattern
indices, we define the noise levels ~j or ~J as the average residual errors:
uniform residual error:..t).
Gaussian residual error:
..t)2.
= _1 '~
"
T
___
1
T
t
'~
"
t
IOEY I,
(16)
(OEY) 2
(17)
OXj
OXj
Actual implementations use stochastic approximation, e. g. for the uniform residual error
?~T+1)
J
= (1 _ T
~ )~~T) + ~ IoEY I.
)
T OXj
(18)
H. G. Zimmennann and R. Neuneier
996
The different residual error levels can be interpreted as follows: A small level ~j may
indicate an unimportant input j or a perfect fit of the network concerning this input j. In
both cases, a small noise level is appropriate. On the other hand, a high value of ~j for an
input j indicates an important but imperfectly fitted input. In this case high noise levels are
advisable. High values of ~j lead to a stiffer regression model and may therefore increase
the generalization perfonnance of the network.
3.3 Cleaning with Noise
Typically, training with noisy inputs takes a data point and adds a random variable drawn
from a fixed or adaptive distribution. This new data point Xt is used as an input to the
network. If we assume, that the data is corrupted by outliers and other influences, it is
preferable to add the noise tenn to the cleaned input. For the case of Gaussian noise the
resulting new input is:
Xt =
+ ~Xt + ~?,
(19)
xf
with ? drawn from the nonnal distribution. The cleaning of the data leads to a corrected
mean of the data and therefore to a more symmetric noise distribution, which also covers
the observed data x t ?
We propose a variant which allows more complicated noise distributions:
(20)
with k as a random number drawn from the indices of the correction vectors (~xtlt=l , ... ,T.
In this way we use a possibly asymmetric and/or dependent noise distribution, which still
covers the observed data Xt by definition of the algorithm.
xf
One might wonder, why to disturb the cleaned input
+ ~Xt with an additional noisy
tenn ~x k. The reason for this is, that we want to benefit from representing the whole input
distribution to the network instead of only using one particular realization.
4 A Unifying Approach
4.1
The Separation of Structure and Noise
In the previous sections we explained how the data can be separated into clean infonnation
and unexplainable noise. Analogous, the neural network is described as a time invariant
structure (otherwise no forecasting is possible) and a noisy part.
- t cleaned data
+ time invariant data noise
data
neural network-ttime invariant parameters+parameter noise
We propose to use cleaning and adaptive noise to separate the data and to use learning and
stochastic search to separate the structure of the neural network.
~ cleaning(neural network) + adaptive noise (neural network)
data
neural network~learning (data)
+ stochastic search(data)
The algorithms analyzing the data depend directly o"n the network whereas the methods
searching for structure are directly related to the data. It should be clear that the model
building process should combine both aspects in an alternate or simultaneous manner. The
interaction of algorithms concerning data analysis and network structure enables the realization of the the concept of the Observer-Observation Dilemma.
The Observer-Observation Dilemma in Neuro-Forecasting
997
The aim of the unified approach can be described, exemplary assuming here a Gaussian
noise model, as the minimization of the error due to both, the structure and the data:
T
2~ I: [(Yt - y1)2 + {Xt - x1)2]
-+
~~fJ
(21)
t=l
Combining the algorithms and approximating the cumulative gradient 9 by g, we receive
data
(22)
structure
(1 -
0:
WIT) -
)gH
+ O:(Yt -
7]g(T)
'-....-'
learning
yt)
P-
-7](9t - g0-))
---......-.noise
The cleaning of the data by the network computes an individual correction term for each
training pattern. The adaptive noise procedure according to eq. 20 generates a potentially
asymmetric and dependent noise distribution which also covers the observed data. The
implied curvature penalty, whose strength depends on the individual liability of the input
variables, can improve the generalization performance of the neural network.
The learning of the structure searches for time invariant parameters characterized by
-j; L 9t = O. The parameter noise supports this exploration as a stochastic search to find
better "global" minima. Additionally, the generalization performance may be further improved by the implied curvature penalty depending on the local liability of the parameters.
Note that, although the description of the weight updates collapses to the simple form of
eq. 4, we preferred the formula above to emphasize the analogy between the mechanism
which handles the data and the structure.
In searching for an optimal combination of data and parameters, the noise of both parts is
not a disastrous failure to build a perfect model but it is an important element to control the
interaction of data and structure.
4.2 Pruning
The neural network topology represents only a hypothesis of the true underlying class of
functions. Due to possible misspecification, we may have defects of the parameter noise
distribution. Pruning algorithms are not only a way to limit the memory of the network, but
they also appear useful to correct the noise distribution in different ways.
Stochastic-Pruning [2] is basically a t-test on the weights w. Weights with low testw values
constitute candidates for pruning to cancel weights with low liability measured by the size
of the weight divided by the standard deviation of its fluctuations. By this, we get a stabilization of the learning against resampling of the training data. A further weight pruning
method is EBD, Early-Brain-Damage [10], which is based on the often cited OBD pruning method of [6]. In contrast to OBD, EBD allows its application before the training has
reached a local minimum. One of the advantages of EBD over OBD is the possibility to
perform the testing while being slidely away from a local minimum. In our training procedure we propose to use noise even in the final part of learning and therefore we are only
nearby a local minimum. Furthermore, EBD is also able to revive already pruned weights.
Similar to Stochastic Pruning, EBD favors weights with a low rate of fluctuations. If a
weight is pushed around by a high noise, the implicit curvature penalty would favor a flat
minimum around this weight which leads to its elimination by EBD.
998
H. G. Zimmermann and R. Neuneier
5 Experiments
In a research project sponsored by the European Community we are applying the proposed
approach to estimate the returns of 3 financial markets for each of the G7 countries subsequently using these estimations in an asset allocation scheme to create a Markowitz-optimal
portfolio [4]. This paper reports the 6 month forecasts of the German bond rate, which is
one of the more difficult tasks due to the reunification of Germany and GDR. The inputs
consist of 39 variables achieved by preprocessing 16 relevant financial time series. The
training set covers the time from April, 1974 to December, 1991, the test set runs from J anuary, 1992 to May, 1996. The network arcitecture consists of one hidden layer (20 neurons,
tanh transfer function) and one linear output. First, we trained the neural network until
convergence with pattern-by-pattern learning using a small batch size of 20 patterns (classical approach). Then, we trained the network using the unified approach as described in
section 4.1 using pattern-by-pattern learning. We compare the resulting predictions of the
networks on the basis of four performance measures (see table). First, the hit rate counts
how often the sign of the return of the bond has been correctly predicted. As to the other
measures, the step from the forecast model to a trading system is here kept very simple. If
the output is positive, we buy shares of the bond, otherwise we sell them. The potential
realized is the ratio of the return to the maximum possible return over the test (training)
set. The annualized return is the average yearly profit of the trading systems. Our approach
turns out to be superior: we almost doubled the annualized return from 4.5% to 8.5% on
the test set. The figure compares the accumulated return of the two approaches on the test
set. The unified approach not only shows a higher profitability, but also has by far a less
maximal draw down.
35 i -- -- ~ -- -.., -
I
approach
II
our
30\
classical
- ,
-
-.
81%(96%)
66%(93%)
realized potential
75%(100%)
44%(96%)
annualized return
8.5% (11.2%)
4.5%(10.1%)
!!!20
~'5'JI
j 'O' ., ' .
1i
I
" '-
-
-.
.
I
~~~/.//
.3: 25 r
hit rate
-
-
"
"
.,
::. .... . ..... ... .
.{--'0? - 20 -" 30
...
40 -
sO
I
- -So
11...
References
[I] Christopher M. Bishop. Neural Networks for Pattern Recognition. Clarendon Press, 1994.
[2] W. Finnoff, F. Hergert, and H. G. Zimmennann. Improving generalization perfonnance by
nonconvergentmodel selection methods. In proc. of ICANN-92, 1992.
[3] W. Finnoff, F. Hergert, and H. G. Zimmennann. Neuronale Lemverfahren mit variabler Schrittweite. 1993. Tech. report, Siemens AG.
[4] P. Herve, P. Nairn, and H. G. Zimmennann. Advanced Adaptive Architectures for Asset Allocation: A Trial Application. In Forecasting Financial Markets, 1996.
[5] S. Hochreiter and J. Schmid huber. Flat minima. Neural Computation, 9(1): 1-42, 1997.
[6] Y. Ie Cun, J. S. Denker, and S. A. Solla. Optimal brain damage. NIPS*89, 1990.
[7] J. E. Moody and T. S. Rognvaldsson . Smoothing regularizers for projective basis function
networks. NIPS 9, 1997.
[8] R. Neuneier and H. G. Zimmennann. How to Train Neural Networks. In Tricks of the Trade:
How to make algorithms really to work. Springer Verlag, Berlin, 1998.
[9] B. Tang, W. Hsieh, and F. Tangang. Cleaming neural networks with continuity constraints for
prediction of noisy time series. ICONIP '96, 1996.
[10] V. Tresp, R. Neuneier, and H. G. Zimmennann. Early brain damage. NIPS 9, 1997.
[II] A. S. Weigend, H. G. Zimmennann, and R. Neuneier. Cleaming. Neural Networks in Financial
Engineering, (NNCM95), 1995.
| 1385 |@word trial:1 version:3 hsieh:1 profit:1 series:3 suppressing:1 neuneier:8 written:1 enables:1 sponsored:1 update:3 resampling:1 tenn:2 mathematical:1 consists:1 combine:1 manner:1 huber:1 market:3 expected:2 behavior:1 brain:3 td:1 actual:1 considering:2 project:1 underlying:1 easiest:1 interpreted:1 unified:5 ag:2 act:2 expands:1 exactly:2 preferable:1 hit:2 control:6 appear:1 before:1 positive:1 engineering:1 local:9 mistake:1 consequence:1 limit:1 analyzing:1 fluctuation:2 might:1 annualized:3 limited:1 collapse:1 g7:1 projective:1 testing:1 practice:1 backpropagation:1 procedure:5 get:1 doubled:1 close:1 selection:1 context:1 influence:1 applying:1 yt:11 independently:1 wit:1 rule:1 financial:6 stability:1 searching:2 anyway:1 handle:1 increment:4 analogous:2 target:1 exact:1 cleaning:15 hypothesis:1 trick:1 assure:1 element:4 approximated:2 recognition:1 asymmetric:2 observed:4 solla:1 decrease:1 trade:1 trained:2 depend:1 dilemma:8 creates:1 basis:2 rognvaldsson:1 tx:1 train:2 separated:3 fast:2 effective:1 artificial:1 choosing:1 whose:3 solve:1 otherwise:2 favor:2 gi:2 revive:1 noisy:5 itself:1 final:2 sequence:1 advantage:1 exemplary:1 propose:5 interaction:3 maximal:1 adaptation:1 relevant:1 combining:1 realization:2 achieve:1 description:1 convergence:3 empty:1 disturb:1 perfect:2 hergert:2 depending:1 advisable:1 measured:1 eq:6 predicted:1 memorize:1 indicate:1 trading:2 concentrate:1 correct:1 stochastic:9 subsequently:1 exploration:1 human:2 stabilization:1 elimination:1 generalization:8 really:1 adjusted:1 correction:4 around:5 wdi:1 normal:1 achieves:1 vary:1 early:2 estimation:1 proc:1 bond:4 iw:1 infonnation:1 tanh:1 individually:1 correctness:2 create:1 minimization:2 mit:1 gaussian:3 aim:1 improvement:1 likelihood:1 indicates:1 tech:1 contrast:1 dependent:2 accumulated:1 typically:3 hidden:1 favoring:1 interested:1 germany:3 nonnal:1 ralph:2 favored:1 priori:1 smoothing:1 field:1 represents:1 sell:1 nairn:1 cancel:1 markowitz:1 report:2 serious:1 randomly:1 ve:1 individual:3 phase:1 acceptance:1 possibility:1 evaluation:1 nonnally:1 introduces:2 regularizers:1 held:1 necessary:1 herve:1 perfonnance:2 taylor:2 initialized:1 clearning:2 minimal:1 fitted:1 psychological:2 eta:5 cover:4 disadvantage:1 cost:1 deviation:1 imperfectly:1 uniform:6 wonder:1 successful:1 reported:1 corrupted:2 combined:1 cited:1 ie:1 connecting:1 together:1 moody:1 again:2 possibly:1 external:1 derivative:1 return:8 potential:2 de:2 includes:1 vario:5 depends:2 piece:1 later:1 try:1 observer:6 view:1 reached:1 complicated:1 minimize:2 variance:4 characteristic:1 produced:1 basically:1 asset:2 explain:1 simultaneous:1 canceling:1 definition:1 failure:1 ty:1 against:1 finnoff:2 knowledge:3 clarendon:1 higher:1 follow:1 improved:1 april:1 done:2 evaluated:1 furthermore:1 profitability:1 implicit:3 until:1 hand:3 christopher:1 nonlinear:1 continuity:1 believe:1 gti:1 building:2 effect:3 concept:1 true:1 regularization:1 symmetric:2 iteratively:1 during:2 everybody:1 iconip:1 gh:1 fj:1 superior:1 ji:1 smoothness:1 inclusion:1 portfolio:1 han:1 stiffer:1 operating:1 gt:7 add:3 curvature:4 optimizing:1 driven:1 verlag:1 minimum:9 additional:1 ii:3 smoother:1 corporate:2 smooth:1 xf:4 characterized:1 divided:1 concerning:2 prediction:2 variant:1 neuro:4 regression:3 expectation:1 achieved:1 hochreiter:1 receive:1 whereas:3 want:1 fine:1 annealing:1 country:1 operate:1 december:1 identically:1 easy:1 fit:2 architecture:1 identified:1 topology:1 economic:1 ytd:1 forecasting:9 penalty:15 hessian:2 constitute:2 dramatically:1 useful:3 clear:1 unimportant:1 tune:1 locally:1 sign:1 estimated:1 correctly:1 ebd:6 write:1 georg:2 taught:1 four:1 achieving:1 drawn:3 changing:2 clean:2 kept:1 defect:1 sum:2 weigend:1 run:1 pattem:4 almost:1 hjj:2 separation:1 draw:1 bit:1 pushed:1 layer:1 strength:1 constraint:2 flat:4 nearby:1 generates:1 aspect:1 pruned:1 according:2 alternate:1 combination:4 wi:1 cun:1 making:1 outlier:2 invariant:5 explained:1 zimmermann:4 heart:1 turn:2 german:2 mechanism:1 count:1 know:1 denker:1 munchen:2 away:1 appropriate:1 hii:2 batch:4 encounter:1 assumes:1 newton:1 unifying:1 yx:1 yearly:1 giving:1 build:2 especially:5 approximating:1 classical:2 implied:2 g0:1 already:1 quantity:1 realized:2 damage:3 usual:1 diagonal:2 gradient:4 separate:2 berlin:1 hxx:1 reason:1 assuming:1 index:2 relationship:1 ratio:1 difficult:1 disastrous:1 potentially:1 implementation:1 tj2:1 perform:1 observation:6 neuron:1 descent:2 situation:1 misspecification:1 y1:1 community:1 drift:2 introduced:1 cleaned:6 learned:1 nip:3 able:2 pattern:23 summarize:1 memory:1 xtlt:1 force:2 disturbance:1 residual:5 advanced:1 representing:1 scheme:1 improve:2 technology:2 lvar:1 schmid:1 tresp:1 understanding:2 law:1 allocation:2 proven:1 var:8 analogy:1 principle:1 share:1 row:1 course:1 obd:3 free:2 liability:4 neuronale:1 characterizing:1 mchp:2 distributed:1 benefit:1 world:1 cumulative:1 computes:1 disturbing:1 adaptive:7 simplified:2 preprocessing:1 far:1 approximate:1 pruning:8 emphasize:1 implicitly:2 preferred:1 keep:1 global:1 overfitting:4 incoming:1 buy:1 search:6 vet:1 why:2 reviewed:3 additionally:1 table:1 learn:2 transfer:1 improving:1 expansion:2 european:1 icann:1 motivation:1 noise:47 whole:1 x1:1 advice:1 slow:1 experienced:1 candidate:1 tang:1 oxj:3 formula:1 down:1 xt:10 bishop:1 consist:2 gdr:1 forecast:3 springer:1 quasinewton:1 determines:1 viewed:1 presentation:2 month:1 unexplainable:2 change:1 typical:2 infinite:1 corrected:1 wt:5 called:1 tendency:1 siemens:4 support:1 violated:1 evaluate:2 phenomenon:2 |
423 | 1,386 | Automated Aircraft Recovery
via Reinforcement Learning:
Initial Experiments
Jeffrey F. Monaco
Barron Associates, Inc.
Jordan Building
1160 Pepsi Place, Suite 300
Charlottesville VA 22901
monaco@bainet.com
David G. Ward
Barron Associates, Inc.
Jordan Building
1160 Pepsi Place, Suite 300
Charlottesville VA 22901
ward@bainet.com
Andrew G. Barto
Department of Computer Science
University of Massachusetts
Amherst MA 01003
barto@cs.umass.edu
Abstract
Initial experiments described here were directed toward using reinforcement learning (RL) to develop an automated recovery system (ARS) for
high-agility aircraft. An ARS is an outer-loop flight-control system designed to bring an aircraft from a range of out-of-control states to straightand-level flight in minimum time while satisfying physical and physiological constraints. Here we report on results for a simple version
of the problem involving only single-axis (pitch) simulated recoveries.
Through simulated control experience using a medium-fidelity aircraft
simulation, the RL system approximates an optimal policy for pitch-stick
inputs to produce minimum-time transitions to straight-and-Ievel flight in
unconstrained cases while avoiding ground-strike. The RL system was
also able to adhere to a pilot-station acceleration constraint while executing simulated recoveries.
Automated Aircraft Recovery via Reinforcement Learning
1023
1 INTRODUCTION
An emerging use of reinforcement learning (RL) is to approximate optimal policies for
large-scale control problems through extensive simulated control experience. Described
here are initial experiments directed toward the development of an automated recovery system (ARS) for high-agility aircraft. An ARS is an outer-loop flight control system designed
to bring the aircraft from a range of initial states to straight, level, and non-inverted flight in
minimum time while satisfying constraints such as maintaining altitude and accelerations
within acceptable limits. Here we describe the problem and present initial results involving
only single-axis (pitch) recoveries. Through extensive simulated control experience using
a medium-fidelity simulation of an F-16, the RL system approximated an optimal policy
for longitudinal-stick inputs to produce near-minimum-time transitions to straight and level
flight in unconstrained cases, as well as while meeting a pilot-station acceleration constraint.
2
AIRCRAFT MODEL
The aircraft was modeled as a dynamical system with state vector x = {q, 0, p, r, {3, Vi},
where q = body-axes pitch rate, 0 = angle of attack, p = body-axes roll rate, r =
body-axes yaw rate, {3 = angle of sideslip, Vi = total airspeed, and control vector fl =
{fl se , flae, fla/' Orud} of effector and pseudo-effector displacements. The controls are defined as: flse = symmetric elevon, oae = asymmetric elevon, oal = asymmetric flap, and
Orud = rudder. (A pseudo-effector is a mathematically convenient combination of real effectors that, e.g., contributes to motion in a limited number of axes.) The following additional descriptive variables were used in the RL problem formulation: h = altitude, h =
vertical component of velocity, e = pitch attitude, N z = pilot-station normal acceleration.
For the initial pitch-axis experiment described here, five discrete actions were available
to the learning agent in each state; these were longitudinal-stick commands selected from
{-6, -3,0, +3, +6} lbf. The command chosen by the learning agent was converted into
a desired normal-acceleration command through the standard F-16 longitudinal-stick command gradient with software breakout. This gradient maps pounds-of-force inputs into desired acceleration responses. We then produce an approximate relationship between normal
acceleration and body-axes pitch rate to yield a pitch-rate flying-qualities model. Given this
model, an inner-loop linear-quadratic (LQ) tracking control algorithm determined the actuator commands to result in optimal model-following of the desired pitch-rate response.
The aircraft model consisted of complete translational and rotational dynamics, including
nonlinear terms owing to inertial cross-coupling and orientation-dependent gravitational effects. These were obtained from a modified linear F-16 model with dynamics of the form
j; = Ax + Bfl + b + N
where A and B were the F-16 aero-inertial parameters (stability derivatives) and effector
sensitivities (control derivatives). These stability and control derivatives and the bias vector, b, were obtained from linearizations of a high-fidelity nonlinear, six-degree-of-freedom
model. Nonlinearities owing to inertial cross-coupling and orientation-dependent gravitational effects were accounted for through the term N, which depended nonlinearly on the
state. Nonlinear actuator dynamics were modeled via the incorporation ofF-16 effector-rate
and effector-position limits. See Ward et al. (1996) for additional details.
3
PROBLEM FORMULATION
The RL problem was to approximate a minimum-time control policy capable of bringing the
aircraft from a range of initial states to straight, level, and non-inverted flight, while satisfying given constraints, e.g., maintaining the normal acceleration at the pilot station within
1. F. Monaco, D. G. Ward and A G. Barto
1024
an acceptable range. For the single-axis (pitch-axis) flight control problem considered here,
recovered flight was defined by:
q
= q = it = h = i't
=
o.
(1)
Successful recovery was achieved when all conditions in Eq. 1 were satisfied simultaneously within pre-specified tolerances.
Because we wished to distinguish between recovery supplied by the LQ tracker and that
learned by the RL system, special attention was given to formulating a meaningful test to
avoid falsely attributing successes to the RL system. For example, if initial conditions were
specified as off-trim perturbations in body-axes pitch rate, pitch acceleration, and true airspeed, the RL system may not have been required because the LQ controller would provide
all the necessary recovery, i.e., zero longitudinal-stick input would result in a commanded
body-axes pitch rate of zero deg./ sec. Because this controller is designed to be highly responsive, its tracking and integral-error penalties usually ensure that the aircraft responses
attain the desired state in a relatively short time. The problem was therefore formulated to
demand recovery from aircraft orientations where the RL system was primarily responsible
for recovery, and the goal state was not readily achieved via the stabilizing action of the LQ
control law.
A pitch-axis recovery problem of interest is one in which initial pitch attitude, e, is selected
to equal e trim +U(80 ,n' 8 0Tn a:l:)' where e trim == atrim (by definition), U is a uniformly
distributed random number, and eO Tnin and eoTnaz define the boundaries of the training region, and other variables are set so that when the aircraft is parallel to the earth (8 0 = 0),
it is "pancaking" toward the ground (with positive trim angle of attack). Other initial conditions correspond to purely-translational climb or descent of the aircraft. For initial conditions where eo < atrim, the flight vehicle will descend, and in the absence of any corrective
longitudinal-stick force, strike the ground or water. Because it imposes no constraints on altitude or pitch-angle variations, the stabilizing response of the LQ controller is inadequate
for providing the necessary recovery.
Tn
4
REINFORCEMENT LEARNING ALGORITHM
Several candidate RL algorithms were evaluated for the ARS. Initial efforts focused primarily on (1) Q-Learning, (2) alternative means for approximating the action-value function (Q
function), and (3) use of discrete versus continuous-action controls. During subsequent investigations, an extension of Q-Learning called Residual Advantage Learning (Baird, 1995;
Harmon & Baird, 1996) was implemented and successfully applied to the pitch-axis ARS
problem. As with action-values in Q-Learning, the advantage function, A(x, u), may be
represented by a function approximation system of the form
A(x,u) = ?(x,ufO,
(2)
where ?( x, u) is a vector of relevant features and 0 are the corresponding weights. Here, the
advantage function is linear in the weights, 0, and these weights are the modifiable, learned
parameters.
For advantage functions of the form in Eq. 2, the update rule is:
Ok+l
Ok - a ((r
? (
+ "Y~t A(y, b*))
K~t + (1 - K~t) A(x, a*) -
~"Y~t?(y, b*) K~t + ~ (1 - K~t) ?(x, a*) -
A(x, a))
?(x, a)) ,
where a* = argminaA(x, a) and b* = argminbA(y, b), !l.t is the system rate (0.02 sec.
in the ARS), "Y~t is the discount factor, and K is an fixed scale factor. In the above notation,
Automated Aircraft Recovery via Reinforcement Learning
1025
y is the resultant state, i.e., the execution of action a results in a transition from state x to
its successor y.
The Residual Advantage Learning update collapses to the Q-Learning update for the case
~ = 0, K =
The parameter ~ is a scalar that controls the trade-off between residualgradient descent when ~ = 1, and a faster, direct algorithm when ~ = O. Harmon & Baird
(1996) address the choice of ~, suggesting the following computation of ~ at each time step:
L.
;F,.
'J!'=
l:o WdWrg
+J-L
l:o(Wd - wrg)wrg
where Wd and Wrg are traces (one for each (J of the function approximation system) associated with the direct and residual gradient algorithms, respectively, and J-L is a small, positive
constant that dictates how rapidly the system forgets. The traces are updated during each
cycle as follows
Wd
f-
(1-J-L)Wd-J-L[(r+'Y~tA(y,b*)) K~t+(1- K~t)A(X,a*)]
? [- :(JA(x, a*)]
wrg
f-
(1-J-L)Wrg-J-L[(r+'Y~tA(y,b*?K~t+(1- K~t)A(x,a*)-A(X,a)]
?
['Y~t ;(JA(y,b*) K~t + (1- K~t) ;(JA(x,a*) -
;(JA(X, a)] .
Advantage Learning updates of the weights, including the calculation of an adaptive ~ as
discussed above, were implemented and interfaced with the aircraft simulation. The Advantage Learning algorithm consistently outperformed its Q-Learning counterpart. For this
reason, most of our efforts have focused on the application of Advantage Learning to the
solution of the ARS. The feature vector 4>(x, u) consisted of normalized (dimensionless)
states and controls, and functions ofthese variables. Use ofthese nondimensionalized variables (obtained via the Buckingham 7r-theorem; e.g., Langharr, 1951) was found to enhance
greatly the stability and robustness of the learning process. Furthermore, the RL system appeared to be less sensitive to changes in parameters such as the learning rate when these
techniques were employed.
5
TRAINING
Training the RL system for arbitrary orientations was accomplished by choosing random
initial conditions on e as outlined above. With the exception of h, all other initial conditions corresponded to trim values for a Mach 0.6, 5 kIt. flight condition. Rewards were
-1 per-time-step until the goal state was reached. In preliminary experiments, the training
region was restricted to ? 0.174 rad.(l0 deg.) from the trim pitch angle. For this range of
initial conditions, the system was able to learn an appropriate policy given only a handful of
features (approximately 30). The policy was significantly mature after 24 hours oflearning
on an HP-730 workstation and appeared to be able to achieve the goal for arbitrary initial
conditions in the aforementioned domain.
We then expanded the training region and considered initial e values within
? 0.785 rad. (45 deg .) of trim. The policy previously learned for the more restricted
training domain performed well here too, and learning to recover for these more drastic
off-trim conditions was trivial. No boundary restrictions were imposed on the system, but
a report of whether the aircraft would have struck the ground was maintained. It was noted
1. R Monaco, D. G. Ward and A. G. Barto
1026
that recovery from all possible initial conditions could not be achieved without hitting
the ground. Episodes in which the ground would have been encountered were a result
of inadequate control authority and not an inadequate RL policy. For example, when the
initial pitch angle was at its maximum negative value, maximum-allowable positive stick
(6 lbf.) was not sufficient to pull up the aircraft nose in time. To remedy this in subsequent
experiments, the number of admissible actions was increased to include larger-magnitude
commands: {-12, -9, -6, -3,0, +3, +6, +9, +12} lbf.
Early attempts at solving the pitch-axis recovery problem with the expanded initial conditions in conjunction with this augmented action set proved challenging. The policy that
worked well in the two previous experiments was no longer able to attain the goal state;
it was only able to come close and oscillate indefinitely about the goal region. The agent
learned to pitch up and down appropriately, e.g., when h was negative it applied a corrective
positive action, and vice versa. However, because of system and actuator dynamics modeled in the simulation, the transient response caused the aircraft to pass through the goal
state. Once beyond the goal region, the agent applied an opposite action, causing it to approach the goal state again, repeating the process indefinitely (until the system was reset
and a new trial was started). Thus, the availability of large-amplitude commands and the
presence of actuator dynamics made it difficult for the agent to fonnulate a consistent policy that afforded all goal state criteria being satisfied simultaneously. One might remedy
the problem by removing the actuator dynamics; however, we did not wish to compromise
simulation fidelity, and chose to use an expanded feature set to improve RL perfonnance.
Using a larger collection offeatures with approximately 180 inputs, the RL agent was able
to formulate a consistent recovery policy. The learning process required approximately 72
hours on an HP-730 workstation. (On this platform, the combined aircraft simulation and
RL software execution rate was approximately twice that of real-time.) At this point performance was evaluated. The simulation was run in evaluation mode, i.e., learning rate was
set to zero and random exploration was disabled. Performance is summarized below.
6
6.1
RESULTS
UNCONSTRAINED PITCH-AXIS RECOVERY
Fig. 1 shows the transition times from off-trim orientations to the goal state as a function
of initial pitch (inclination) angle. Recovery times were approximately 11-12 sec. for the
worst-case scenarios. i.e .? 180 1= 45 deg. off-trim. and decrease (almost) monotonically
for points closer to the unperturbed initial conditions. The occasional "blips" in the figure
suggest that additional learning would have improved the global RL performance slightly.
For 180 1 = 45 deg. off-trim, maximum altitude loss and gain were each approximately 1667
ft. (0.33 x 5000 f t. ). These excursions may seem substantial. but when one looks atthe time
histories for these maneuvers, it is apparent that the RL-derived policy was perfonning well.
The policy effectively minimizes any altitude variation; the magnitude of these changes are
principally governed by available control authority and the severity of the flight condition
from which the policy must recover.
Fig. 2 shows time histories of relevant variables for one of the limiting cases. The first column shows body-axes pitch rate (Qb) and commanded body-axes pitch rate (Qbmodel)
in (deg./sec.), pilot station nonnal acceleration (Nz) in (g), angle of attack (Alpha) in
(deg.). and pitch attitude (Theta) in (deg.), respectively. The second column shows the
longitudinal stick action executed by the RL system (lbf.), the left and right elevator deflections (deg.). total airspeed (ft./ sec.), and altitude (ft.). The majority ofthe 1600+ ft.
altitude loss occurs between zero and five sec.; during this time, the RL system is applying
maximum (allowable) positive stick. Thus, this altitude excursion is principally attributed
to limited control authority as well as significant off-trim initial orientations.
Automated Aircraft Recovery via Reinforcement Learning
1027
20
18
16
14
Recovery Time (sec.)
12
10
8
6
4
2
O~~rn~~~~~~~~~~~~~~~~
?50
?40
?30
?20
?10
0
10
30
20
40
50
Figure 1: Simulated Aircraft Recovery Times for Unconstrained Pitch-Axis ARS
-
_. 00.-
00.
L
?
-
~
?
2
?
"
,
12
j
11?
?
-
-IU
jV ~
?
2
?
?
?
2
?
?
-:t: ~
,
"
12
1.
12
?
2
?
?
?
"
cC."
-_
,'z
.......
...........
he
o
--
2
?
I
10
,
-.-
~,
",,--
,
"
t==
?
2
?
-"0
Figure 2: TIme Histories During Unconstrained Pitch-Axis Recovery for 8
45 deg .
0
=
8 trim
-
I. R Monaco, D. G. Ward and A G. Barto
1028
6.2
CONSTRAINED PITCH-AXIS RECOVERY
The requirement to execute aircraft recoveries while adhering to pilot-safety constraints was
a deciding factor in using RL to demonstrate the automated recovery system concept. The
need to recover an aircraft while minimizing injury and, where possible, discomfort to the
flight crew, requires that the controller incorporate constraints that can be difficult or impossible to express in forms suitable for linear and nonlinear programming methods.
In subsequent ARS investigations, allowable pilot-station normal acceleration was restricted to the range -1.5 9 ~ N z ~ 3.5 g. These values were selected because the unconstrained ARS was observed to exceed these limits. Several additional features (for a
total of 189) were chosen, and the learning process was continued. Initial weights for the
original 180 inputs corresponded to those from the previously learned policy; the new features were chosen to have zero weights initially. Here, the RL system learned to avoid the
normal acceleration limits and consistently reach the goal state for initial pitch angles in the
region [-45 + 8 trim , 35 + 8 trim ] deg. Additional learning should result in improved recovery policies in this bounded acceleration domain for all initial conditions. Nonetheless,
the results showed how an RL system can learn to satisfy these kinds of constraints.
7 CONCLUSION
In addition to the results reported here, we conducted extensive analysis of the degree to
which the learned policy successfully generalized to a range of initial conditions not experienced in training. In all cases, aircraft responses to novel recovery scenarios were stable
and qualitatively similar to those previously executed in the training region. We are also
conducting experiments with a multi-axes ARS, in which longitudinal-stick and lateral-stick
sequences must be coordinated to recover the aircraft. Initial results are promising, but substantially longer training times are required. In summary, we believe that the results presented here demonstrate the feasibility of using RL algorithms to develop robust recovery
strategies for high-agility aircraft, although substantial further research is needed.
Acknowledgments
This work was supported by the Naval Air Warfare Center Aircraft Division (NAWCAD),
Flight Controls/Aeromechanics Division under Contract N62269-96-C-0080. The authors
thank Marc Steinberg, the Program Manager and Chief Technical Monitor. The authors also
express appreciation to Rich Sutton and Mance Harmon for their valuable help, and to Lockheed Martin Tactical Aircraft Systems for authorization to use their ATLAS software, from
which F-16 parameters were extracted.
References
Baird, L. C. (1995) Residual algorithms: reinforcement learning with function approximation. In A. Prieditis and S. Russell (eds.), Machine Learning: Proceedings of the Twelfth
International Conference, pp. 30-37. San Francisco, CA: Morgan Kaufmann.
Harmon, M. E. & Baird, L. C. (1996) Multi-agent residual advantage learning with general
function approximation. Wright Laboratory Technical Report, WPAFB, OH.
Langharr, H. L. (1951) Dimensional Analysis and Theory of Models. New York: Wiley and
Sons.
Ward, D. G., Monaco, J. E, Barron, R. L., Bird, R.A., Virnig, J.C., & Landers, T.E (1996)
Self-designing controller. Final Tech. Rep. for Directorate of Mathematics and Computer
Sciences, AFOSR, Contract F49620-94-C-0087. Barron Associates, Inc.
| 1386 |@word aircraft:30 trial:1 version:1 oae:1 twelfth:1 simulation:7 initial:28 uma:1 longitudinal:7 recovered:1 com:2 wd:4 buckingham:1 must:2 readily:1 subsequent:3 designed:3 atlas:1 update:4 selected:3 short:1 indefinitely:2 authority:3 attack:3 five:2 direct:2 rudder:1 falsely:1 multi:2 manager:1 notation:1 bounded:1 medium:2 kind:1 minimizes:1 emerging:1 substantially:1 suite:2 pseudo:2 stick:11 control:22 maneuver:1 positive:5 safety:1 limit:4 depended:1 sutton:1 mach:1 approximately:6 might:1 chose:1 twice:1 nz:1 bird:1 challenging:1 limited:2 commanded:2 collapse:1 range:7 directed:2 acknowledgment:1 responsible:1 displacement:1 attain:2 wrg:5 convenient:1 dictate:1 pre:1 significantly:1 suggest:1 close:1 dimensionless:1 applying:1 impossible:1 restriction:1 map:1 imposed:1 center:1 attention:1 focused:2 formulate:1 stabilizing:2 adhering:1 recovery:30 rule:1 continued:1 pull:1 oh:1 stability:3 variation:2 updated:1 limiting:1 programming:1 designing:1 associate:3 velocity:1 satisfying:3 approximated:1 asymmetric:2 observed:1 ft:4 aero:1 descend:1 worst:1 region:7 cycle:1 episode:1 trade:1 decrease:1 russell:1 valuable:1 substantial:2 reward:1 dynamic:6 solving:1 compromise:1 flying:1 purely:1 division:2 represented:1 corrective:2 attitude:3 describe:1 corresponded:2 choosing:1 apparent:1 larger:2 ward:7 final:1 descriptive:1 advantage:9 sequence:1 reset:1 causing:1 relevant:2 loop:3 rapidly:1 achieve:1 breakout:1 requirement:1 produce:3 executing:1 help:1 coupling:2 andrew:1 develop:2 wished:1 eq:2 implemented:2 c:1 crew:1 come:1 owing:2 exploration:1 successor:1 transient:1 ja:4 investigation:2 preliminary:1 sideslip:1 mathematically:1 extension:1 gravitational:2 tracker:1 considered:2 ground:6 normal:6 deciding:1 wright:1 early:1 earth:1 outperformed:1 sensitive:1 vice:1 successfully:2 modified:1 avoid:2 barto:5 command:7 conjunction:1 ax:11 l0:1 derived:1 naval:1 consistently:2 greatly:1 tech:1 warfare:1 dependent:2 initially:1 nonnal:1 iu:1 translational:2 fidelity:4 orientation:6 aforementioned:1 development:1 platform:1 special:1 constrained:1 equal:1 once:1 look:1 yaw:1 report:3 primarily:2 residualgradient:1 simultaneously:2 elevator:1 jeffrey:1 attempt:1 freedom:1 interest:1 highly:1 evaluation:1 integral:1 capable:1 closer:1 necessary:2 experience:3 perfonnance:1 harmon:4 desired:4 effector:7 increased:1 column:2 ar:12 injury:1 oflearning:1 successful:1 inadequate:3 conducted:1 too:1 reported:1 combined:1 international:1 amherst:1 sensitivity:1 contract:2 off:8 enhance:1 offeatures:1 again:1 satisfied:2 derivative:3 suggesting:1 converted:1 nonlinearities:1 sec:7 summarized:1 tactical:1 availability:1 blip:1 inc:3 coordinated:1 satisfy:1 baird:5 caused:1 vi:2 fonnulate:1 vehicle:1 performed:1 reached:1 recover:4 parallel:1 air:1 roll:1 kaufmann:1 ufo:1 conducting:1 yield:1 correspond:1 interfaced:1 ofthe:1 cc:1 straight:4 history:3 reach:1 ed:1 definition:1 nonetheless:1 pp:1 resultant:1 associated:1 attributed:1 wpafb:1 workstation:2 gain:1 pilot:7 proved:1 massachusetts:1 amplitude:1 inertial:3 ok:2 monaco:6 ta:2 response:6 improved:2 formulation:2 evaluated:2 pound:1 execute:1 furthermore:1 until:2 flight:14 nonlinear:4 mode:1 fla:1 quality:1 disabled:1 believe:1 building:2 effect:2 consisted:2 true:1 remedy:2 counterpart:1 concept:1 normalized:1 symmetric:1 laboratory:1 during:4 self:1 maintained:1 noted:1 criterion:1 generalized:1 allowable:3 complete:1 demonstrate:2 tn:2 motion:1 discomfort:1 bring:2 novel:1 rl:26 physical:1 discussed:1 he:1 approximates:1 significant:1 versa:1 unconstrained:6 outlined:1 mathematics:1 hp:2 stable:1 longer:2 showed:1 scenario:2 rep:1 success:1 meeting:1 accomplished:1 inverted:2 morgan:1 minimum:5 additional:5 kit:1 eo:2 employed:1 strike:2 monotonically:1 technical:2 faster:1 calculation:1 cross:2 va:2 feasibility:1 pitch:30 involving:2 controller:5 achieved:3 addition:1 lander:1 adhere:1 appropriately:1 bringing:1 bfl:1 mature:1 climb:1 seem:1 jordan:2 near:1 presence:1 exceed:1 automated:7 opposite:1 inner:1 prieditis:1 whether:1 six:1 effort:2 penalty:1 york:1 oscillate:1 action:11 se:1 ievel:1 repeating:1 discount:1 supplied:1 per:1 modifiable:1 discrete:2 express:2 monitor:1 jv:1 oal:1 linearizations:1 angle:9 run:1 deflection:1 place:2 almost:1 excursion:2 acceptable:2 fl:2 distinguish:1 quadratic:1 encountered:1 constraint:9 pepsi:2 incorporation:1 handful:1 worked:1 software:3 flap:1 afforded:1 formulating:1 qb:1 expanded:3 relatively:1 martin:1 department:1 combination:1 slightly:1 son:1 ofthese:2 restricted:3 principally:2 altitude:8 previously:3 needed:1 nose:1 drastic:1 available:2 mance:1 actuator:5 occasional:1 barron:4 appropriate:1 responsive:1 alternative:1 robustness:1 original:1 ensure:1 include:1 maintaining:2 approximating:1 occurs:1 strategy:1 gradient:3 thank:1 simulated:6 lateral:1 majority:1 outer:2 perfonning:1 trivial:1 toward:3 water:1 reason:1 modeled:3 relationship:1 rotational:1 providing:1 minimizing:1 difficult:2 executed:2 trace:2 negative:2 policy:17 appreciation:1 vertical:1 descent:2 agility:3 severity:1 rn:1 perturbation:1 station:6 arbitrary:2 david:1 nonlinearly:1 required:3 specified:2 extensive:3 struck:1 rad:2 inclination:1 learned:7 hour:2 address:1 able:6 beyond:1 dynamical:1 usually:1 below:1 appeared:2 program:1 including:2 suitable:1 force:2 residual:5 improve:1 theta:1 axis:12 started:1 charlottesville:2 law:1 afosr:1 loss:2 versus:1 agent:7 degree:2 sufficient:1 consistent:2 imposes:1 summary:1 accounted:1 supported:1 bias:1 tolerance:1 distributed:1 boundary:2 f49620:1 transition:4 rich:1 author:2 made:1 reinforcement:8 adaptive:1 collection:1 qualitatively:1 san:1 approximate:3 alpha:1 trim:15 lockheed:1 deg:11 global:1 francisco:1 continuous:1 chief:1 promising:1 learn:2 robust:1 ca:1 contributes:1 domain:3 marc:1 did:1 directorate:1 body:8 augmented:1 fig:2 wiley:1 experienced:1 position:1 wish:1 lq:5 candidate:1 governed:1 forgets:1 steinberg:1 admissible:1 theorem:1 down:1 removing:1 unperturbed:1 physiological:1 effectively:1 magnitude:2 execution:2 demand:1 attributing:1 hitting:1 tracking:2 scalar:1 extracted:1 ma:1 goal:11 formulated:1 acceleration:13 absence:1 change:2 determined:1 uniformly:1 total:3 called:1 pas:1 meaningful:1 exception:1 lbf:4 incorporate:1 avoiding:1 |
424 | 1,387 | Local Dimensionality Reduction
Stefan Schaal
1,2,4
sschaal@usc.edu
http://www-slab.usc.edulsschaal
Sethu Vijayakumar 3, I
Christopher G. Atkeson 4
sethu@cs.titech.ac.jp
http://ogawawww.cs.titech.ac.jp/-sethu
cga@cc.gatech.edu
http://www.cc.gatech.edul
fac/Chris.Atkeson
IERATO Kawato Dynamic Brain Project (IST), 2-2 Hikaridai, Seika-cho, Soraku-gun, 619-02 Kyoto
2Dept. of Comp. Science & Neuroscience, Univ. of South. California HNB-I 03, Los Angeles CA 90089-2520
3Department of Computer Science, Tokyo Institute of Technology, Meguro-ku, Tokyo-I 52
4College of Computing, Georgia Institute of Technology, 801 Atlantic Drive, Atlanta, GA 30332-0280
Abstract
If globally high dimensional data has locally only low dimensional distributions, it is advantageous to perform a local dimensionality reduction before
further processing the data. In this paper we examine several techniques for
local dimensionality reduction in the context of locally weighted linear regression. As possible candidates, we derive local versions of factor analysis
regression, principle component regression, principle component regression
on joint distributions, and partial least squares regression. After outlining the
statistical bases of these methods, we perform Monte Carlo simulations to
evaluate their robustness with respect to violations of their statistical assumptions. One surprising outcome is that locally weighted partial least
squares regression offers the best average results, thus outperforming even
factor analysis, the theoretically most appealing of our candidate techniques.
1 INTRODUCTION
Regression tasks involve mapping a n-dimensional continuous input vector x E ~n onto
a m-dimensional output vector y E ~m ? They form a ubiquitous class of problems found
in fields including process control, sensorimotor control, coordinate transformations, and
various stages of information processing in biological nervous systems. This paper will
focus on spatially localized learning techniques, for example, kernel regression with
Gaussian weighting functions. Local learning offer advantages for real-time incremental
learning problems due to fast convergence, considerable robustness towards problems of
negative interference, and large tolerance in model selection (Atkeson, Moore, & Schaal,
1997; Schaal & Atkeson, in press). Local learning is usually based on interpolating data
from a local neighborhood around the query point. For high dimensional learning problems, however, it suffers from a bias/variance dilemma, caused by the nonintuitive fact
that " ... [in high dimensions] if neighborhoods are local, then they are almost surely
empty, whereas if a neighborhood is not empty, then it is not local." (Scott, 1992, p.198).
Global learning methods, such as sigmoidal feedforward networks, do not face this
634
S. School, S. Vijayakumar and C. G. Atkeson
problem as they do not employ neighborhood relations, although they require strong
prior knowledge about the problem at hand in order to be successful.
Assuming that local learning in high dimensions is a hopeless, however, is not necessarily warranted: being globally high dimensional does not imply that data remains high dimensional if viewed locally. For example, in the control of robot anns and biological
anns we have shown that for estimating the inverse dynamics of an ann, a globally 21dimensional space reduces on average to 4-6 dimensions locally (Vijayakumar & Schaal,
1997). A local learning system that can robustly exploit such locally low dimensional
distributions should be able to avoid the curse of dimensionality.
In pursuit of the question of what, in the context of local regression, is the "right"
method to perfonn local dimensionality reduction, this paper will derive and compare
several candidate techniques under i) perfectly fulfilled statistical prerequisites (e.g.,
Gaussian noise, Gaussian input distributions, perfectly linear data), and ii) less perfect
conditions (e.g., non-Gaussian distributions, slightly quadratic data, incorrect guess of
the dimensionality of the true data distribution). We will focus on nonlinear function approximation with locally weighted linear regression (L WR), as it allows us to adapt a variety of global linear dimensionality reduction techniques, and as L WR has found widespread application in several local learning systems (Atkeson, Moore, & Schaal, 1997;
Jordan & Jacobs, 1994; Xu, Jordan, & Hinton, 1996). In particular, we will derive and
investigate locally weighted principal component regression (L WPCR), locally weighted
joint data principal component analysis (L WPCA), locally weighted factor analysis
(L WF A), and locally weighted partial least squares (LWPLS). Section 2 will briefly outline these methods and their theoretical foundations, while Section 3 will empirically
evaluate the robustness of these methods using synthetic data sets that increasingly violate some of the statistical assumptions of the techniques.
2 METHODS OF DIMENSIONALITY REDUCTION
We assume that our regression data originate from a generating process with two sets of
observables, the "inputs" i and the "outputs" y. The characteristics of the process ensure a functional relation y = f(i). Both i and yare obtained through some measurement device that adds independent mean zero noise of different magnitude in each observable, such that x == i + Ex and y = y + Ey ? For the sake of simplicity, we will only focus on one-dimensional output data (m=l) and functions / that are either linear or
slightly quadratic, as these cases are the most common in nonlinear function approximation with locally linear models. Locality of the regression is ensured by weighting the error of each data point with a weight from a Gaussian kernel:
Wi = exp(-O.5(Xi - Xqf D(Xi - Xq))
(1)
Xtt denotes the query point, and D a positive semi-definite distance metric which determmes the size and shape of the neighborhood contributing to the regression (Atkeson et
aI., 1997). The parameters Xq and D can be determined in the framework of nonparametric statistics (Schaal & Atkeson, in press) or parametric maximum likelihood estimations
(Xu et aI, 1995}- for the present study they are determined manually since their origin is
secondary to the results of this paper. Without loss of generality, all our data sets will set
!,q to the zero vector, compute the weights, and then translate the input data such that the
locally weighted mean, i = L WI Xi / L Wi , is zero. The output data is equally translated to
be mean zero. Mean zero data is necessary for most of techniques considered below. The
(translated) input data is summarized in the rows of the matrix X, the corresponding
(translated) outputs are the elements of the vector y, and the corresponding weights are in
the diagonal matrix W. In some cases, we need the joint input and output data, denoted
as Z=[X y).
Local Dimensionality Reduction
635
2.1 FACTORANALYSIS(LWFA)
Factor analysis (Everitt, 1984) is a technique of dimensionality reduction which is the
most appropriate given the generating process of our regression data. It assumes the observed data z was produced. by a mean zero independently distributed k -dimensional
vector of factors v, transformed by the matrix U, and contaminated by mean zero independent noise f: with diagonal covariance matrix Q:
z=Uv+f:, where
z=[xT,yt
and
f:=[f:~,t:yr
(2)
If both v and f: are normally distributed, the parameters Q and U can be obtained iteratively by the Expectation-Maximization algorithm (EM) (Rubin & Thayer, 1982). For a
linear regression problem, one assumes that z was generated with U=[I, f3 Y and v = i,
where f3 denotes the vector of regression coefficients of the linear model y = f31 x, and I
the identity matrix. After calculating Q and U by EM in joint data space as formulated in
(2), an estimate of f3 can be derived from the conditional probability p(y I x). As all
distributions are assumed to be normal, the expected value ofy is the mean of this conditional distribution. The locally weighted version (L WF A) of f3 can be obtained together
with an estimate of the factors v from the joint weighted covariance matrix 'I' of z and v:
E{[:]
+[}
~ ~ ~,,~,;'x,
Q+UU T
=
UT
[
where
~ ~ [ZT, VT~~Jft: w; ~
(3)
U] ['I'II(=n x n)
'I'12(=nX(m+k?)]
I = '?21(= (m + k) x n) '1'22(= (m + k) x (m + k?)
where E { .} denotes the expectation operator and B a matrix of coefficients involved in
estimating the factors v. Note that unless the noise f: is zero, the estimated f3 is different
from the true f3 as it tries to average out the noise in the data.
2.2 JOINT-SPACE PRINCIPAL COMPONENT ANALYSIS (LWPCA)
An alternative way of determining the parameters f3 in a reduced space employs locally
weighted principal component analysis (LWPCA) in the joint data space. By defining the .
largest k+ 1 principal components of the weighted covariance matrix ofZ as U:
U= [eigenvectors(I Wi (Zi - ZXZi - Z)T II Wi)]
(4)
max(l :k+1l
and noting that the eigenvectors in U are unit length, the matrix inversion theorem (Hom
& Johnson, 1994) provides a means to derive an efficient estimate of f3
T
T(
T
f3=U x(Uy -Uy UyUy -I
)-1 UyU yt
T\
[Ux(=nXk)]
where U= Uy(=mxk)
(5)
In our one dimensional output case, U y is just a (1 x k) -dimensional row vector and the
evaluation of (5) does not require a matrix inversion anymore but rather a division.
If one assumes normal distributions in all variables as in LWF A, LWPCA is the special
case of L WF A where the noise covariance Q is spherical, i.e., the same magnitude of
noise in all observables. Under these circumstances, the subspaces spanned by U in both
methods will be the same. However, the regression coefficients of LWPCA will be different from those of LWF A unless the noise level is zero, as LWFA optimizes the coefficients according to the noise in the data (Equation (3? . Thus, for normal distributions
and a correct guess of k, LWPCA is always expected to perform worse than LWFA.
S. Schaal, S. Vijayakumar and C. G. Atkeson
636
2.3 PARTIAL LEAST SQUARES (LWPLS, LWPLS_I)
Partial least squares (Wold, 1975; Frank & Friedman, 1993) recursively computes orthogonal projections of the input data and performs single variable regressions along
these projections on the residuals of the previous iteration step. A locally weighted version of partial least squares (LWPLS) proceeds as shown in Equation (6) below.
As all single variable regressions are ordinary uniFor Training:
For Lookup:
variate least-squares minim izations,
L WPLS
Initialize:
Initialize:
makes the same statistical assumption as ordinary
Do = X, eo = y do = x, y=
linear regressions, i.e., that only output variables
have additive noise, but input variables are noiseFor i = 1 to k:
For i = 1 to k:
less. The choice of the projections u, however, ins. = dT.u.
troduces an element in LWPLS that remains statistically still debated (Frank & Friedman, 1993), although, interestingly, there exists a strong similarity with the way projections are chosen in Cascade
Correlation (Fahlman & Lebiere, 1990). A peculiarity of LWPLS is that it also regresses the inputs
of the previous step against the projected inputs s
in order to ensure the orthogonality of all the projections u. Since LWPLS chooses projections in a
(6)
very powerful way, it can accomplish optimal
function fits with only one single projections (i.e.,
k= 1) for certain input distributions. We will address this issue in our empirical evaluations by comparing k-step LWPLS with I-step LWPLS, abbreviated LWPLS_I.
?
I
1-
I
2.4 PRINCIPAL COMPONENT REGRESSION (L WPCR)
Although not optimal, a computationally efficient techniques of dimensionality reduction
for linear regression is principal component regression (LWPCR) (Massy, 1965). The inputs are projected onto the largest k principal components of the weighted covariance
matrix of the input data by the matrix U:
U = [eig envectors(2: Wi (Xi -
xX xt /2:
Xi -
(7)
Wi )]
max(l:k)
The regression coefficients f3 are thus calculated as:
f3
= (UTXTwxUtUTXTWy
(8)
Equation (8) is inexpensive to evaluate since after projecting X with U, UTXTWXU becomes a diagonal matrix that is easy to invert. LWPCR assumes that the inputs have additive spherical noise, which includes the zero noise case. As during dimensionality reduction LWPCR does not take into account the output data, it is endangered by clipping
input dimensions with low variance which nevertheless have important contribution to
the regression output. However, from a statistical point of view, it is less likely that low
variance inputs have significant contribution in a linear regression, as the confidence
bands of the regression coefficients increase inversely proportionally with the variance of
the associated input. If the input data has non-spherical noise, L WPCR is prone to focus
the regression on irrelevant projections.
3 MONTE CARLO EVALUATIONS
In order to evaluate the candidate methods, data sets with 5 inputs and 1 output were randomly generated. Each data set consisted of 2,000 training points and 10,000 test points,
distributed either uniformly or nonuniformly in the unit hypercube. The outputs were
Local Dimensionality Reduction
637
generated by either a linear or quadratic function. Afterwards, the 5-dimensional input
space was projected into a to-dimensional space by a randomly chosen distance preserving linear transformation. Finally, Gaussian noise of various magnitudes was added
to both the 10-dimensional inputs and one dimensional output. For the test sets, the additive noise in the outputs was omitted. Each regression technique was localized by a
Gaussian kernel (Equation (1)) with a to-dimensional distance metric D=IO*I (D was
manually chosen to ensure that the Gaussian kernel had sufficiently many data points and
no "data holes" in the fringe areas of the kernel) . The precise experimental conditions
followed closely those suggested by Frank and Friedman (1993):
{g.I for:
131.. = [I, I, I, I, If ,
?
2 kinds of linear functions y =
?
2 kinds of quadratic functions y = f3J.I + f3::.aAxt ,xi
i)
?
1311.
i)
= [I, I, I, I, Wand f3q.ad = 0.1 [I, I, I, I, If, and ii)
ii)
I3Ii. = [1,2,3,4, sf
,xi ,X;,X;]T for:
131.. = [1,2,3,4, sf
and f3q uad = 0.1 [I, 4, 9, 16, 2sf
3 kinds of noise conditions, each with 2 sub-conditions:
local signal/noise ratio Isnr=20,
i) only output noise:
a) low noise:
and
b) high noise:
Isnr=2,
ii) equal noise in inputs and outputs:
a) low noise Ex ?? = Sy = N(O,O.Ot2), n e[I,2, ... ,10],
and
b) high noise Ex ?? =sy=N(0,0.1 2),ne[I,2, ... ,10],
iii) unequal noise in inputs and outputs:
a) low noise : Ex .? = N(0,(0.0In)2), n e[I,2, . .. ,1O] and Isnr=20,
and
?
b) high noise:
Ex .?
= N(0,(0.0In)2), n e[I,2, ... ,1O] and Isnr=2,
2 kinds of input distributions: i) uniform in unit hyper cube, ii) uniform in unit hyper cube excluding data
points which activate a Gaussian weighting function (I) at c = [O.S,O,o,o,of with D=IO*I more than
w=0.2 (this forms a "hyper kidney" shaped distribution)
Every algorithm was run * 30 times on each of the 48 combinations of the conditions.
Additionally, the complete test was repeated for three further conditions varying the dimensionality--called factors in accordance with LWFA-that the algorithms assumed to
be the true dimensionality of the to-dimensional data from k=4 to 6, i.e., too few, correct,
and too many factors. The average results are summarized in Figure I.
Figure I a,b,c show the summary results of the three factor conditions. Besides averaging
over the 30 trials per condition, each mean of these charts also averages over the two input distribution conditions and the linear and quadratic function condition, as these four
cases are frequently observed violations of the statistical assumptions in nonlinear function approximation with locally linear models. In Figure I b the number of factors equals
the underlying dimensionality of the problem, and all algorithms are essentially performing equally well. For perfectly Gaussian distributions in all random variables (not
shown separately), LWFA's assumptions are perfectly fulfilled and it achieves the best
results, however, almost indistinguishable closely followed by LWPLS. For the ''unequal
noise condition", the two PCA based techniques, LWPCA and LWPCR, perform the
worst since--as expected-they choose suboptimal projections. However, when violating the statistical assumptions, L WF A loses parts of its advantages, such that the summary results become fairly balanced in Figure lb.
The quality of function fitting changes significantly when violating the correct number of
factors, as illustrated in Figure I a,c. For too few factors (Figure la), LWPCR performs
worst because it randomly omits one of the principle components in the input data, without respect to how important it is for the regression. The second worse is LWF A: according to its assumptions it believes that the signal it cannot model must be noise, leading to a degraded estimate of the data's subspace and, consequently, degraded regression
results. LWPLS has a clear lead in this test, closely followed by LWPCA and LWPLS_I.
* Except for LWFA, all methods can evaluate a data set in non-iterative calculations. LWFA was trained with EM for maximally 1000 iterations or until the log-likelihood increased less than I.e-lOin one iteration.
S. Schaal, S. Vljayakumar and C. G. Atkeson
638
For too many factors than necessary (Figure Ie), it is now LWPCA which degrades. This
effect is due to its extracting one very noise contaminated projection which strongly influences the recovery of the regression parameters in Equation (4). All other algorithms
perform almost equally well, with LWF A and LWPLS taking a small lead.
Equal NoIse In ell
In puIS end OutpUIS
OnlyOutpul
Noise
Unequel NoIse In ell
Inputs end OutpulS
0.1
c
o
~
::::;;
0.01
c
II>
C>
~ 0.001
~
0.0001
fl- I. E>O
~I. ? >>(I
~ J. &>O
~J , E ? O
~ J .E>O
fl- I.&? O ~I . & >O
~ I . & > >O
~I.& >O
~ I .?>>o
p,. 1. s>O
tJ-J .?>>O
e) RegressIon Results with 4 Factors
?
LWFA
?
LWPCA
?
LWPCR
0
LWPLS
?
LWPLS_1
0.1
c::
o
W
~c::
0.01
II>
C)
~ 0.001
~
0.0001
~
~
0.1
W
0.01
8
~c::
g,
~
0.001
~
0.0001
c) RegressIon Results with 6 Feclors
jj
il
0.1
f-
a
~
::::;;
0.01
c
II>
C)
~ 0.001
~
0.0001
d) Summery Results
Figure I: Average summary results of Monte Carlo experiments. Each chart is primarily
divided into the three major noise conditions, cf. headers in chart (a). In each noise condition, there are four further subdivision: i) coefficients of linear or quadratic model are
equal with low added noise; ii) like i) with high added noise; iii) coefficients oflinear or
quadratic model are different with low noise added; iv) like iii) with high added noise.
Refer to text and descriptions of Monte Carlo studies for further explanations.
Local Dimensionality Reduction
639
4 SUMMARY AND CONCLUSIONS
Figure 1d summarizes all the Monte Carlo experiments in a final average plot. Except for
LWPLS, every other technique showed at least one clear weakness in one of our "robustness" tests. It was particularly an incorrect number of factors which made these weaknesses apparent. For high-dimensional regression problems, the local dimensionality, i.e.,
the number of factors, is not a clearly defined number but rather a varying quantity, depending on the way the generating process operates. Usually, this process does not need
to generate locally low dimensional distributions, however, it often "chooses" to do so,
for instance, as human ann movements follow stereotypic patterns despite they could
generate arbitrary ones. Thus, local dimensionality reduction needs to find autonomously
the appropriate number of local factor. Locally weighted partial least squares turned out
to be a surprisingly robust technique for this purpose, even outperforming the statistically
appealing probabilistic factor analysis. As in principal component analysis, LWPLS's
number of factors can easily be controlled just based on a variance-cutoff threshold in input space (Frank & Friedman, 1993), while factor analysis usually requires expensive
cross-validation techniques. Simple, variance-based control over the number of factors
can actually improve the results of LWPCA and LWPCR in practice, since, as shown in
Figure I a, LWPCR is more robust towards overestimating the number of factors, while
L WPCA is more robust towards an underestimation. If one is interested in dynamically
growing the number of factors while obtaining already good regression results with too
few factors, L WPCA and, especially, L WPLS seem to be appropriate-it should be
noted how well one factor LWPLS (LWPLS_l) already performed in Figure I!
In conclusion, since locally weighted partial least squares was equally robust as local
weighted factor analysis towards additive noise in. both input and output data, and,
moreover, superior when mis-guessing the number of factors, it seems to be a most favorable technique for local dimensionality reduction for high dimensional regressions.
Acknowledgments
The authors are grateful to Geoffrey Hinton for reminding them of partial least squares. This work was supported by the ATR Human Information Processing Research Laboratories. S. Schaal's support includes the
German Research Association, the Alexander von Humboldt Foundation, and the German Scholarship Foundation. S. Vijayakumar was supported by the Japanese Ministry of Education, Science, and Culture (Monbusho).
C. G. Atkeson acknowledges the Air Force Office of Scientific Research grant F49-6209410362 and a National
Science Foundation Presidential Young Investigators Award.
References
Atkeson, C. G., Moore, A. W., & Schaal, S, (1997a).
"Locally weighted learning." ArtifiCial Intelligence Review, 11, 1-5, pp.II-73.
Atkeson, C. G., Moore, A. W., & Schaal, S, (1997c).
"Locally weighted learning for control." ArtifiCial Intelligence Review, 11, 1-5, pp.75-113.
Belsley, D. A., Kuh, E., & Welsch, R. E, (1980). Regression diagnostics: Identifying influential data and
sources of collinearity. New York: Wiley.
Everitt, B. S, (1984). An introduction to latent variable
models. London: Chapman and Hall.
Fahlman, S. E. ,Lebiere, C, (1990). "The cascadecorrelation learning architecture." In: Touretzky, D. S.
(Ed.), Advances in Neural Information Processing
Systems II, pp.524-532. Morgan Kaufmann.
Frank, I. E., & Friedman, 1. H, (1993). "A statistical
view of some chemometric regression tools." Technometrics, 35, 2, pp.l09-135.
Geman, S., Bienenstock, E., & Doursat, R. (1992).
"Neural networks and the bias/variance dilemma."
Neural Computation, 4, pp.I-58.
Hom, R. A., & Johnson, C. R, (1994). Matrix analySis.
Press Syndicate of the University of Cambridge.
Jordan, M.I., & Jacobs, R, (1994). "Hierarchical mix-
tures of experts and the EM algorithm." Neural Computation, 6, 2, pp.181-214.
Massy, W. F, (1965). "Principle component regression
in exploratory statistical research." Journal ofthe
American Statistical Association, 60, pp.234-246.
Rubin, D. B., & Thayer, D. T, (l982). "EM algorithms
for ML factor analysis." Psychometrika, 47, I, 69-76.
Schaal, S., & Atkeson, C. G, (in press). "Constructive
incremental learning from only local information."
Neural Computation.
Scott, D. W, (1992). Multivariate Density Estimation.
New York: Wiley.
Vijayakumar, S., & Schaal, S, (1997). "Local dimensionality reduction for locally weighted learning." In:
International Conference on Computational Intelligence in Robotics and Automation, pp.220-225, Monteray, CA, July 10-11, 1997.
Wold, H. (1975). "Soft modeling by latent variables:
the nonlinear iterative partial least squares approach."
In: Gani, J. (Ed.), Perspectives in Probability and Statistics, Papers in Honour ofM S. Bartlett. Aca<j. Press.
Xu, L., Jordan, M.l., & Hinton, G. E, (1995). "An alternative model for mixture of experts." In: Tesauro,
G., Touretzky, D. S., & Leen, T. K. (Eds.), Advances in
Neural Information Processing Systems 7, pp.633-640.
Cambridge, MA: MIT Press.
| 1387 |@word trial:1 collinearity:1 briefly:1 version:3 inversion:2 advantageous:1 seems:1 simulation:1 jacob:2 covariance:5 recursively:1 reduction:15 interestingly:1 atlantic:1 comparing:1 surprising:1 must:1 additive:4 shape:1 plot:1 intelligence:3 guess:2 nervous:1 device:1 yr:1 provides:1 sigmoidal:1 along:1 wpls:2 become:1 incorrect:2 fitting:1 theoretically:1 expected:3 seika:1 examine:1 frequently:1 brain:1 growing:1 globally:3 spherical:3 curse:1 becomes:1 project:1 estimating:2 xx:1 underlying:1 moreover:1 psychometrika:1 what:1 kind:4 massy:2 transformation:2 perfonn:1 every:2 ensured:1 control:5 normally:1 unit:4 grant:1 before:1 positive:1 local:25 accordance:1 io:2 despite:1 dynamically:1 statistically:2 uy:3 acknowledgment:1 practice:1 definite:1 isnr:4 area:1 empirical:1 significantly:1 cascade:1 projection:10 confidence:1 onto:2 ga:1 selection:1 operator:1 cannot:1 context:2 influence:1 www:2 yt:2 independently:1 kidney:1 simplicity:1 recovery:1 identifying:1 spanned:1 exploratory:1 coordinate:1 mxk:1 origin:1 element:2 expensive:1 particularly:1 geman:1 observed:2 worst:2 autonomously:1 movement:1 balanced:1 dynamic:2 trained:1 grateful:1 dilemma:2 division:1 observables:2 translated:3 easily:1 joint:7 l982:1 various:2 univ:1 fac:1 fast:1 activate:1 monte:5 london:1 query:2 artificial:2 hyper:3 outcome:1 neighborhood:5 header:1 apparent:1 presidential:1 statistic:2 final:1 advantage:2 turned:1 translate:1 meguro:1 description:1 los:1 convergence:1 empty:2 wpca:3 generating:3 incremental:2 perfect:1 derive:4 depending:1 ac:2 school:1 strong:2 c:2 uu:1 closely:3 tokyo:2 correct:3 peculiarity:1 human:2 education:1 humboldt:1 require:2 endangered:1 reminding:1 biological:2 around:1 considered:1 sufficiently:1 normal:3 exp:1 hall:1 mapping:1 slab:1 sschaal:1 major:1 achieves:1 omitted:1 purpose:1 estimation:2 favorable:1 largest:2 tool:1 weighted:20 stefan:1 mit:1 clearly:1 gaussian:10 always:1 rather:2 avoid:1 varying:2 gatech:2 office:1 derived:1 focus:4 schaal:13 likelihood:2 wf:4 bienenstock:1 relation:2 nxk:1 transformed:1 interested:1 issue:1 denoted:1 special:1 initialize:2 fairly:1 cube:2 field:1 equal:4 f3:12 shaped:1 ell:2 manually:2 chapman:1 contaminated:2 overestimating:1 employ:2 few:3 primarily:1 randomly:3 national:1 summery:1 usc:2 friedman:5 technometrics:1 atlanta:1 investigate:1 evaluation:3 weakness:2 violation:2 mixture:1 diagnostics:1 tj:1 partial:10 necessary:2 culture:1 orthogonal:1 unless:2 iv:1 theoretical:1 increased:1 instance:1 soft:1 modeling:1 maximization:1 ordinary:2 clipping:1 uniform:2 successful:1 johnson:2 too:5 accomplish:1 synthetic:1 cho:1 chooses:2 density:1 international:1 ie:1 vijayakumar:6 probabilistic:1 factoranalysis:1 together:1 von:1 choose:1 thayer:2 worse:2 expert:2 american:1 leading:1 account:1 lookup:1 summarized:2 includes:2 coefficient:8 automation:1 caused:1 chemometric:1 ad:1 performed:1 try:1 view:2 aca:1 contribution:2 air:1 square:11 chart:3 degraded:2 variance:7 characteristic:1 kaufmann:1 sy:2 il:1 ofthe:1 produced:1 carlo:5 comp:1 cc:2 drive:1 anns:2 suffers:1 touretzky:2 ed:3 against:1 inexpensive:1 sensorimotor:1 pp:9 involved:1 regress:1 lebiere:2 associated:1 mi:1 knowledge:1 ut:1 dimensionality:21 ubiquitous:1 f3j:1 loin:1 actually:1 dt:1 violating:2 follow:1 maximally:1 leen:1 wold:2 strongly:1 generality:1 just:2 stage:1 correlation:1 until:1 hand:1 christopher:1 nonlinear:4 eig:1 widespread:1 quality:1 scientific:1 effect:1 consisted:1 true:3 spatially:1 moore:4 iteratively:1 laboratory:1 syndicate:1 illustrated:1 indistinguishable:1 during:1 noted:1 outline:1 complete:1 performs:2 common:1 superior:1 kawato:1 functional:1 empirically:1 jp:2 association:2 measurement:1 significant:1 lwpca:10 refer:1 cambridge:2 ai:2 everitt:2 wpcr:3 uv:1 unifor:1 had:1 robot:1 similarity:1 base:1 add:1 multivariate:1 showed:1 perspective:1 optimizes:1 irrelevant:1 tesauro:1 certain:1 outperforming:2 vt:1 preserving:1 ministry:1 morgan:1 ey:1 eo:1 surely:1 signal:2 ii:13 semi:1 violate:1 ofy:1 afterwards:1 kyoto:1 reduces:1 mix:1 july:1 adapt:1 calculation:1 offer:2 cross:1 jft:1 divided:1 equally:4 award:1 controlled:1 regression:41 circumstance:1 titech:2 metric:2 expectation:2 essentially:1 iteration:3 kernel:5 lwpls_i:3 invert:1 robotics:1 whereas:1 separately:1 source:1 doursat:1 south:1 seem:1 jordan:4 extracting:1 noting:1 feedforward:1 iii:3 easy:1 variety:1 variate:1 zi:1 fit:1 architecture:1 perfectly:4 suboptimal:1 l09:1 angeles:1 xqf:1 pca:1 bartlett:1 soraku:1 york:2 jj:1 cga:1 involve:1 eigenvectors:2 proportionally:1 clear:2 nonparametric:1 locally:23 band:1 reduced:1 http:3 generate:2 neuroscience:1 fulfilled:2 wr:2 estimated:1 per:1 ist:1 four:2 nevertheless:1 threshold:1 cutoff:1 wand:1 run:1 inverse:1 powerful:1 izations:1 almost:3 ofm:1 summarizes:1 fl:2 hom:2 followed:3 quadratic:7 orthogonality:1 sake:1 performing:1 edul:1 department:1 influential:1 according:2 combination:1 cascadecorrelation:1 slightly:2 increasingly:1 em:5 wi:7 appealing:2 honour:1 projecting:1 interference:1 computationally:1 equation:5 remains:2 abbreviated:1 uyu:1 german:2 end:2 pursuit:1 prerequisite:1 yare:1 hierarchical:1 appropriate:3 robustly:1 anymore:1 alternative:2 robustness:4 denotes:3 assumes:4 ensure:3 cf:1 calculating:1 exploit:1 scholarship:1 especially:1 hikaridai:1 hypercube:1 question:1 added:5 quantity:1 already:2 parametric:1 degrades:1 nonuniformly:1 diagonal:3 guessing:1 subspace:2 distance:3 ot2:1 atr:1 nx:1 chris:1 gun:1 sethu:3 originate:1 evaluate:5 assuming:1 length:1 besides:1 ratio:1 nonintuitive:1 frank:5 negative:1 zt:1 minim:1 perform:5 uad:1 defining:1 hinton:3 excluding:1 precise:1 lb:1 arbitrary:1 california:1 unequal:2 omits:1 address:1 able:1 suggested:1 proceeds:1 usually:3 below:2 scott:2 pattern:1 including:1 max:2 explanation:1 belief:1 force:1 residual:1 lwf:4 improve:1 technology:2 imply:1 inversely:1 ne:1 acknowledges:1 xq:2 text:1 prior:1 review:2 xtt:1 contributing:1 determining:1 loss:1 tures:1 geoffrey:1 localized:2 outlining:1 validation:1 foundation:4 rubin:2 principle:4 row:2 prone:1 hopeless:1 summary:4 surprisingly:1 fahlman:2 supported:2 bias:2 institute:2 face:1 taking:1 lwfa:8 tolerance:1 distributed:3 f31:1 dimension:4 calculated:1 computes:1 author:1 made:1 projected:3 atkeson:14 observable:1 kuh:1 ml:1 global:2 assumed:2 xi:7 continuous:1 iterative:2 latent:2 additionally:1 ku:1 robust:4 ca:2 obtaining:1 warranted:1 interpolating:1 necessarily:1 japanese:1 gani:1 noise:39 f3q:2 repeated:1 xu:3 georgia:1 oflinear:1 wiley:2 sub:1 sf:3 debated:1 candidate:4 weighting:3 young:1 theorem:1 xt:2 exists:1 lwpcr:8 magnitude:3 hole:1 locality:1 welsch:1 likely:1 ux:1 lwpls:15 loses:1 ma:1 conditional:2 fringe:1 viewed:1 identity:1 formulated:1 ann:2 consequently:1 towards:4 considerable:1 change:1 determined:2 except:2 uniformly:1 operates:1 averaging:1 principal:9 belsley:1 called:1 stereotypic:1 secondary:1 experimental:1 hnb:1 la:1 subdivision:1 underestimation:1 college:1 support:1 alexander:1 investigator:1 constructive:1 dept:1 ex:5 |
425 | 1,388 | On Efficient Heuristic Ranking of
Hypotheses
Steve Chien, Andre Stechert, and Darren Mutz
Jet Propulsion Laboratory, California Institute of Technology
4800 Oak Grove Drive, MIS 525-3660, Pasadena, CA 91109-8099
steve.chien@jpl.nasa.gov, Voice: (818) 306-6144 FAX: (818) 306-6912
Content Areas: Applications (Stochastic Optimization),Model Selection Algorithms
Abstract
This paper considers the problem of learning the ranking of a set
of alternatives based upon incomplete information (e.g., a limited
number of observations). We describe two algorithms for hypothesis ranking and their application for probably approximately correct (PAC) and expected loss (EL) learning criteria. Empirical
results are provided to demonstrate the effectiveness of these ranking procedures on both synthetic datasets and real-world data from
a spacecraft design optimization problem.
1
INTRODUCTION
In many learning applications, the cost of information can be quite high, imposing
a requirement that the learning algorithms glean as much usable information as
possible with a minimum of data. For example:
? In speedup learning, the expense of processing each training example can
be significant [Tadepalli921.
? In decision tree learning, ihe cost of using all available training examples
when evaluating potential attributes for partitioning can be computationally ex.pensive [Musick93].
? In evaruating medical treatment policies, additional training examples imly suboptimal treatment of human subjects.
? n data-poor applications, training data may be very scarce and learning
as well as possible from limited data may be key.
E
This paper provides a statistical decision-theoretic framework for the ranking of
parametric distributions. This framework will provide the answers to a wide range
of questions about algorithms such as: how much information is enough? At what
point do we have adequate information to rank the alternatives with some requested
confidence?
445
On Efficient Heuristic Ranking of Hypotheses
The remainder of this paper is structured as follows. First, we describe the hypothesis ranking problem more formally, including definitions for the probably approximately correct (PAC) and expected loss (EL) decision criteria. We then define two
algorithms for establishing these criteria for the hypothesis ranking problem - a recursive hypothesis selection algorithm and an adjacency based algorithm. Next, we
describe empirical tests demonstrating the effectiveness of these algorithms as well
as documenting their improved performance over a standard algorithm from the statistical ranking literature. Finally, we describe related work and future extensions
to the algorithms.
2
HYPOTHESIS RANKING PROBLEMS
Hypothesis ranking problems, an extension of hypothesis selection problems, are an
abstract class of learning problems where an algorithm is given a set of hypotheses
to rank according to expected utility over some unknown distribution, where the
expected utility must be estimated from training data.
In many of these applications, a system chooses a single alternative and never revisits the decision. However, some systems require the ability to investigate several
options (either serially or in parallel), such as in beam search or iterative broadening, where the ranking formulation is most appropriate. Also, as is the case
with evolutionary approaches, a system may need to populate future alternative
hypotheses on the basis of the ranking of the current population[Goldberg89] .
In any hypothesis evaluation problem, always achieving a correct ranking is impossible in practice, because the actual underlying probability distributions are
unavailable and there is always a (perhaps vanishingly) small chance that the algorithms will be unlucky because only a finite number of samples can be taken.
Consequently, rather than always requiring an algorithm to output a correct ranking, we impose probabilistic criteria on the rankings to be produced. While several
families of such requirements exist, in this paper we examine two, the probably
approximately correct (PAC) requirement from the computational learning theory
community [Valiant84] and the expected loss (EL) requirement frequently used in
decision theory and gaming problems [Russe1l92] .
The expected utility of a hypothesis can be estimated by observing its values over a
finite set of training examples. However, to satisfy the PAC and EL requirements,
an algorithm must also be able to reason about the potential difference between
the estimated and true utilities of each hypotheses. Let Ui be the true expected
utility of hypothesis i and let Ui be the estimated expected utility of hypothesis i.
Without loss of generality, let us presume that the proposed ranking of hypotheses
is U1 > U2 >, ... , > Uk-I> Uk. The PAC requirement states that for some userspecified ?. with probability 1 - 8:
k-l
/\ [(Ui
+ f) > MAX(Ui+I, ... ,UIe)]
(1)
;=1
Correspondingly, let the loss L of selecting a hypothesis HI to be the best from a
set of k hypotheses HI, ... , Hk be as follows .
L(HI' {HI, ... ,HIe}) = MAX(O, MAX(U2 , ... ,UIe) - UI)
(2)
and let the loss RL of a ranking H 1, ... , H k be as follows.
Ie-I
RL(Hl, ... , Hie) =
L L(Hi, {Hi+l, .. ., Hie})
(3)
i=1
A hypothesis ranking algorithm which obeys the expected loss requirement must
produce rankings that on average have less than the requested expected loss bound.
S. Chien, A. Stechert and D. Mutz
446
Consider ranking the hypotheses with expected utilities: U1 = 1.0, U2 = 0.95, U3 =
0.86. The ranking U2 > U1 > U3 is a valid PAC ranking for { = 0.06 but not for
{ = 0.01 and has an observed loss of 0.05 + 0 = 0.05.
However, while the confidence in a pairwise comparison between two hypotheses is
well understood, it is less clear how to ensure that desired confidence is met in the
set of comparisons required for a selection or the more complex set of comparisons
required for a ranking. Equation 4 defines the confidence that Ui + { > Uj, when
the distribution underlying the utilities is normally distributed with unknown and
unequal variances.
(4)
where ? represents the cumulative standard normal distribution function, and n,
Ui-j, and Si-j are the size, sample mean, and sample standard deviation of the
blocked differential distribution, respectively 1 .
Likewise, computation of the expected loss for asserting an ordering between a pair
of hypotheses is well understood, but the estimation of expected loss for an entire
ranking is less clear. Equation 5 defines the expected loss for drawing the conclusion
Ui > Uj, again under the assumption ~fnormality (see [Chien95] for further details).
EL(Ui
> Ujl
~
'e
U'-i
-O . 6 n
( ) :l
Si_j
= .-]
';21rn
-j
fJ
+~
.,j2;
oc
_ \~irn
e- O ? 6 ?
:l
dz
(5)
'-J
In the next two subsections, we describe two interpretations for estimating the likelihood that an overall ranking satisfies the PAC or EL requirements by estimating
and combining pairwise PAC errors or EL estimates. Each of these interpretations
lends itself directly to an algorithmic implementation as described below.
2.1 RANKING AS RECURSIVE SELECTION
One way to determine a ranking HI, ... , Hk is to view ranking as recursive selection
from the set of remaining candidate hypotheses. In this view, the overall ranking
error, as specified by the desired confidence in PAC algorithms and the loss threshhold in EL algorithms, is first distributed among k - 1 selection errors which are
then further subdivided into pairwise comparison errors. Data is then sampled until the estimates of the pairwise comparison error (as dictated by equation 4 or 5)
satisfy the bounds set by the algorithm.
Thus, another degree of freedom in the design of recursive ranking algorithms is
the method by which the overall ranking error is ultimately distributed among
individual pairwise comparisons between hypotheses. Two factors influence the
way in which we compute error distribution. First, our model of error combination
determines how the error allocated for individual comparisons or selections combines
into overall ranking error and thus how many candidates are available as targets
for the distribution. Using Bonferroni's inequality, one combine errors additively,
but a more conservative approach might be to assert that because the predicted
"best" hypothesis may change during sampling in the worst case the conclusion
might depend on all possible pairwise comparisons and thus the error should be
distributed among all (~) pairs of hypotheses 2 ).
INote that in our approach we block examples to further reduce sampling complexity.
Blocking forms estimates by using the difference in utility between competing hypotheses
on each observed example. Blocking can significantly reduce the variance in the data when
the hypotheses are not independent. It is trivial to modify the formulas to address the
cases in which it is not possible to block data (see [Moore94, Chien95] for further details).
2For a discussion of this issue, see pp. 18-20 of [Gratch93].
447
On Efficient Heuristic Ranking of Hypotheses
Second, our policy with respect to allocation of error among the candidate comparisons or selections determines how samples will be distributed. For example, in
some contexts, the consequences of early selections far outweigh those of later selections. For these scenarios, we have implemented ranking algorithms that divide
overall ranking error unequally in favor of earlier selections3 . Also, it is possible to
divide selection error into pairwise error unequally based on estimates of hypothesis
parameters in order to reduce sampling cost (for example, [Gratch94] allocates error
rationally) .
Within the scope of this paper, we only consider algorithms that: (1) combine
pairwise error into selection error additively, (2) combine selection error into overall
ranking error additively and (3) allocate error equally at each level.
One disadvantage of recursive selection is that once a hypothesis has been selected,
it is removed from the pool of candidate hypotheses. This causes problems in rare
instances when, while sampling to increase the confidence of some later selection,
the estimate for a hypothesis' mean changes enough that some previously selected
hypothesis no longer dominates it. In this case, the algorithm is restarted taking
into account the data sampled so far.
These assumptions result in the following formulations (where d(U11>? {U2' ... , Uk})
is used to denote the error due to the action of selecting hypothesis 1 under Equation
1 from the set {HI, ... , Hk} and d(UII>{U2, ... , Uk}) denotes the error due to selection
loss in situations where Equation 2 applies):
t5 rec (UI > U2 > ... > Uk) =
t5 rec (U2 > U3 > ... > Uk)
+t5(UI t>. {U2 , ??? ,Uk})
(6)
where drec(Uk) = 0 (the base case for the recursion) and the selection error is as
defined in [Chien95]:
k
t5(Ul t>. {U2 , ??? ,Uk}) =
L 15
1 ,.
(7)
.=2
using Equation 4 to compute pairwise confidence.
Algorithmically, we implement this by:
1. sampling a default number of times to seed the estimates for each hypothesis
mean and variance,
2. allocating the error to selection and pairwise comparisons as indicated
above,
3. sampling until the desired confidences for successive selections is met, and
4. restarting the algorithm if any of the hypotheses means changed significantly enough to change the overall ranking.
An analogous recursive selection algorithm based on expected loss is defined as
follows.
EL rec (U2 > U3 > ... > Uk)
(8)
+EL(U1 t> {U2 , ??? ,Uk})
where ELrec(Uk) = 0 and the selection EL is as defined in [Chien95]:
k
EL(U1 I> {U2, ... , Uk})
=L
i=2
3Space constraints preclude their description here.
EL(Ut, Ud
(9)
448
S. Chien, A. StecheT1 and D. Mutz
2.2 RANKING BY COMPARISON OF ADJACENT ELEMENTS
Another interpretation of ranking confidence (or loss) is that only adjacent elements
in the ranking need be compared. In this case, the overall ranking error is divided
directly into k -1 pairwise comparison errors. This leads to the following confidence
equation for the PAC criteria:
k-l
(10)
dac(i(Ul > U2 > ... > Uk) = Ldi,i+1
i=l
And the following equation for the EL criteria.k_l
ELac(i(Ul
> U2 > ... > Uk)
= '2: EL (Ui,Ui+d
(11)
i=l
Because ranking by comparison of adjacent hypotheses does not establish the dominance between non-adjacent hypotheses (where the hypotheses are ordered byobserved mean utility), it has the advantage of requiring fewer comparisons than
recursive selection (and thus may require fewer samples than recursive selection).
However, for the same reason, adjacency algorithms may be less likely to correctly
bound probability of correct selection (or average loss) than the recursive selection
algorithms. In the case of the PAC algorithms, this is because f-dominance is not
necessarily transitive. In the case of the EL algorithms, it is because expected loss is
not additive when considering two hypothesis relations sharing a common hypothesis. For instance, the size of the blocked differential distribution may be different
for each of the pairs of hypotheses being compared.
2.3 OTHER RELEVANT APPROACHES
Most standard statistical ranking/selection approaches make strong assumptions
about the form of the problem (e.g., the variances associated with underlying utility
distribution of the hypotheses might be assumed known and equal). Among these,
Turnbull and Weiss [Turnbull84] is most comparable to our PAC-based approach4.
Turnbull and Weiss treat hypotheses as normal random variables with unknown
mean and unknown and unequal variance. However, they make the additional
stipulation that hypotheses are independent. So, while it is still reasonable to
use this approach when the candidate hypotheses are not independent, excessive
statistical error or unnecessarily large training set sizes may result.
3
EMPIRICAL PERFORMANCE EVALUATION
We now turn to empirical evaluation of the hypothesis ranking techniques on realworld datasets. This evaluation serves three purposes. First, it demonstrates that
the techniques perform as predicted (in terms of bounding the probability of incorrect selection or expected loss). Second, it validates the performance of the techniques as compared to standard algorithms from the statistical literature. Third,
the evaluation demonstrates the robustness of the new approaches to real-world
hypothesis ranking problems.
An experimental trial consists of solving a hypothesis ranking problem with a given
technique and a given set of problem and control parameters. We measure performance by (1) how well the algorithms satisfy their respective criteria; and (2) the
number of samples taken. Since the performance of these statistical algorithms on
any single trial provides little information about their overall behavior, each trial
is repeated multiple times and the results are averaged across 100 trials. Because
4 PAC-based approaches have been investigated extensively in the statistical ranking and
selection literature under the topic of confidence interval based algorithms (see [Haseeb85]
for a review of the recent literature).
On Efficient Heuristic Ranking of Hypotheses
449
Table 1: Estimated expected total number of observations to rank DS-2 spacecraft
. sown
h 'm parenthesis.
designs. Achieved pro b a bTt
I I yof correc t ran k'mg IS
k
10
10
10
'Y
!!..
0.75
0 .90
0.95
2
2
2
TURNtlULL
534 {0 .96
667 (0 .98
793 (0.99
PAC rec
144 1.00
160 1.00
177 1.00
PACod '
92 0.98
98 1.00
103 0.99
Table 2: Estimated expected total number of observations and expected loss of an
incorrect ranking of DS-2 penetrator designs.
Parameters
k
~
10
0.10
0 .05
10
0 .02
10
EL rec
Samples
152
200
378
EL a d '
Loss
0.005
0 .003
0 .003
l:)amples
77
90
139
l..oss
0 .014
0 .006
0 .003
the PAC and expected loss criteria are not directly comparable, the approaches are
analyzed separately.
Experimental results from synthetic datasets are reported in [Chien97]. The evaluation of our approach on artificially generated data is used to show that: (1) the
techniques correctly bound probability of incorrect ranking and expected loss as predicted when the underlying assumptions are valid even when the underlying utility
distributions are inherently hard to rank , and (2) that the PAC techniques compare favorably to the algorithm of Thrnbull and Weiss in a wide variety of problem
configurations.
The test of real-world applicability is based on data drawn from an actual NASA
spacecraft design optimization application. This data provides a strong test of the
applicability of the techniques in that all of the statistical techniques make some
form of normality assumption - yet the data in this application is highly non-normal.
Tables 1 and 2 show the results of ranking 10 penetrator designs using the PACbased, Thrnbull, and expected loss algorithms In this problem the utility function
is the depth of penetration of the penetrator, with those cases in which the penetrator does not penetrate being assigned zero utility. As shown in Table 1, both
PAC algorithms significantly outperformed the Thrnbull algorithm, which is to be
expected because the hypotheses are somewhat correlated (via impact orientations
and soil densities). Table 2 shows that the EL rec expected loss algorithm effectively
bounded actual loss but the ELad,i algorithm was inconsistent.
4
DISCUSSION AND CONCLUSIONS
There are a number of areas of related work. First, there has been considerable
analysis of hypothesis selection problems. Selection problems have been formalized
using a Bayesian framework [Moore94, Rivest88] that does not require an initial
sample, but uses a rigorous encoding of prior knowledge. Howard [Howard70] also
details a Bayesian framework for analyzing learning cost for selection problems. If
one uses a hypothesis selection framework for ranking, allocation of pairwise errors
can be performed rationally [Gratch94]. Reinforcement learning work [Kaelbling93]
with immediate feedback can also be viewed as a hypothesis selection problem.
In su~mary, this paper has described the hypothesis ranking problem, an extension
to the hypothesis selection problem. We defined the application of two decision
criteria, probably approximately correct and expected loss, to this problem. We then
defined two families of algorithms, recursive selection and adjacency, for solution of
hypothesis ranking problems. Finally, we demonstrated the effectiveness of these
algorithms on both synthetic and real-world datasets, documenting improved performance over existing statistical approaches.
450
S. Chien, A. Stechert and D. Mutz
References
[Bechhofer54] R.E. Bechhofer, "A Single-sample Multiple Decision Procedure for Ranking
Means of Normal Populations with Known Variances," Annals of Math. Statistics (25)
1, 1954 pp. 16-39.
[Chien95] S. A. Chien, J . M. Gratch and M. C. Burl, "On the Efficient Allocation of
Resources for Hypothesis Evaluation: A Statistical Approach," IEEE Trans. Pattern
Analysis and Machine Intelligence 17 (7), July 1995, pp. 652-665.
[Chien97] S. Chien, A. Stechert, and D. Mutz, "Efficiently Ranking Hypotheses
in Machine Learning," JPL-D-14661, June 1997. Available online at http://wwwaig.jpl.nasa.gov/public/www/pas-bibliography.html
[Goldberg89] D. Goldberg, Genetic Algorithms in Search, Optimization and Machine
Learning, Add. Wes., 1989.
[Govind81] Z. Govindarajulu, "The Sequential Statistical Analysis," American Sciences
Press, Columbus, OH, 1981.
[Gratch92] J . Gratch and G. Dejong, "COMPOSER: A Probabilistic Solution to the Utility Problem in Speed-up Learning," Proc. AAAI92, San Jose, CA, July 1992, pp. 235240.
[Gratch93] J. Gratch, "COMPOSER: A Decision-theoretic Approach to Adaptive Problem
Solving," Tech. Rep. UIUCDCS-R-93-1806, Dept. Compo Sci., Univ. Illinois, May 1993.
[Gratch94] J. Gratch, S. Chien, and G. Dejong, "Improving Learning Performance
Through Rational Resource Allocation," Proc. AAAI94, Seattle, WA, August 1994,
pp. 576-582.
[Greiner92] R. Greiner and I. Jurisica, "A Statistical Approach to Solving the EBL Utility
Problem," Proc. AAAI92, San Jose, CA, July 1992, pp. 241-248.
[Haseeb85] R. M. Haseeb, Modern Statistical Selection, Columbus, OH: Am. Sciences
Press, 1985.
[Hogg78] R. V. Hogg and A. T. Craig, Introduction to Mathematical Statistics, Macmillan
Inc., London, 1978.
[Howard70] R. A. Howard, Decision Analysis: Perspectives on Inference, Decision, and
Experimentation," Proceedings of the IEEE 58, 5 (1970), pp. 823-834.
[Kaelbling93] L. P. Kaelbling, Learning in Embedded Systems, MIT Press, Cambridge,
MA,1993.
[Minton88] S. Minton, Learning Search Control Knowledge: An Explanation-Based Approach, Kluwer Academic Publishers, Norwell, MA, 1988.
[Moore94] A. W. Moore and M. S. Lee, "Efficient Algorithms for Minimizing Cross Validation Error," Proc. ML94, New Brunswick, MA, July 1994.
[Musick93] R. Musick, J. Catlett and S. Russell, "Decision Theoretic Subsampling for
Induction on Large Databases," Proc. ML93, Amhert, MA, June 1993, pp. 212-219.
[Rivest88] R. L. Rivest and R. Sloan, A New Model for Inductive Inference," Proc. 2nd
Conference on Theoretical Aspects of Reasoning about Knowledge, 1988.
[Russell92] S. Russell and E . Wefald, Do the Right Thing: Studies in Limited Rationality,
MIT Press, MA.
[Tadepalli92] P. Tadepalli, "A theory of unsupervised speedup learning," Proc. AAAI92"
pp. 229-234.
[Turnbull84] Turnbull and Weiss, "A class of sequential procedures for k-sample problems
concerning normal means with unknown unequal variances," in Design of Experiments:
ranking and selection, T. J. Santner and A. C. Tamhane (eds. ), Marcel Dekker, 1984.
[Valiant84] L. G. Valiant, "A Theory of the Learnable," Communications of the ACM 27,
(1984), pp. 1134-1142.
| 1388 |@word trial:4 nd:1 tadepalli:1 dekker:1 additively:3 u11:1 initial:1 configuration:1 selecting:2 genetic:1 existing:1 current:1 si:1 yet:1 must:3 additive:1 intelligence:1 selected:2 fewer:2 compo:1 provides:3 math:1 successive:1 oak:1 mathematical:1 differential:2 incorrect:3 consists:1 combine:4 pairwise:12 spacecraft:3 expected:26 os:1 behavior:1 examine:1 frequently:1 gov:2 actual:3 little:1 preclude:1 considering:1 provided:1 estimating:2 underlying:5 bounded:1 rivest:1 what:1 dejong:2 assert:1 demonstrates:2 uk:15 partitioning:1 normally:1 medical:1 control:2 understood:2 modify:1 treat:1 consequence:1 encoding:1 analyzing:1 establishing:1 approximately:4 might:3 limited:3 range:1 obeys:1 averaged:1 recursive:10 practice:1 block:2 implement:1 procedure:3 area:2 empirical:4 significantly:3 confidence:11 selection:38 context:1 impossible:1 influence:1 www:1 outweigh:1 demonstrated:1 dz:1 penetrate:1 formalized:1 oh:2 jurisica:1 population:2 analogous:1 annals:1 target:1 rationality:1 us:2 goldberg:1 hypothesis:61 pa:1 element:2 rec:6 pensive:1 database:1 blocking:2 observed:2 worst:1 ordering:1 russell:2 removed:1 ebl:1 ran:1 correc:1 ui:13 complexity:1 ultimately:1 depend:1 solving:3 upon:1 basis:1 unequally:2 univ:1 describe:5 london:1 quite:1 heuristic:4 elad:1 drawing:1 ability:1 favor:1 statistic:2 itself:1 validates:1 online:1 advantage:1 mg:1 vanishingly:1 remainder:1 j2:1 relevant:1 combining:1 description:1 seattle:1 requirement:8 produce:1 ldi:1 strong:2 implemented:1 predicted:3 marcel:1 met:2 correct:7 attribute:1 stochastic:1 human:1 public:1 adjacency:3 require:3 subdivided:1 extension:3 normal:5 seed:1 algorithmic:1 scope:1 u3:4 early:1 catlett:1 purpose:1 estimation:1 proc:7 outperformed:1 mit:2 ujl:1 always:3 rather:1 minton:1 june:2 rank:4 likelihood:1 hk:3 tech:1 rigorous:1 am:1 inference:2 el:19 entire:1 pasadena:1 irn:1 relation:1 overall:9 among:5 issue:1 orientation:1 html:1 equal:1 once:1 never:1 sampling:6 represents:1 unnecessarily:1 unsupervised:1 excessive:1 future:2 btt:1 modern:1 individual:2 stechert:4 freedom:1 investigate:1 highly:1 evaluation:7 unlucky:1 analyzed:1 allocating:1 norwell:1 grove:1 allocates:1 respective:1 tree:1 incomplete:1 divide:2 desired:3 theoretical:1 instance:2 earlier:1 disadvantage:1 turnbull:3 cost:4 applicability:2 deviation:1 kaelbling:1 rare:1 reported:1 answer:1 synthetic:3 chooses:1 density:1 ie:1 probabilistic:2 lee:1 pool:1 again:1 american:1 usable:1 account:1 potential:2 amples:1 inc:1 satisfy:3 sloan:1 ranking:58 later:2 view:2 performed:1 observing:1 option:1 parallel:1 variance:7 likewise:1 efficiently:1 bayesian:2 produced:1 craig:1 drive:1 presume:1 andre:1 sharing:1 ed:1 definition:1 pp:10 associated:1 mi:1 sampled:2 rational:1 treatment:2 subsection:1 ut:1 knowledge:3 nasa:3 steve:2 improved:2 wei:4 formulation:2 generality:1 until:2 d:2 dac:1 su:1 gaming:1 defines:2 indicated:1 columbus:2 perhaps:1 mary:1 requiring:2 true:2 inductive:1 burl:1 documenting:2 assigned:1 laboratory:1 moore:1 adjacent:4 during:1 bonferroni:1 oc:1 criterion:9 theoretic:3 demonstrate:1 fj:1 pro:1 reasoning:1 common:1 rl:2 hie:3 interpretation:3 kluwer:1 significant:1 blocked:2 cambridge:1 imposing:1 hogg:1 illinois:1 longer:1 base:1 add:1 recent:1 dictated:1 perspective:1 scenario:1 inequality:1 rep:1 minimum:1 additional:2 somewhat:1 impose:1 determine:1 ud:1 july:4 multiple:2 jet:1 academic:1 cross:1 divided:1 concerning:1 equally:1 parenthesis:1 impact:1 santner:1 achieved:1 beam:1 separately:1 interval:1 allocated:1 publisher:1 probably:4 subject:1 gratch:4 thing:1 inconsistent:1 effectiveness:3 enough:3 variety:1 competing:1 suboptimal:1 reduce:3 allocate:1 utility:16 ul:3 cause:1 action:1 adequate:1 clear:2 extensively:1 wes:1 http:1 exist:1 estimated:6 algorithmically:1 correctly:2 glean:1 dominance:2 key:1 demonstrating:1 achieving:1 drawn:1 realworld:1 jose:2 family:2 reasonable:1 decision:11 comparable:2 bound:4 hi:8 stipulation:1 constraint:1 uii:1 bibliography:1 u1:5 speed:1 aspect:1 speedup:2 structured:1 according:1 combination:1 poor:1 across:1 inote:1 penetration:1 hl:1 taken:2 computationally:1 equation:8 resource:2 previously:1 turn:1 ihe:1 serf:1 aaai92:3 available:3 experimentation:1 appropriate:1 k_l:1 alternative:4 voice:1 robustness:1 denotes:1 remaining:1 ensure:1 subsampling:1 uj:2 establish:1 question:1 parametric:1 evolutionary:1 rationally:2 lends:1 sci:1 propulsion:1 topic:1 considers:1 trivial:1 reason:2 induction:1 minimizing:1 expense:1 favorably:1 design:7 implementation:1 policy:2 unknown:5 perform:1 observation:3 datasets:4 howard:2 finite:2 immediate:1 situation:1 communication:1 rn:1 august:1 community:1 pair:3 required:2 userspecified:1 specified:1 california:1 unequal:3 trans:1 address:1 able:1 below:1 pattern:1 including:1 max:3 explanation:1 serially:1 scarce:1 drec:1 recursion:1 normality:1 technology:1 transitive:1 fax:1 review:1 literature:4 prior:1 embedded:1 loss:27 allocation:4 validation:1 degree:1 changed:1 soil:1 populate:1 institute:1 wide:2 taking:1 correspondingly:1 distributed:5 feedback:1 default:1 depth:1 world:4 evaluating:1 valid:2 cumulative:1 asserting:1 t5:4 reinforcement:1 san:2 adaptive:1 far:2 restarting:1 chien:8 assumed:1 search:3 iterative:1 table:5 correlated:1 inherently:1 ca:3 improving:1 unavailable:1 requested:2 broadening:1 investigated:1 complex:1 necessarily:1 artificially:1 composer:2 bounding:1 revisits:1 repeated:1 candidate:5 third:1 formula:1 pac:17 learnable:1 jpl:3 dominates:1 sequential:2 effectively:1 valiant:1 likely:1 greiner:1 ordered:1 macmillan:1 u2:15 restarted:1 applies:1 darren:1 chance:1 satisfies:1 determines:2 ma:5 uiucdcs:1 acm:1 viewed:1 consequently:1 content:1 change:3 hard:1 considerable:1 conservative:1 total:2 experimental:2 formally:1 brunswick:1 dept:1 ex:1 |
426 | 1,389 | Local Dimensionality Reduction
Stefan Schaal
1,2,4
sschaal@usc.edu
http://www-slab.usc.edulsschaal
Sethu Vijayakumar 3, I
Christopher G. Atkeson 4
sethu@cs.titech.ac.jp
http://ogawawww.cs.titech.ac.jp/-sethu
cga@cc.gatech.edu
http://www.cc.gatech.edul
fac/Chris.Atkeson
IERATO Kawato Dynamic Brain Project (IST), 2-2 Hikaridai, Seika-cho, Soraku-gun, 619-02 Kyoto
2Dept. of Comp. Science & Neuroscience, Univ. of South. California HNB-I 03, Los Angeles CA 90089-2520
3Department of Computer Science, Tokyo Institute of Technology, Meguro-ku, Tokyo-I 52
4College of Computing, Georgia Institute of Technology, 801 Atlantic Drive, Atlanta, GA 30332-0280
Abstract
If globally high dimensional data has locally only low dimensional distributions, it is advantageous to perform a local dimensionality reduction before
further processing the data. In this paper we examine several techniques for
local dimensionality reduction in the context of locally weighted linear regression. As possible candidates, we derive local versions of factor analysis
regression, principle component regression, principle component regression
on joint distributions, and partial least squares regression. After outlining the
statistical bases of these methods, we perform Monte Carlo simulations to
evaluate their robustness with respect to violations of their statistical assumptions. One surprising outcome is that locally weighted partial least
squares regression offers the best average results, thus outperforming even
factor analysis, the theoretically most appealing of our candidate techniques.
1 INTRODUCTION
Regression tasks involve mapping a n-dimensional continuous input vector x E ~n onto
a m-dimensional output vector y E ~m ? They form a ubiquitous class of problems found
in fields including process control, sensorimotor control, coordinate transformations, and
various stages of information processing in biological nervous systems. This paper will
focus on spatially localized learning techniques, for example, kernel regression with
Gaussian weighting functions. Local learning offer advantages for real-time incremental
learning problems due to fast convergence, considerable robustness towards problems of
negative interference, and large tolerance in model selection (Atkeson, Moore, & Schaal,
1997; Schaal & Atkeson, in press). Local learning is usually based on interpolating data
from a local neighborhood around the query point. For high dimensional learning problems, however, it suffers from a bias/variance dilemma, caused by the nonintuitive fact
that " ... [in high dimensions] if neighborhoods are local, then they are almost surely
empty, whereas if a neighborhood is not empty, then it is not local." (Scott, 1992, p.198).
Global learning methods, such as sigmoidal feedforward networks, do not face this
634
S. School, S. Vijayakumar and C. G. Atkeson
problem as they do not employ neighborhood relations, although they require strong
prior knowledge about the problem at hand in order to be successful.
Assuming that local learning in high dimensions is a hopeless, however, is not necessarily warranted: being globally high dimensional does not imply that data remains high dimensional if viewed locally. For example, in the control of robot anns and biological
anns we have shown that for estimating the inverse dynamics of an ann, a globally 21dimensional space reduces on average to 4-6 dimensions locally (Vijayakumar & Schaal,
1997). A local learning system that can robustly exploit such locally low dimensional
distributions should be able to avoid the curse of dimensionality.
In pursuit of the question of what, in the context of local regression, is the "right"
method to perfonn local dimensionality reduction, this paper will derive and compare
several candidate techniques under i) perfectly fulfilled statistical prerequisites (e.g.,
Gaussian noise, Gaussian input distributions, perfectly linear data), and ii) less perfect
conditions (e.g., non-Gaussian distributions, slightly quadratic data, incorrect guess of
the dimensionality of the true data distribution). We will focus on nonlinear function approximation with locally weighted linear regression (L WR), as it allows us to adapt a variety of global linear dimensionality reduction techniques, and as L WR has found widespread application in several local learning systems (Atkeson, Moore, & Schaal, 1997;
Jordan & Jacobs, 1994; Xu, Jordan, & Hinton, 1996). In particular, we will derive and
investigate locally weighted principal component regression (L WPCR), locally weighted
joint data principal component analysis (L WPCA), locally weighted factor analysis
(L WF A), and locally weighted partial least squares (LWPLS). Section 2 will briefly outline these methods and their theoretical foundations, while Section 3 will empirically
evaluate the robustness of these methods using synthetic data sets that increasingly violate some of the statistical assumptions of the techniques.
2 METHODS OF DIMENSIONALITY REDUCTION
We assume that our regression data originate from a generating process with two sets of
observables, the "inputs" i and the "outputs" y. The characteristics of the process ensure a functional relation y = f(i). Both i and yare obtained through some measurement device that adds independent mean zero noise of different magnitude in each observable, such that x == i + Ex and y = y + Ey ? For the sake of simplicity, we will only focus on one-dimensional output data (m=l) and functions / that are either linear or
slightly quadratic, as these cases are the most common in nonlinear function approximation with locally linear models. Locality of the regression is ensured by weighting the error of each data point with a weight from a Gaussian kernel:
Wi = exp(-O.5(Xi - Xqf D(Xi - Xq))
(1)
Xtt denotes the query point, and D a positive semi-definite distance metric which determmes the size and shape of the neighborhood contributing to the regression (Atkeson et
aI., 1997). The parameters Xq and D can be determined in the framework of nonparametric statistics (Schaal & Atkeson, in press) or parametric maximum likelihood estimations
(Xu et aI, 1995}- for the present study they are determined manually since their origin is
secondary to the results of this paper. Without loss of generality, all our data sets will set
!,q to the zero vector, compute the weights, and then translate the input data such that the
locally weighted mean, i = L WI Xi / L Wi , is zero. The output data is equally translated to
be mean zero. Mean zero data is necessary for most of techniques considered below. The
(translated) input data is summarized in the rows of the matrix X, the corresponding
(translated) outputs are the elements of the vector y, and the corresponding weights are in
the diagonal matrix W. In some cases, we need the joint input and output data, denoted
as Z=[X y).
Local Dimensionality Reduction
635
2.1 FACTORANALYSIS(LWFA)
Factor analysis (Everitt, 1984) is a technique of dimensionality reduction which is the
most appropriate given the generating process of our regression data. It assumes the observed data z was produced. by a mean zero independently distributed k -dimensional
vector of factors v, transformed by the matrix U, and contaminated by mean zero independent noise f: with diagonal covariance matrix Q:
z=Uv+f:, where
z=[xT,yt
and
f:=[f:~,t:yr
(2)
If both v and f: are normally distributed, the parameters Q and U can be obtained iteratively by the Expectation-Maximization algorithm (EM) (Rubin & Thayer, 1982). For a
linear regression problem, one assumes that z was generated with U=[I, f3 Y and v = i,
where f3 denotes the vector of regression coefficients of the linear model y = f31 x, and I
the identity matrix. After calculating Q and U by EM in joint data space as formulated in
(2), an estimate of f3 can be derived from the conditional probability p(y I x). As all
distributions are assumed to be normal, the expected value ofy is the mean of this conditional distribution. The locally weighted version (L WF A) of f3 can be obtained together
with an estimate of the factors v from the joint weighted covariance matrix 'I' of z and v:
E{[:]
+[}
~ ~ ~,,~,;'x,
Q+UU T
=
UT
[
where
~ ~ [ZT, VT~~Jft: w; ~
(3)
U] ['I'II(=n x n)
'I'12(=nX(m+k?)]
I = '?21(= (m + k) x n) '1'22(= (m + k) x (m + k?)
where E { .} denotes the expectation operator and B a matrix of coefficients involved in
estimating the factors v. Note that unless the noise f: is zero, the estimated f3 is different
from the true f3 as it tries to average out the noise in the data.
2.2 JOINT-SPACE PRINCIPAL COMPONENT ANALYSIS (LWPCA)
An alternative way of determining the parameters f3 in a reduced space employs locally
weighted principal component analysis (LWPCA) in the joint data space. By defining the .
largest k+ 1 principal components of the weighted covariance matrix ofZ as U:
U= [eigenvectors(I Wi (Zi - ZXZi - Z)T II Wi)]
(4)
max(l :k+1l
and noting that the eigenvectors in U are unit length, the matrix inversion theorem (Hom
& Johnson, 1994) provides a means to derive an efficient estimate of f3
T
T(
T
f3=U x(Uy -Uy UyUy -I
)-1 UyU yt
T\
[Ux(=nXk)]
where U= Uy(=mxk)
(5)
In our one dimensional output case, U y is just a (1 x k) -dimensional row vector and the
evaluation of (5) does not require a matrix inversion anymore but rather a division.
If one assumes normal distributions in all variables as in LWF A, LWPCA is the special
case of L WF A where the noise covariance Q is spherical, i.e., the same magnitude of
noise in all observables. Under these circumstances, the subspaces spanned by U in both
methods will be the same. However, the regression coefficients of LWPCA will be different from those of LWF A unless the noise level is zero, as LWFA optimizes the coefficients according to the noise in the data (Equation (3? . Thus, for normal distributions
and a correct guess of k, LWPCA is always expected to perform worse than LWFA.
S. Schaal, S. Vijayakumar and C. G. Atkeson
636
2.3 PARTIAL LEAST SQUARES (LWPLS, LWPLS_I)
Partial least squares (Wold, 1975; Frank & Friedman, 1993) recursively computes orthogonal projections of the input data and performs single variable regressions along
these projections on the residuals of the previous iteration step. A locally weighted version of partial least squares (LWPLS) proceeds as shown in Equation (6) below.
As all single variable regressions are ordinary uniFor Training:
For Lookup:
variate least-squares minim izations,
L WPLS
Initialize:
Initialize:
makes the same statistical assumption as ordinary
Do = X, eo = y do = x, y=
linear regressions, i.e., that only output variables
have additive noise, but input variables are noiseFor i = 1 to k:
For i = 1 to k:
less. The choice of the projections u, however, ins. = dT.u.
troduces an element in LWPLS that remains statistically still debated (Frank & Friedman, 1993), although, interestingly, there exists a strong similarity with the way projections are chosen in Cascade
Correlation (Fahlman & Lebiere, 1990). A peculiarity of LWPLS is that it also regresses the inputs
of the previous step against the projected inputs s
in order to ensure the orthogonality of all the projections u. Since LWPLS chooses projections in a
(6)
very powerful way, it can accomplish optimal
function fits with only one single projections (i.e.,
k= 1) for certain input distributions. We will address this issue in our empirical evaluations by comparing k-step LWPLS with I-step LWPLS, abbreviated LWPLS_I.
?
I
1-
I
2.4 PRINCIPAL COMPONENT REGRESSION (L WPCR)
Although not optimal, a computationally efficient techniques of dimensionality reduction
for linear regression is principal component regression (LWPCR) (Massy, 1965). The inputs are projected onto the largest k principal components of the weighted covariance
matrix of the input data by the matrix U:
U = [eig envectors(2: Wi (Xi -
xX xt /2:
Xi -
(7)
Wi )]
max(l:k)
The regression coefficients f3 are thus calculated as:
f3
= (UTXTwxUtUTXTWy
(8)
Equation (8) is inexpensive to evaluate since after projecting X with U, UTXTWXU becomes a diagonal matrix that is easy to invert. LWPCR assumes that the inputs have additive spherical noise, which includes the zero noise case. As during dimensionality reduction LWPCR does not take into account the output data, it is endangered by clipping
input dimensions with low variance which nevertheless have important contribution to
the regression output. However, from a statistical point of view, it is less likely that low
variance inputs have significant contribution in a linear regression, as the confidence
bands of the regression coefficients increase inversely proportionally with the variance of
the associated input. If the input data has non-spherical noise, L WPCR is prone to focus
the regression on irrelevant projections.
3 MONTE CARLO EVALUATIONS
In order to evaluate the candidate methods, data sets with 5 inputs and 1 output were randomly generated. Each data set consisted of 2,000 training points and 10,000 test points,
distributed either uniformly or nonuniformly in the unit hypercube. The outputs were
Local Dimensionality Reduction
637
generated by either a linear or quadratic function. Afterwards, the 5-dimensional input
space was projected into a to-dimensional space by a randomly chosen distance preserving linear transformation. Finally, Gaussian noise of various magnitudes was added
to both the 10-dimensional inputs and one dimensional output. For the test sets, the additive noise in the outputs was omitted. Each regression technique was localized by a
Gaussian kernel (Equation (1)) with a to-dimensional distance metric D=IO*I (D was
manually chosen to ensure that the Gaussian kernel had sufficiently many data points and
no "data holes" in the fringe areas of the kernel) . The precise experimental conditions
followed closely those suggested by Frank and Friedman (1993):
{g.I for:
131.. = [I, I, I, I, If ,
?
2 kinds of linear functions y =
?
2 kinds of quadratic functions y = f3J.I + f3::.aAxt ,xi
i)
?
1311.
i)
= [I, I, I, I, Wand f3q.ad = 0.1 [I, I, I, I, If, and ii)
ii)
I3Ii. = [1,2,3,4, sf
,xi ,X;,X;]T for:
131.. = [1,2,3,4, sf
and f3q uad = 0.1 [I, 4, 9, 16, 2sf
3 kinds of noise conditions, each with 2 sub-conditions:
local signal/noise ratio Isnr=20,
i) only output noise:
a) low noise:
and
b) high noise:
Isnr=2,
ii) equal noise in inputs and outputs:
a) low noise Ex ?? = Sy = N(O,O.Ot2), n e[I,2, ... ,10],
and
b) high noise Ex ?? =sy=N(0,0.1 2),ne[I,2, ... ,10],
iii) unequal noise in inputs and outputs:
a) low noise : Ex .? = N(0,(0.0In)2), n e[I,2, . .. ,1O] and Isnr=20,
and
?
b) high noise:
Ex .?
= N(0,(0.0In)2), n e[I,2, ... ,1O] and Isnr=2,
2 kinds of input distributions: i) uniform in unit hyper cube, ii) uniform in unit hyper cube excluding data
points which activate a Gaussian weighting function (I) at c = [O.S,O,o,o,of with D=IO*I more than
w=0.2 (this forms a "hyper kidney" shaped distribution)
Every algorithm was run * 30 times on each of the 48 combinations of the conditions.
Additionally, the complete test was repeated for three further conditions varying the dimensionality--called factors in accordance with LWFA-that the algorithms assumed to
be the true dimensionality of the to-dimensional data from k=4 to 6, i.e., too few, correct,
and too many factors. The average results are summarized in Figure I.
Figure I a,b,c show the summary results of the three factor conditions. Besides averaging
over the 30 trials per condition, each mean of these charts also averages over the two input distribution conditions and the linear and quadratic function condition, as these four
cases are frequently observed violations of the statistical assumptions in nonlinear function approximation with locally linear models. In Figure I b the number of factors equals
the underlying dimensionality of the problem, and all algorithms are essentially performing equally well. For perfectly Gaussian distributions in all random variables (not
shown separately), LWFA's assumptions are perfectly fulfilled and it achieves the best
results, however, almost indistinguishable closely followed by LWPLS. For the ''unequal
noise condition", the two PCA based techniques, LWPCA and LWPCR, perform the
worst since--as expected-they choose suboptimal projections. However, when violating the statistical assumptions, L WF A loses parts of its advantages, such that the summary results become fairly balanced in Figure lb.
The quality of function fitting changes significantly when violating the correct number of
factors, as illustrated in Figure I a,c. For too few factors (Figure la), LWPCR performs
worst because it randomly omits one of the principle components in the input data, without respect to how important it is for the regression. The second worse is LWF A: according to its assumptions it believes that the signal it cannot model must be noise, leading to a degraded estimate of the data's subspace and, consequently, degraded regression
results. LWPLS has a clear lead in this test, closely followed by LWPCA and LWPLS_I.
* Except for LWFA, all methods can evaluate a data set in non-iterative calculations. LWFA was trained with EM for maximally 1000 iterations or until the log-likelihood increased less than I.e-lOin one iteration.
S. Schaal, S. Vljayakumar and C. G. Atkeson
638
For too many factors than necessary (Figure Ie), it is now LWPCA which degrades. This
effect is due to its extracting one very noise contaminated projection which strongly influences the recovery of the regression parameters in Equation (4). All other algorithms
perform almost equally well, with LWF A and LWPLS taking a small lead.
Equal NoIse In ell
In puIS end OutpUIS
OnlyOutpul
Noise
Unequel NoIse In ell
Inputs end OutpulS
0.1
c
o
~
::::;;
0.01
c
II>
C>
~ 0.001
~
0.0001
fl- I. E>O
~I. ? >>(I
~ J. &>O
~J , E ? O
~ J .E>O
fl- I.&? O ~I . & >O
~ I . & > >O
~I.& >O
~ I .?>>o
p,. 1. s>O
tJ-J .?>>O
e) RegressIon Results with 4 Factors
?
LWFA
?
LWPCA
?
LWPCR
0
LWPLS
?
LWPLS_1
0.1
c::
o
W
~c::
0.01
II>
C)
~ 0.001
~
0.0001
~
~
0.1
W
0.01
8
~c::
g,
~
0.001
~
0.0001
c) RegressIon Results with 6 Feclors
jj
il
0.1
f-
a
~
::::;;
0.01
c
II>
C)
~ 0.001
~
0.0001
d) Summery Results
Figure I: Average summary results of Monte Carlo experiments. Each chart is primarily
divided into the three major noise conditions, cf. headers in chart (a). In each noise condition, there are four further subdivision: i) coefficients of linear or quadratic model are
equal with low added noise; ii) like i) with high added noise; iii) coefficients oflinear or
quadratic model are different with low noise added; iv) like iii) with high added noise.
Refer to text and descriptions of Monte Carlo studies for further explanations.
Local Dimensionality Reduction
639
4 SUMMARY AND CONCLUSIONS
Figure 1d summarizes all the Monte Carlo experiments in a final average plot. Except for
LWPLS, every other technique showed at least one clear weakness in one of our "robustness" tests. It was particularly an incorrect number of factors which made these weaknesses apparent. For high-dimensional regression problems, the local dimensionality, i.e.,
the number of factors, is not a clearly defined number but rather a varying quantity, depending on the way the generating process operates. Usually, this process does not need
to generate locally low dimensional distributions, however, it often "chooses" to do so,
for instance, as human ann movements follow stereotypic patterns despite they could
generate arbitrary ones. Thus, local dimensionality reduction needs to find autonomously
the appropriate number of local factor. Locally weighted partial least squares turned out
to be a surprisingly robust technique for this purpose, even outperforming the statistically
appealing probabilistic factor analysis. As in principal component analysis, LWPLS's
number of factors can easily be controlled just based on a variance-cutoff threshold in input space (Frank & Friedman, 1993), while factor analysis usually requires expensive
cross-validation techniques. Simple, variance-based control over the number of factors
can actually improve the results of LWPCA and LWPCR in practice, since, as shown in
Figure I a, LWPCR is more robust towards overestimating the number of factors, while
L WPCA is more robust towards an underestimation. If one is interested in dynamically
growing the number of factors while obtaining already good regression results with too
few factors, L WPCA and, especially, L WPLS seem to be appropriate-it should be
noted how well one factor LWPLS (LWPLS_l) already performed in Figure I!
In conclusion, since locally weighted partial least squares was equally robust as local
weighted factor analysis towards additive noise in. both input and output data, and,
moreover, superior when mis-guessing the number of factors, it seems to be a most favorable technique for local dimensionality reduction for high dimensional regressions.
Acknowledgments
The authors are grateful to Geoffrey Hinton for reminding them of partial least squares. This work was supported by the ATR Human Information Processing Research Laboratories. S. Schaal's support includes the
German Research Association, the Alexander von Humboldt Foundation, and the German Scholarship Foundation. S. Vijayakumar was supported by the Japanese Ministry of Education, Science, and Culture (Monbusho).
C. G. Atkeson acknowledges the Air Force Office of Scientific Research grant F49-6209410362 and a National
Science Foundation Presidential Young Investigators Award.
References
Atkeson, C. G., Moore, A. W., & Schaal, S, (1997a).
"Locally weighted learning." ArtifiCial Intelligence Review, 11, 1-5, pp.II-73.
Atkeson, C. G., Moore, A. W., & Schaal, S, (1997c).
"Locally weighted learning for control." ArtifiCial Intelligence Review, 11, 1-5, pp.75-113.
Belsley, D. A., Kuh, E., & Welsch, R. E, (1980). Regression diagnostics: Identifying influential data and
sources of collinearity. New York: Wiley.
Everitt, B. S, (1984). An introduction to latent variable
models. London: Chapman and Hall.
Fahlman, S. E. ,Lebiere, C, (1990). "The cascadecorrelation learning architecture." In: Touretzky, D. S.
(Ed.), Advances in Neural Information Processing
Systems II, pp.524-532. Morgan Kaufmann.
Frank, I. E., & Friedman, 1. H, (1993). "A statistical
view of some chemometric regression tools." Technometrics, 35, 2, pp.l09-135.
Geman, S., Bienenstock, E., & Doursat, R. (1992).
"Neural networks and the bias/variance dilemma."
Neural Computation, 4, pp.I-58.
Hom, R. A., & Johnson, C. R, (1994). Matrix analySis.
Press Syndicate of the University of Cambridge.
Jordan, M.I., & Jacobs, R, (1994). "Hierarchical mix-
tures of experts and the EM algorithm." Neural Computation, 6, 2, pp.181-214.
Massy, W. F, (1965). "Principle component regression
in exploratory statistical research." Journal ofthe
American Statistical Association, 60, pp.234-246.
Rubin, D. B., & Thayer, D. T, (l982). "EM algorithms
for ML factor analysis." Psychometrika, 47, I, 69-76.
Schaal, S., & Atkeson, C. G, (in press). "Constructive
incremental learning from only local information."
Neural Computation.
Scott, D. W, (1992). Multivariate Density Estimation.
New York: Wiley.
Vijayakumar, S., & Schaal, S, (1997). "Local dimensionality reduction for locally weighted learning." In:
International Conference on Computational Intelligence in Robotics and Automation, pp.220-225, Monteray, CA, July 10-11, 1997.
Wold, H. (1975). "Soft modeling by latent variables:
the nonlinear iterative partial least squares approach."
In: Gani, J. (Ed.), Perspectives in Probability and Statistics, Papers in Honour ofM S. Bartlett. Aca<j. Press.
Xu, L., Jordan, M.l., & Hinton, G. E, (1995). "An alternative model for mixture of experts." In: Tesauro,
G., Touretzky, D. S., & Leen, T. K. (Eds.), Advances in
Neural Information Processing Systems 7, pp.633-640.
Cambridge, MA: MIT Press.
Serial Order in Reading Aloud:
Connectionist Models and Neighborhood
Structure
Jeanne C. Milostan
Computer Science & Engineering 0114
University of California San Diego
La Jolla, CA 92093-0114
Garrison W. Cottrell
Computer Science & Engineering 0114
University of California San Diego
La Jolla, CA 92093-0114
Abstract
Dual-Route and Connectionist Single-Route models ofreading have
been at odds over claims as to the correct explanation of the reading
process. Recent Dual-Route models predict that subjects should
show an increased naming latency for irregular words when the irregularity is earlier in the word (e.g. chef is slower than glow) - a
prediction that has been confirmed in human experiments. Since
this would appear to be an effect of the left-to-right reading process,
Coltheart & Rastle (1994) claim that Single-Route parallel connectionist models cannot account for it. A refutation of this claim is
presented here, consisting of network models which do show the
interaction, along with orthographic neighborhood statistics that
explain the effect.
1
Introduction
A major component of the task of learning to read is the development of a mapping
from orthography to phonology. In a complete model of reading, message understanding must playa role, but many psycholinguistic phenomena can be explained
in the context of this simple mapping task. A difficulty in learning this mapping
is that in a language such as English, the mapping is quasiregular (Plaut et al.,
1996); there are a wide range of exceptions to the general rules. As with nearly
all psychological phenomena, more frequent stimuli are processed faster, leading to
shorter naming latencies. The regularity of mapping interacts with this variable,
a robust finding that is well-explained by connectionist accounts (Seidenberg and
M.cClelland, 1989; Taraban and McClelland, 1987).
In this paper we consider a recent effect that seems difficult to account for in terms
of the standard parallel network models. Coltheart & Rastle (1994) have shown
1. C. Milostan and G. W. Cottrell
60
Position
of
2
Irregular
3
Phoneme
1
4
5
Irregular
Regular Control
Difference
554
502
52
542
516
26
530
518
12
529
523
6
537
525
12
Irregular
Regular Control
Difference
Avg. Difl'.
545
500
45
48.5
524
503
21
23.5
528
503
25
18.5
526
515
11
8.5
528
524
4
8
Filler
Nonword
Exception
Table 1: Naming Latency vs. Irregularity Position
that the amount of delay experienced in naming an exception word is related to the
phonemic position of the irregularity in pronunciation. Specifically, the earlier the
exception occurs in the word, the longer the latency to the onset of pronouncing
the word. Table 1, adapted from (Coltheart and Rastle, 1994) shows the response
latencies to two-syllable words by normal subjects. There is a clear left-to-right
ranking of the latencies compared to controls in the last row of the Table. Coltheart
et al. claim this delay ranking cannot be achieved by standard connectionist models.
This paper shows this claim to be false, and shows that the origin of the effect lies in
a statistical regularity of English, related to the number of "friends" and "enemies"
of the pronunciation within the word's neighborhood 1.
2
Background
Computational modeling of the reading task has been approached from a number
of different perspectives. Advocates of a dual-route model of oral reading claim
that two separate routes, one lexical (a lexicon, often hypothesized to be an associative network) and one rule-based, are required to account for certain phenomena
in reaction times and nonword pronunciation seen in human subjects (Coltheart
et al., 1993). Connectionist modelers claim that the same phenomena can be captured in a single-route model which learns simply by exposure to a representative
dataset (Seidenberg and McClelland, 1989).
In the Dual-Route Cascade model (DRC) (Coltheart et al., 1993), the lexical route
is implemented as an Interactive Activation (McClelland and Rumelhart, 1981)
system, while the non-lexical route is implemented by a set of grapheme-phoneme
correspondence (GPC) rules learned from a dataset. Input at the letter identification layer is activated in a left-to-right sequential fashion to simulate the reading
direction of English, and fed simultaneously to the two pathways in the model.
Activation from both the GPC route and the lexicon route then begins to interact
at the output (phoneme) level, starting with the phonemes at the beginning of the
word. If the GPC and the lexicon agree on pronunciation, the correct phonemes
will be activated quickly. For words with irregular pronunciation, the lexicon and
GPC routes will activate different phonemes: the GPC route will try to activate
the regular pronunciation while the lexical route will activate the irregular (correct)
1 Friends are words with the same pronunciations for the ambiguous letter-ta-sound
correspondence; enemies are words with different pronunciations.
Serial Ortier in Reading Aloud
61
pronunciation. Inhibitory links between alternate phoneme pronunciations will slow
down the rise in activation, causing words with inconsistencies to be pronounced
more slowly than regular words. This slowing will not occur, however, when an
irregularity appears late in a word. This is because in the model the lexical node
spreads activation to all of a word's phonemes as soon as it becomes active. If an
irregularity is late in a word, the correct pronunciation will begin to be activated
before the GPC route is able to vote against it. Hence late irregularities will not be
as affected by conflicting information. This result is validated by simulations with
the one-syllable DRC model (Coltheart and Rastle, 1994).
Several connectionist systems have been developed to model the orthography to
phonology process (Seidenberg and McClelland, 1989; Plaut et al., 1996). These
connectionist models provide evidence that the task, with accompanying phenomena, can be learned through a single mechanism. In particular, Plaut et al. (henceforth PMSP) develop a recurrent network which duplicates the naming latencies
appropriate to their data set, consisting of approximately 3000 one-syllable English words (monosyllabic words with frequency greater than 1 in the Kucera &
Francis corpus (Kucera and Francis, 1967?. Naming latencies are computed based
on time-t~settle for the recurrent network, and based on MSE for a feed-forward
model used in some simulations. In addition to duplicating frequency and regularity
interactions displayed in previous human studies, this model also performs appr~
priately in providing pronunciation of pronounceable nonwords. This provides an
improvement over, and a validation of, previous work with a strictly feed-forward
network (Seidenberg and McClelland, 1989). However, to date, no one has shown
that Coltheart's naming latency by irregularity of position interaction can be accounted for by such a model. Indeed, it is difficult to see how such a model could
account for such a phenomenon, as its explanation (at least in the DRC model)
seems to require the serial, left-t~right nature of processing in the model, whereas
networks such as PMSP present the word orthography all at once. In the following,
we fill this gap in the literature, and explain why a parallel, feed-forward model can
account for this result.
Experiments & Results
3
3.1
The Data
Pronunciations for approximately 100,000 English words were obtained through an
electronic dictionary developed by CMU 2 . The provided format was not amenable
to an automated method for distinguishing the number of syllables in the word. To
obtain syllable counts, English tw~syllable words were gathered from the Medical
Research Council (MRC) Psycholinguistic Database (Coltheart and Rastle, 1994),
which is conveniently annotated with syllable counts and frequency (only those with
Kucera-Francis written frequency of one or greater were selected). Intersecting the
two databases resulted in 5,924 tw~syllable words. There is some noise in the data;
ZONED and AERIAL, for example, are in this database of purported tw~syllable
words. Due to the size of the database and time limitations, we did not prune the
data of these errors, nor did we eliminate proper nouns or foreign words. Singlesyllable words with the same frequency criterion were also selected for comparison
with previous work. 3,284 unique single-syllable words were obtained, in contrast
to 2,998 words used by PMSP. Similar noisy data as in the tw~syllable set exists
in this database. Each word was represented using the orthography and phonology
representation scheme outlined by PMSP.
2 Available
via ftp://ftp.cs.cmu.edu/project/fgdata/dict/
1. C. Milostan and G. W Cottrell
62
1.0
S
0.8
I
0.6
""it.
J...
0.4
r!I
02
>::'
Ii
Figure 1: I-syllable network latency differences & neighborhood statistics
3.2
Methods
For the single syllable words, we used an identical network to the feed-forward network used by PMSP, i.e., a 105-100-61 network, and for the two syllable words, we
simply used the same architecture with the each layer size doubled. We trained each
network for 300 epochs, using batch training with a cross entropy objective function,
an initial learning rate of 0.001, momentum of 0.9 after the first 10 epochs, weight
decay of 0.0001, and delta-bar-delta learning rate adjustment. Training exemplars
were weighted by the log of frequency as found in the Kucera-Francis corpus. After this training, the single syllable feed-forward networks averaged 98.6% correct
outputs, using the same evaluation technique outlined in PMSP. Two syllable networks were trained for 1700 epochs using online training, a learning rate of 0.05,
momentum of 0.9 after the first 10 epochs, and raw frequency weighting. The two
syllable network achieved 85% correct. Naming latency was equated with network
output MSE; for successful results, the error difference between the irregular words
and associated control words should decrease with irregularity position.
3.3
Results
Single Syllable Words First, Coltheart's challenge that a single-route model
cannot produce the latency effects was explored. The single-syllable network described above was tested on the collection of single-syllable words identified as irregular by (Taraban and McClelland, 1987). In (Coltheart and Rastle, 1994), control
words are selected based on equal number of letters, same beginning phoneme, and
Kucera-Francis frequency between 1 and 20 (controls were not frequency matched).
For single syllable words used here, the control condition was modified to allow
frequency from 1 to 70, which is the range of the "low frequency" exception words
in the Taraban & McClelland set. Controls were chosen by drawing randomly from
the words meeting the control criteria.
Each test and control word input vector was presented to the network, and the
MSE at the output layer (compared to the expected correct target) was calculated.
From these values, the differences in MSE for target and matched control words
were calculated and are shown in Figure 1. Note that words with an irregularity
in the first phoneme position have the largest difference from their control words,
with this (exception - regular control) difference decreasing as phoneme position
increases. Contrary to the claims of the Dual-Route model, this network does show
the desired rank-ordering of MSE/latency.
63
Serial Order in Reading Aloud
1.0
02
I'
d
I
1;l
0.1
0.0
::I!
o
4
I'boaeme 1........ larit)' PooIIioa
O.O+--~--r----.---~--'
6
o
2
"
Pbonane lnegularlty PosItioIl
6
Figure 2: 2-syllable network latency differences & neighborhood statistics
Two Syllable Words Testing of the two-syllable network is identical to that
of the one-syllable network. The difference in MSE for each test word and its
corresponding control is calculated, averaging across all test pairs in the position
set. Both test words and their controls are those found in (Coltheart and Rastle,
1994). The 2-syllable network appears to produce approximately the correct linear
trend in the naming MSE/latency (Figure 2), although the results displayed are not
monotonically decreasing with position. Note, however, that the results presented
by Coltheart, when taken separately, also fail to exhibit this trend (Table 1). For
correct analysis, several "subject" networks should be trained, with formal linear
trend analysis then performed with the resulting data. These further simulations
are currently being undertaken.
4
Why the network works: Neighborhood effects
A possible explanation for these results relies on the fact that connectionist networks
tend to extract statistical regularities in the data, and are affected by regularity by
frequency interactions. In this case, we decided to explore the hypothesis that
the results could be explained by a neighborhood effect: Perhaps the number of
"friends" and "enemies" in the neighborhood (in a sense to be defined below) of
the exception word varies in English in a position-dependent way. If there are more
enemies (different pronunciations) than friends (identical pronunciations) when the
exception occurs at the beginning of a word than at the end, then one would expect
a network to reflect this statistical regularity in its output errors. In particular, one
would expect higher errors (and therefore longer latencies in naming) if the word
has a higher proportion of enemies in the neighborhood.
To test this hypothesis, we created some data search engines to collect word neighborhoods based on various criteria. There is no consensus on the exact definition
of the "neighborhood" of a word. There are some common measures, however, so
we explored several of these. Taraban & McClelland (1987) neighborhoods (T&M)
are defined as words containing the same vowel grouping and final consonant cluster. These neighborhoods therefore tend to consist of words that rhyme (MUST,
DUST, TRUST). There is independent evidence that these word-body neighbors
are psychologically relevant for word naming tasks (i.e., pronunciation) (Treiman
and Chafetz, 1987). The neighborhood measure given by Coltheart (Coltheart and
Rastle, 1994), N, counts same-length words which differ by only one letter, taking string position into account. Finally, edit-distance-1 (ED1) neighborhoods are
those words which can be generated from the target word by making one change
J. C. Milostan and G. W Cottrell
64
(Peereman, 1995): either a letter substitution, insertion or deletion. This differs
from the Coltheart N definition in that "TRUST" is in the EDI neighborhood (but
not the N neighborhood) of "RUST" , and provides a neighborhood measure which
considers both pronunciation and spelling similarity. However, the N and the ED-l
measure have not been shown to be psychologically real in terms of affecting naming
latency (Treiman and Chafetz, 1987).
We therefore extended T&M neighborhoods to multi-syllable words. Each vowel
group is considered within the context of its rime, with each syllable considered
separately. Consonant neighborhoods consist of orthographic clusters which correspond to the same location in the word. This results in 4 consonant cluster
locations: first syllable onset, first syllable coda, second syllable onset, and second
syllable coda. Consonant cluster neighborhoods include the preceeding vowel for
coda consonants, and the following vowel for onset consonants.
The notion of exception words is also not universally agreed upon. Precisely which
words are exceptions is a function of the working definition of pronunciation and
regularity for the experiment at hand. Given a definition of neighborhood, then,
exception words can be defined as those words which do not agree with the phonological mapping favored by the majority of items in that particular neighborhood.
Alternatively, in cases assuming a set of rules for grapheme-phoneme correspondence, exception words are those which violate the rules which define thp majority
of pronunciations. For this investigation, single syllable exception words are those
defined as exception by the T&M neighborhood definition. For instance, PINT
would be considered an exception word compared to its neighbors MINT, TINT,
HINT, etc. Coltheart, on the other hand, defines exception words to be those for
which his G PC rules produce incorrect pronunciation. Since we are concerned with
addressing Coltheart's claims, these 2-syllable exception words will also be used
here.
4.1
Results
Single syllable words For each phoneme position, we compare each word with
irregularity at that position with its neighbors, counting the number of enemies
(words with alternate pronunciation at the supposed irregularity) and friends (words
with pronunciation in agreement) that it has. The T &M neighborhood numbers
(words containing the same vowel grouping and final consonant cluster) used in Figure 1 are found in (Taraban and McClelland, 1987). For each word , we calculate its
(enemy) / (friend+enemy) ratio; these ratios are then averaged over all the words in
the position set. The results using neighborhoods as defined in Taraban & McClelland clearly show the desired rank ordering of effect. First-position-irregularity
words have more "enemies" and fewer "friends" than third-position-irregularity
words, with the second-position words falling in the middle as desired. We suggest that this statistical regularity in the data is what the above networks capture.
However convincing these results may be, they do not fully address Coltheart's
data, which is for two syllable words of five phonemes or phoneme clusters, with
irregularities at each of five possible positions. Also, due to the size of the T&M
data set, there are only 2 members in the position I set, and the single-syllable data
only goes up to phoneme position 3. The neighborhoods for the two-syllable data
set were thus examined.
Two syllable results Recall that the two-syllable test words are those used in
the (Coltheart and Rastle, 1994) subject study, for which naming latency differences are shown in Table 1. CoItheart's I-letter-different neighborhood definition
Serial Order in Reading Aloud
65
is not very informative in this case, since by this criterion most of the target words
provided in (Coltheart and Rastle, 1994) are loners (i.e., have no neighbors at
all). However, using a neighborhood based on T&M-2 recreates the desired ranking
(Figure 2) as indicated by the ratio of hindering pronunciations to the total of the
helping and hindering pronunciations. As with the single syllable words, each test
word is compared with its neighbor words and the (enemy)/(friend+enemy) ratio is
calculated. Averaging over the words in each position set, we again see that words
with early irregularities are at a support disadvantage compared to words with late
irregularities.
5
Summary
Dual-Route models claim the irregularity position effect can only be accounted
for by two-route models with left-to-right activation of phonemes, and interaction
between GPC rules and the lexicon. The work presented in this paper refutes this
claim by presenting results from feed-forward connectionist networks which show the
same rank ordering of latency. Further, an analysis of orthographic neighborhoods
shows why the networks can do this: the effect is based on a statistical interaction
between friend/enemy support and position. Words with irregular orthographicphonemic correspondence at word beginning have less support from their neighbors
than words with later irregularities; it is this difference which explains the latency
results. The resulting statistical regularity is then easily captured by connectionist
networks exposed to representative data sets.
References
Coltheart, M., Curitis, B., Atkins, P., and Haller, M. (1993). Models of reading
aloud: Dual-route and parallel-distributed-processing approaches. Psychological Review, 100(4):589-608.
Coltheart, M. and Rastle, K. (1994). Serial processing in reading aloud: Evidence
for dual route models of reading. Journal of Experimental Psychology: Human
Perception and Performance, 20(6):1197-1211.
Kucera, H. and Francis, W. (1967). Computational Analysis of Present-Day American English. Brown University Press, Providence, RI.
McClelland, J. and Rumelhart, D. (1981). An interactive activation model of context
effects in letter perception: Part 1. an account of basic findings. Psychological
Review, 88:375-407.
Peereman, R. (1995). Naming regular and exception words: Further examination
of the effect of phonological dissension among lexical neighbours. European
Journal of Cognitive Psychology, 7(3):307-330.
Plaut, D., McClelland, J., Seidenberg, M., and Patterson, K. (1996). Understanding
normal and impaired word reading: Computational principles in quasi-regular
domains. Psychological Review, 103(1):56-115.
Seidenberg, M. and McClelland, J. (1989). A distributed, developmental model of
word recognition and naming. Psychological Review, 96:523-568.
Taraban, R. and McClelland, J. (1987). Conspiracy effects in word pronunciation.
Journal of Memory and Language, 26:608-631.
Treiman, R. and Chafetz, J. (1987). Are there onset- and rime-like units in printed
words? In Coltheart, M., editor, Attention and Performance XII: The Psychology of Reading. Erlbaum, Hillsdale, NJ.
| 1389 |@word collinearity:1 trial:1 version:3 middle:1 inversion:2 seems:3 proportion:1 advantageous:1 briefly:1 simulation:4 jacob:2 covariance:5 recursively:1 reduction:15 initial:1 substitution:1 interestingly:1 atlantic:1 reaction:1 comparing:1 surprising:1 activation:6 grapheme:2 must:3 written:1 cottrell:4 additive:4 informative:1 shape:1 plot:1 v:1 intelligence:3 selected:3 fewer:1 item:1 guess:2 nervous:1 device:1 yr:1 beginning:4 slowing:1 provides:3 plaut:4 node:1 lexicon:5 location:2 sigmoidal:1 five:2 along:2 wpls:2 become:1 incorrect:3 fitting:1 advocate:1 pathway:1 theoretically:1 expected:4 indeed:1 seika:1 growing:1 multi:1 examine:1 frequently:1 nor:1 brain:1 globally:3 spherical:3 decreasing:2 curse:1 becomes:2 psychometrika:1 xx:1 matched:2 underlying:1 provided:2 project:2 estimating:2 moreover:1 what:2 begin:2 kind:4 massy:2 string:1 developed:2 finding:2 transformation:2 nj:1 perfonn:1 duplicating:1 every:2 interactive:2 coltheart:24 ensured:1 control:20 normally:1 grant:1 medical:1 appear:1 unit:5 before:2 positive:1 engineering:2 local:25 accordance:1 io:2 despite:1 rhyme:1 approximately:3 examined:1 dynamically:1 collect:1 monosyllabic:1 range:2 statistically:2 averaged:2 uy:3 acknowledgment:1 unique:1 decided:1 testing:1 practice:1 orthographic:3 definite:1 differs:1 irregularity:18 isnr:4 area:1 empirical:1 significantly:1 cascade:2 projection:10 printed:1 confidence:1 word:96 regular:7 suggest:1 doubled:1 onto:2 ga:1 selection:1 operator:1 cannot:4 context:5 influence:1 www:2 yt:2 lexical:6 go:1 attention:1 starting:1 ed1:1 independently:1 kidney:1 exposure:1 simplicity:1 preceeding:1 recovery:1 identifying:1 rule:7 spanned:1 fill:1 his:1 exploratory:1 notion:1 coordinate:1 diego:2 target:4 exact:1 mxk:1 distinguishing:1 hypothesis:2 origin:2 agreement:1 drc:3 element:2 rumelhart:2 recognition:1 trend:3 expensive:1 particularly:1 geman:1 database:5 observed:2 role:1 capture:1 worst:2 calculate:1 difl:1 autonomously:1 ordering:3 decrease:1 nonwords:1 movement:1 balanced:1 developmental:1 insertion:1 dynamic:2 trained:4 grateful:1 oral:1 exposed:1 dilemma:2 upon:1 division:1 patterson:1 observables:2 translated:3 l982:1 easily:2 joint:7 represented:1 various:3 univ:1 fast:1 fac:1 activate:4 monte:5 london:1 query:2 artificial:2 approached:1 hyper:3 neighborhood:35 outcome:1 pronunciation:25 aloud:6 header:1 apparent:1 enemy:12 drawing:1 presidential:1 statistic:5 chef:1 noisy:1 final:3 online:1 associative:1 advantage:2 hindering:2 interaction:6 frequent:1 causing:1 turned:1 relevant:1 date:1 translate:1 meguro:1 supposed:1 description:1 pronounced:1 los:1 convergence:1 regularity:9 wpca:3 impaired:1 empty:2 cluster:6 generating:3 produce:3 incremental:2 perfect:1 ftp:2 derive:4 develop:1 ac:2 depending:1 friend:9 exemplar:1 recurrent:2 school:1 pronounceable:1 phonemic:1 strong:2 implemented:2 c:3 uu:1 differ:1 direction:1 closely:3 tokyo:2 correct:12 annotated:1 peculiarity:1 human:6 settle:1 education:1 explains:1 require:3 hillsdale:1 humboldt:1 rime:2 endangered:1 investigation:1 reminding:1 biological:2 strictly:1 helping:1 accompanying:1 around:1 hall:1 considered:4 sufficiently:1 exp:1 normal:5 mapping:7 predict:1 claim:11 slab:1 appr:1 sschaal:1 major:2 achieves:1 early:1 dictionary:1 omitted:1 purpose:1 estimation:2 favorable:1 currently:1 council:1 edit:1 largest:3 tool:1 weighted:21 stefan:1 mit:1 clearly:2 gaussian:10 always:1 modified:1 rather:2 avoid:1 varying:2 gatech:2 office:1 derived:1 focus:4 validated:1 schaal:13 improvement:1 rank:3 likelihood:2 contrast:1 wf:4 sense:1 jeanne:1 dependent:1 foreign:1 eliminate:1 bienenstock:1 relation:2 nxk:1 quasi:1 transformed:1 interested:1 issue:1 among:1 pronouncing:1 dual:8 denoted:1 favored:1 development:1 noun:1 special:1 fairly:1 ell:2 initialize:2 field:1 cube:2 f3:12 phonological:2 equal:5 once:1 manually:2 identical:3 chapman:1 shaped:1 nearly:1 contaminated:2 connectionist:11 overestimating:1 duplicate:1 primarily:1 employ:2 stimulus:1 randomly:4 neighbour:1 hint:1 simultaneously:1 national:1 resulted:1 few:3 summery:1 usc:2 consisting:2 vowel:5 friedman:5 technometrics:1 atlanta:1 message:1 investigate:1 evaluation:4 weakness:2 violation:2 mixture:1 diagnostics:1 pc:1 activated:3 tj:1 amenable:1 partial:10 necessary:2 culture:1 shorter:1 orthogonal:1 unless:2 iv:1 refutation:1 desired:4 theoretical:1 psychological:5 increased:2 instance:2 modeling:2 earlier:2 soft:1 disadvantage:1 maximization:1 ordinary:2 clipping:1 addressing:1 uniform:2 delay:2 successful:2 johnson:2 erlbaum:1 too:5 providence:1 varies:1 accomplish:1 synthetic:1 cho:1 chooses:2 density:1 international:1 ie:1 vijayakumar:6 probabilistic:1 factoranalysis:1 together:1 quickly:1 intersecting:1 again:1 von:1 reflect:1 containing:2 choose:1 slowly:1 thayer:2 henceforth:1 worse:2 glow:1 cognitive:1 expert:2 american:2 leading:2 account:9 lookup:1 summarized:2 includes:2 coefficient:8 automation:1 caused:1 ranking:3 onset:5 chemometric:1 ad:1 performed:2 try:2 view:2 later:1 aca:1 francis:6 parallel:4 thp:1 contribution:2 square:11 il:1 degraded:2 air:1 kaufmann:1 phoneme:17 chart:3 variance:7 correspond:1 ofthe:1 gathered:1 characteristic:1 sy:2 identification:1 raw:1 produced:1 carlo:5 comp:1 cc:2 confirmed:1 drive:1 mrc:1 anns:2 explain:2 suffers:1 psycholinguistic:2 touretzky:2 ed:4 definition:6 inexpensive:1 against:2 sensorimotor:1 pp:9 involved:1 regress:1 frequency:12 lebiere:2 associated:2 mi:1 modeler:1 dataset:2 recall:1 f3j:1 ut:1 dimensionality:21 ubiquitous:1 knowledge:1 loin:1 agreed:1 actually:1 recreates:1 appears:2 feed:6 ta:1 dt:1 violating:2 day:1 follow:1 response:1 maximally:1 higher:2 leen:1 wold:2 strongly:1 generality:1 just:2 stage:1 until:1 correlation:1 working:1 hand:3 christopher:1 trust:2 nonlinear:4 eig:1 widespread:1 defines:1 quality:1 perhaps:1 indicated:1 scientific:1 effect:14 hypothesized:1 consisted:1 true:3 brown:1 hence:1 spatially:1 read:1 moore:4 laboratory:1 iteratively:1 syndicate:1 illustrated:1 indistinguishable:1 during:1 ambiguous:1 noted:1 criterion:4 coda:3 presenting:1 outline:1 complete:2 performs:3 common:2 superior:1 kawato:1 functional:1 empirically:1 rust:1 jp:2 association:2 significant:1 refer:1 measurement:1 cambridge:2 lwpca:10 ai:2 everitt:2 wpcr:3 uv:1 outlined:2 unifor:1 language:2 had:1 robot:1 similarity:2 rastle:11 longer:2 etc:1 base:1 add:1 playa:1 multivariate:1 dust:1 recent:2 showed:1 perspective:2 irrelevant:1 optimizes:1 jolla:2 tesauro:1 mint:1 certain:2 route:22 outperforming:2 vt:1 meeting:1 inconsistency:1 atkins:1 captured:2 preserving:1 ministry:1 seen:1 morgan:1 greater:2 ey:1 prune:1 surely:1 eo:1 monotonically:1 july:1 ii:14 semi:1 afterwards:1 ofy:1 sound:1 reduces:1 kyoto:1 signal:2 violate:2 mix:1 faster:1 adapt:1 calculation:1 offer:2 dept:1 jft:1 cross:2 divided:1 naming:15 serial:6 equally:4 award:1 controlled:1 prediction:1 regression:41 basic:1 circumstance:1 metric:2 essentially:1 cmu:2 titech:2 iteration:3 kernel:5 psychologically:2 lwpls_i:3 orthography:4 expectation:2 robotics:1 irregular:9 nonword:2 addition:1 affecting:1 separately:3 whereas:2 invert:1 achieved:2 background:1 source:1 doursat:1 south:1 subject:5 tend:2 member:1 contrary:1 seem:1 jordan:4 odds:1 extracting:1 noting:1 counting:1 feedforward:1 iii:3 easy:1 concerned:1 automated:1 variety:1 variate:1 zi:1 fit:1 architecture:2 identified:1 l09:1 psychology:3 perfectly:4 suboptimal:1 angeles:1 xqf:1 pca:1 bartlett:1 soraku:1 york:2 jj:1 latency:20 proportionally:1 gpc:7 eigenvectors:2 clear:3 cga:1 involve:1 nonparametric:1 amount:1 band:1 locally:23 processed:1 mcclelland:14 reduced:1 generate:2 http:3 inhibitory:1 estimated:1 neuroscience:1 wr:2 per:1 fulfilled:2 delta:2 xii:1 affected:2 ist:1 group:1 four:2 threshold:1 nevertheless:1 falling:1 cutoff:1 undertaken:1 wand:1 run:1 inverse:1 letter:7 kucera:6 powerful:1 izations:1 almost:3 electronic:1 ofm:1 summarizes:1 purported:1 fl:2 layer:3 hom:2 followed:3 syllable:41 correspondence:4 quadratic:7 adapted:1 occur:1 precisely:1 orthogonality:1 ri:1 sake:1 simulate:1 performing:1 format:1 edul:1 department:1 influential:1 according:2 alternate:2 combination:1 aerial:1 cascadecorrelation:1 across:1 em:5 slightly:2 increasingly:1 wi:7 appealing:2 tw:4 making:1 honour:1 explained:3 projecting:1 interference:1 taken:1 computationally:1 equation:5 agree:2 remains:2 abbreviated:1 uyu:1 mechanism:1 fail:1 count:3 german:2 fed:1 end:3 pursuit:1 available:1 prerequisite:1 yare:1 hierarchical:1 appropriate:4 robustly:1 anymore:1 alternative:2 robustness:4 batch:1 slower:1 denotes:3 assumes:4 ensure:3 cf:1 include:1 calculating:1 exploit:1 phonology:3 scholarship:1 especially:1 hikaridai:1 hypercube:1 objective:1 added:5 quantity:1 already:2 occurs:2 parametric:1 degrades:1 nonuniformly:1 spelling:1 diagonal:3 interacts:1 guessing:1 exhibit:1 subspace:2 distance:4 separate:1 link:1 atr:1 ot2:1 majority:2 nx:1 gun:1 originate:1 sethu:3 chris:1 considers:1 consensus:1 assuming:2 length:2 besides:1 ratio:5 convincing:1 providing:1 difficult:2 nonintuitive:1 frank:5 negative:1 rise:1 proper:1 zt:1 minim:1 perform:5 uad:1 displayed:2 defining:1 hinton:3 excluding:1 precise:1 extended:1 lb:1 arbitrary:1 edi:1 pair:1 required:1 engine:1 california:3 learned:2 deletion:1 unequal:2 omits:1 conflicting:1 address:2 able:2 suggested:1 bar:1 usually:3 proceeds:1 perception:2 below:3 pattern:1 scott:2 reading:15 challenge:1 max:2 memory:1 explanation:4 belief:1 including:1 dict:1 difficulty:1 force:1 examination:1 residual:1 lwf:4 scheme:1 improve:1 technology:2 imply:1 inversely:1 ne:1 created:1 acknowledges:1 extract:1 xq:2 text:1 prior:1 understanding:2 literature:1 review:6 epoch:4 xtt:1 contributing:1 determining:1 fully:1 expect:2 loss:1 limitation:1 tures:1 geoffrey:1 localized:2 outlining:1 validation:2 foundation:4 rubin:2 principle:5 editor:1 row:3 prone:1 hopeless:1 summary:5 accounted:2 surprisingly:1 last:1 fahlman:2 supported:2 english:8 soon:1 bias:2 formal:1 allow:1 institute:2 neighbor:6 wide:1 face:1 taking:2 lwfa:8 distributed:5 tolerance:1 f31:1 calculated:5 dimension:4 computes:1 forward:6 equated:1 made:1 san:2 collection:1 projected:3 atkeson:14 author:1 avg:1 universally:1 observable:1 kuh:1 ml:1 global:2 active:1 question:1 corpus:2 assumed:2 consonant:7 xi:7 alternatively:1 zoned:1 continuous:1 iterative:2 latent:2 seidenberg:6 search:1 why:3 table:5 additionally:1 nature:1 ku:1 robust:5 ca:4 obtaining:1 interact:1 warranted:1 mse:7 interpolating:1 necessarily:1 european:1 domain:1 japanese:1 did:2 spread:1 gani:1 noise:40 f3q:2 repeated:1 body:1 xu:3 representative:2 georgia:1 fashion:1 oflinear:1 garrison:1 slow:1 wiley:2 experienced:1 position:23 momentum:2 refutes:1 sub:1 sf:3 debated:1 lie:1 candidate:4 weighting:4 third:1 late:4 learns:1 young:1 theorem:1 down:1 xt:2 explored:2 decay:1 evidence:3 grouping:2 lwpcr:8 consist:2 exists:2 false:1 sequential:1 magnitude:3 hole:1 gap:1 locality:1 entropy:1 simply:2 likely:1 welsch:1 explore:1 conveniently:1 adjustment:1 ux:1 lwpls:15 loses:1 relies:1 ma:1 conditional:2 fringe:1 identity:1 formulated:1 viewed:1 ann:2 consequently:1 towards:4 considerable:1 change:2 specifically:1 except:2 determined:2 uniformly:1 averaging:3 operates:1 principal:9 total:1 belsley:1 stereotypic:1 called:1 experimental:2 hnb:1 subdivision:1 la:3 vote:1 secondary:1 underestimation:1 exception:18 pint:1 college:1 support:4 alexander:1 filler:1 investigator:1 constructive:1 evaluate:5 tested:1 phenomenon:6 ex:5 |
427 | 139 | 11
AN OPTIMALITY PRINCIPLE FOR
UNSUPERVISED LEARNING
Terence D. Sanger
MIT AI Laboratory, NE43-743
Cambridge, MA 02139
(tds@wheaties.ai.mit.edu)
ABSTRACT
We propose an optimality principle for training an unsupervised feedforward neural network based upon maximal
ability to reconstruct the input data from the network outputs. We describe an algorithm which can be used to train
either linear or nonlinear networks with certain types of
nonlinearity. Examples of applications to the problems of
image coding, feature detection, and analysis of randomdot stereograms are presented.
1. INTRODUCTION
There are many algorithms for unsupervised training of neural networks, each of
which has a particular optimality criterion as its goal. (For a partial review, see
(Hinton, 1987, Lippmann, 1987).) We have presented a new algorithm for training
single-layer linear networks which has been shown to have optimality properties
associated with the Karhunen-Loeve expansion (Sanger, 1988b). We now show
that a similar algorithm can be applied to certain types of nonlinear feedforward
networks, and we give some examples of its behavior.
The optimality principle which we will use to describe the algorithm is based on the
idea of maximizing information which was first proposed as a desirable property of
neural networks by Linsker (1986, 1988). Unfortunately, measuring the information
in network outputs can be difficult without precise knowledge of the distribution
on the input data, so we seek another measure which is related to information
but which is easier to compute. If instead of maximizing information, we try to
maximize our ability to reconstruct the input (with minimum mean-squared error)
given the output of the network, we are able to obtain some useful results. Note
that this is not equivalent to maximizing information except in some special cases.
However, it contains the intuitive notion that the input data is being represented
by the network in such a way that very little of it has been "lost".
12
Sanger
2. LINEAR CASE
We now summarize some of the results in (Sanger, 1988b). A single-layer linear
feedforward network is described by an M xN matrix C of weights such that if x is
a vector of N inputs and y is a vector of M outputs with M < N, we have y Cx.
As mentioned above, we choose an optimality principle defined so that we can best
reconstruct the inputs to the network given the outputs. We want to minimize the
mean squared error E[(x - x)2] where x is the actual input which is zero-mean with
correlation matrix Q E[xxT], and x is a linear estimation of this input given the
output y. The linear least squares estimate (LLSE) is given by
=
=
and we will assume that x is computed in this way for any matrix C of weights
which we choose. The mean-squared error for the LLSE is given by
and it is well known that this is minimized if the rows of C are a linear combination of
the first M eigenvectors of the correlation matrix Q. One optimal choice of C is the
Singular Value Decomposition (SVD) of Q, for which the output correlation matrix
E[yyT] CQCT will be the diagonal matrix of eigenvalues of Q. In this case, the
outputs are uncorrelated and the sum of their variances (traceE[yyT]) is maximal for
any set of M un correlated outputs. We can thus think of the eigenvectors as being
obtained by any process which maximizes the output variance while maintaining
the outputs uncorrelated.
=
We now define the optimal single-layer linear network as that network whose weights
represent the first M eigenvectors of the input correlation matrix Q. The optimal
network thus minimizes the mean-squared approximation error E[(x - x)2] given
the shape constraint that M < N.
2.1 LINEAR ALGORITHM
We have previously proposed a weight-update rule called the "Generalized Hebbian
Algorithm" , and proven that this algorithm causes the rows of the weight matrix C
to converge to the eigenvectors of the input correlation matrix Q (Sanger, 1988a,b).
The algorithm is given by:
C(t
+ 1) = C(t) + I
(y(t)xT(t) - LT[y(t)yT(t)]C(t?)
(1)
where I is a rate constant which decreases as l/t, x(t) is an input sample vector,
yet) = C(t)x(t), and LTD is an operator which makes its matrix argument lower
triangular by setting all entries above the diagonal to zero. This algorithm can be
implemented using only a local synaptic learning rule (Sanger, 1988b).
Since the Generalized Hebbian Algorithm computes the eigenvectors of the input
correlation matrix Q, it is related to the Singular Value Decomposition (SVD),
An Optimality Principle for Unsupervised Learning
c.
Figure 1: (aJ original image. (bJ image coded at .36 bits per pixel. (cJ masks
learned by the network which were used for vector quantized coding of 8x8 blocks of
the image.
Principal Components Analysis (PCA), and the Karhunen-Loeve Transform (KLT).
(For a review of several related algorithms for performing the KLT, see (Oja, 1983).)
2.2 IMAGE CODING
We present one example of the behavior of a single-layer linear network. (This
example appears in (Sanger, 1988b).) Figure 1a shows an original 256x256x8bit
image which was used for training a network. 8x8 blocks of the image were chosen
by scanning over the image, and these were used as training inputs to a network with
64 inputs and 8 outputs. After training, the set of weights for each output (figure
lc) represents a vector quantizing mask. Each 8x8 block of the input image is then
coded using the outputs of the network. Each output is quantized with a number of
bits related to the log of the variance, and the original figure is approximated from
the quantized outputs. The reconstruction of figure 1b uses a total of 23 bits per 8x8
region, which gives a data rate of 0.36 bits per pixel. The fact that the image could
be represented using such a low bit rate indicates that the masks that were found
represent significant features which are useful for recognition. This image coding
technique is equivalent to block-coded KLT methods common in the literature.
13
14
Sanger
3. NONLINEAR CASE
In general, training a nonlinear unsupervised network to approximate nonlinear
functions is very difficult. Because of the large (infinite-dimensional) space of possible functions, it is important to have detailed knowledge of the class of functions
which are useful in order to design an efficient network algorithm. (Several people pointed out to me that the talk implied such knowledge is not necessary, but
unfortunately such an implication is false.)
The network structure we consider is a linear layer represented by a matrix C (which
is perhaps an interior layer of a larger network) followed by node nonlinearities (1'(Yi)
where Yi is the ith linear output, followed by another linear layer (perhaps followed
by more layers). We assume that the nonlinearities (1'0 are fixed, and that the only
parameters susceptible to training are the linear weights C.
If z is the M-vector of outputs after the nonlinearity, then we can write each component Zi = (1'(Yi)
(1'(CiX) where Ci is the ith row of the matrix C. Note that
the level contours of each function Zi are determined entirely by the vector Ci, and
that the effect of (1'0 is limited to modifying the output value. Intuitively, we thus
expect that if Yi encodes a useful parameter of the input x, then Zi will encode the
same parameter, although scaled by the nonlinearity (1'0.
=
This can be formalized, and if we choose our optimality principle to again be minimum mean-squared linear approximation of the original input x given the output
z, the best solution remains when the rows of C are a linear combination of the first
M eigenvectors of the input correlation matrix Q (Bourlard and Kamp, 1988) .
In two of the simulations, the nonlinearity (1'0 which we use is a rectification nonlinearity, given by
Yi if Yi 20
(1'(Yd { 0 if Yi <0
=
Note that at most one of {(1'(Yi), (1'( -Yi)} is nonzero at any time, so these two values
are uncorrelated. Therefore, if we maximize the variance of y (before the nonlinearity) while maintaining the elements of Z (after the nonlinearity) uncorrelated,
we need 2M outputs in order to represent the data available from an M-vector y.
Note that 2M may be greater than the number of inputs N, so that the "hidden
layer" Z can have more elements than the input.
3.1 NONLINEAR ALGORITHM
The nonlinear Generalized Hebbian Algorithm has exactly the same form as for
the linear case, except that we substitute the use of the output values after the
nonlinearity for the linear values. The algorithm is thus given by:
C(t
+ 1) = C(t) + 'Y (z(t)xT(t) -
LT[z(t)zT(t)]C(t))
where the elements of z are given by Zi(t) = (1'(Yi(t)), with y(t)
(2)
= C(t)x(t).
Although we have not proven that this algorithm converges, a heuristic analysis
of its behavior (for a rectification nonlinearity and Gaussian input distribution)
An Optimality Principle for Unsupervised Learning
shows that stable points may exist for which each row of C is proportional to
an eigenvector of Q, and pairs of rows are either the negative of each other or
orthogonal. In practice, the rows of C are ordered by decreasing output variance,
and occur in pairs for which one member is the negative of the other. This choice
of C will maximize the sum of the output variances for uncorrelated outputs, so
long as the input is Gaussian. It also allows optimal linear estimation of the input
given the output, so long as both polarities of each of the eigenvectors are present.
3.2 NONLINEAR EXAMPLES
3.2.1 Encoder Problem
We compare the performance of two nonlinear networks which have learned to
perform an identity mapping (the "encoder" problem). One is trained by backpropagation, (Rumelhart et a/., 1986) and the other has two hidden layers trained
using the unsupervised Hebbian algorithm, while the output layer is trained using
a supervised LMS algorithm (Widrow and Hoff, 1960). The network has 5 inputs,
two hidden layers of 3 units each, and 5 outputs. There is a sigmoid nonlinearity at
each hidden layer, but the thresholds are all kept at zero. The task is to minimize
the mean-squared difference between the inputs and the outputs. The input is a
zero-mean correlated Gaussian random 5-vector, and both algorithms are presented
with the same sequence of inputs. The unsupervised-trained network converged to a
steady state after 1600 examples, and the backpropagation network converged after
2400 (convergence determined by no further decrease in average error). The RMS
error at steady state was 0.42 for both algorithms (this figure should be compared to
the sum of the variances of the inputs, which was 5.0). Therefore, for this particular
task, there is no significant difference in performance between backpropagation and
the Generalized Hebbian Algorithm. This is an encouraging result, since if we can
use an unsupervised algorithm to solve other problems, the training time will scale
at most linearly with the number of layers.
3.2.2 Nonlinear Receptive Fields
Several investigators have shown that Hebbian algorithms can discover useful image
features related to the receptive fields of cells in primate visual cortex (see for
example (Bienenstock et a/., 1982, Linsker, 1986, Barrow, 1987?. One of the more
recent methods uses an algorithm very similar to the one proposed here to find the
principal component of the input (Linsker, 1986). We performed an experiment to
find out what types of nonlinear receptive fields could be learned by the Generalized
Hebbian Algorithm if provided with similar input to that used by Linsker.
We used a single-layer nonlinear network with 4096 inputs arranged in a 64x64
grid, and 16 outputs with a rectification nonlinearity. The input data consisted of
images of low-pass filtered white Gaussian noise multiplied by a Gaussian window.
After 5000 samples, the 16 outputs learned the masks shown in figure 2. These
masks possess qualitative similarity to the receptive fields of cells found in the visual
cortex of cat and monkey (see for example (Andrews and Pollen, 1979?. They are
equivalent to the masks learned by a purely linear network (Sanger, 1988b), except
that both positive and negative polarities of most mask shapes are present here.
15
16
Sanger
Figure 2: Nonlinear receptive fields ordered from left-to-right and top-to-bottom.
3.2.3 Stereo
We now show how the nonlinear Generalized Hebbian Algorithm can be used to
train a two-layer network to detect disparity edges. The network has 128 inputs,
8 types of unit in the hidden layer with a rectification nonlinearity, and 4 types of
output unit. A 128x128 pixel random-dot stereo pair was generated in which the
left half had a disparity of two pixels, and the right half had zero disparity. The
image was convolved with a vertically-oriented elliptical Gaussian mask to remove
high-frequency vertical components. Corresponding 8x8 blocks of the left and right
images (64 pixels from each image) were multiplied by a Gaussian window function
and presented as input to the network, which was allowed to learn the first layer
according to the unsupervised algorithm. After 4000 iterations, the first layer had
converged to a set of 8 pairs of masks. These masks were convolved with the images
(the left mask was convolved with the left image, and the right mask with the right
image, and the two results were summed and rectified) to produce a pattern of
activity at the hidden layer. (Although there were only 8 types of hidden unit, we
now allow one of each type to be centered at every input image location to obtain a
pattern of total activity.) Figure 3 shows this activity, and we can see that the last
four masks are disparity-sensitive since they respond preferentially to either the 2
pixel disparity or the zero disparity region of the image.
An Optimality Principle for Unsupervised Learning
Figure 3: Hidden layer response for a two-layer nonlinear network trained on stereo
images. The left half of the input random dot image has a 2 pixel disparity, and the
right half has zero disparity.
Figure 4: Output layer response for a two-layer nonlinear network trained on stereo
Images.
Since we were interested in disparity, we trained the second layer using only the last
four hidden unit types. The second layer had 1024 (=4x16x16) inputs organized as
a 16x16 receptive field in each of the four hidden unit "planes". The outputs did not
have any nonlinearity. Training was performed by scanning over the hidden unit
activity pattern (successive examples overlapped by 8 pixels) and 6000 iterations
were used to produce the second-layer weights. The masks that were learned were
then convolved with the hidden unit activity pattern to produce an output unit
activity pattern, shown in figure 4.
The third output is clearly sensitive to a change in disparity (a depth edge). If we
generate several different random-dot stereograms and average the output results,
17
18
Sanger
Figure 5: Output layer response averaged over ten stereograms with a central 2 pixel
disparity square and zero disparity surround.
we see that the other outputs are also sensitive (on average) to disparity changes, but
not as much as the third. Figure 5 shows the averaged response to 10 stereograms
with a central 2 pixel disparity square against a zero disparity background. Note
that the ability to detect disparity edges requires the rectification nonlinearity at
the hidden layer, since no linear function has this property.
4. CONCLUSION
We have shown that the unsupervised Generalized Hebbian Algorithm can produce
useful networks. The algorithm has been proven to converge only for single-layer
linear networks. However, when applied to nonlinear networks with certain types
of nonlinearity, it appears to converge to good results. In certain cases, it operates
by maintaining the outputs uncorrelated while maximizing their variance. We have
not investigated its behavior on nonlinearities other than rectification or sigmoids,
so we can make no predictions about its general utility. Nevertheless, the few
examples presented for the nonlinear case are encouraging, and suggest that further
investigation of this algorithm will yield interesting results.
Acknowledgements
I would like to express my gratitude to the many people at the NIPS conference
and elsewhere whose comments, criticisms, and suggestions have increased my understanding of these results. In particular, thanks are due to Ralph Linsker for
pointing out to me an important error in the presentation and for his comments on
the manuscript, as well as to John Denker, Steve Nowlan, Rich Sutton, Tom Breuel,
and my advisor Tomaso Poggio.
This report describes research done at the MIT Artificial Intelligence Laboratory,
and sponsored by a grant from the Office of Naval Research (ONR), Cognitive
and Neural Sciences Division; by the Alfred P. Sloan Foundation; by the National
Science Foundation; by the Artificial Intelligence Center of Hughes Aircraft Corporation (SI-801534-2); and by the NATO Scientific Affairs Division (0403/87).
Support for the A. I. Laboratory's artificial intelligence research is provided by the
An Optimality Principle for Unsupervised Learning
Advanced Research Projects Agency of the Department of Defense under Army contract DACA76-85-C-0010, and in part by ONR contract N00014-85-K-0124. The
author was supported during part of this research by a National Science Foundation
Graduate fellowship, and later by a Medical Scientist Training Program grant.
References
Andrews B. W., Pollen D. A., 1979, Relationship between spatial frequency selectivity and receptive field profile of simple cells, J. Physiol., 287:163-176.
Barrow H. G., 1987, Learning receptive fields, In Pmc. IEEE 1st Ann. Conference
on Neural Networks, volume 4, pages 115-121, San Diego, CA.
Bienenstock E. L., Cooper L. N., Munro P. W., 1982, Theory for the development
of neuron selectivity: Orientation specificity and binocular interaction in visual
cortex, J. Neuroscience, 2(1):32-48.
Bourlard H., Kamp Y., 1988, Auto-association by multilayer perceptrons and singular value decomposition, Biological Cybernetics, 59:291-294.
Hinton G. E., 1987, Connectionist learning procedures, CMU Tech. Report CS-87115.
Linsker R., 1986, From basic network principles to neural architecture, Proc. Natl.
Acad. Sci. USA, 83:7508-7512.
Linsker R., 1988, Self-organization in a perceptual network, Computer, 21(3):105117.
Lippmann R. P., 1987, An introduction to computing with neural nets, IEEE A SSP
Magazine, pages 4-22.
Oja E., 1983, Subspace Methods of Pattern Recognition, Research Studies Press,
UK.
Rumelhart D. E., Hinton G. E., Williams R. J., 1986, Learning representations by
back-propagating errors, Nature, 323(9):533-536.
Sanger T. D., 1988a, Optimal unsupervised learning, Neural Networks, 1(Sl):127,
Proc. 1st Ann. INNS meeting, Boston, MA.
Sanger T. D., 1988b, Optimal unsupervised learning in a single-layer linear feedforward neural network, submitted to Neural Networks.
Widrow B., Hoff M. E., 1960, Adaptive switching circuits, In IRE WESCON Conv.
Record, Part 4, pages 96-104.
19
| 139 |@word aircraft:1 simulation:1 seek:1 decomposition:3 contains:1 disparity:16 elliptical:1 nowlan:1 si:1 yet:1 john:1 physiol:1 shape:2 remove:1 sponsored:1 update:1 half:4 intelligence:3 plane:1 affair:1 ith:2 record:1 filtered:1 ire:1 quantized:3 node:1 location:1 successive:1 x128:1 qualitative:1 mask:14 tomaso:1 behavior:4 decreasing:1 little:1 actual:1 encouraging:2 window:2 conv:1 provided:2 discover:1 project:1 maximizes:1 circuit:1 what:1 minimizes:1 eigenvector:1 monkey:1 corporation:1 every:1 exactly:1 scaled:1 uk:1 unit:9 grant:2 medical:1 before:1 positive:1 scientist:1 local:1 vertically:1 acad:1 sutton:1 switching:1 yd:1 limited:1 graduate:1 averaged:2 lost:1 block:5 practice:1 hughes:1 backpropagation:3 procedure:1 specificity:1 suggest:1 interior:1 operator:1 equivalent:3 yt:1 maximizing:4 center:1 williams:1 formalized:1 rule:2 his:1 notion:1 x64:1 diego:1 magazine:1 us:2 overlapped:1 element:3 rumelhart:2 approximated:1 recognition:2 bottom:1 region:2 decrease:2 mentioned:1 ne43:1 agency:1 stereograms:4 trained:7 purely:1 upon:1 division:2 represented:3 cat:1 talk:1 xxt:1 train:2 describe:2 artificial:3 whose:2 heuristic:1 larger:1 solve:1 reconstruct:3 triangular:1 ability:3 encoder:2 think:1 transform:1 sequence:1 eigenvalue:1 quantizing:1 net:1 breuel:1 inn:1 propose:1 reconstruction:1 interaction:1 maximal:2 intuitive:1 randomdot:1 convergence:1 produce:4 converges:1 widrow:2 andrew:2 propagating:1 implemented:1 c:1 modifying:1 centered:1 investigation:1 biological:1 daca76:1 mapping:1 bj:1 lm:1 pointing:1 estimation:2 proc:2 sensitive:3 mit:3 clearly:1 gaussian:7 office:1 encode:1 klt:3 naval:1 indicates:1 tech:1 criticism:1 detect:2 hidden:13 bienenstock:2 interested:1 pixel:10 ralph:1 orientation:1 development:1 spatial:1 special:1 summed:1 hoff:2 field:8 tds:1 represents:1 unsupervised:15 linsker:7 minimized:1 report:2 connectionist:1 few:1 oriented:1 oja:2 national:2 detection:1 organization:1 natl:1 implication:1 edge:3 partial:1 necessary:1 poggio:1 orthogonal:1 increased:1 advisor:1 measuring:1 entry:1 scanning:2 my:3 thanks:1 st:2 contract:2 terence:1 squared:6 again:1 central:2 choose:3 cognitive:1 nonlinearities:3 coding:4 sloan:1 performed:2 try:1 later:1 minimize:2 square:3 variance:8 yield:1 kamp:2 rectified:1 cybernetics:1 converged:3 submitted:1 synaptic:1 against:1 frequency:2 associated:1 knowledge:3 yyt:2 cj:1 organized:1 back:1 appears:2 manuscript:1 steve:1 supervised:1 tom:1 response:4 arranged:1 done:1 binocular:1 correlation:7 nonlinear:18 aj:1 perhaps:2 scientific:1 usa:1 effect:1 consisted:1 laboratory:3 nonzero:1 white:1 during:1 self:1 steady:2 criterion:1 generalized:7 image:24 common:1 sigmoid:1 volume:1 association:1 significant:2 cix:1 cambridge:1 surround:1 ai:2 grid:1 pointed:1 nonlinearity:15 had:4 dot:3 stable:1 cortex:3 similarity:1 recent:1 selectivity:2 certain:4 n00014:1 onr:2 meeting:1 yi:10 minimum:2 greater:1 converge:3 maximize:3 desirable:1 llse:2 hebbian:9 long:2 coded:3 prediction:1 basic:1 multilayer:1 cmu:1 iteration:2 represent:3 cell:3 background:1 fellowship:1 want:1 singular:3 posse:1 comment:2 member:1 feedforward:4 zi:4 architecture:1 idea:1 pca:1 rms:1 utility:1 defense:1 munro:1 ltd:1 stereo:4 cause:1 useful:6 detailed:1 eigenvectors:7 ten:1 generate:1 sl:1 exist:1 neuroscience:1 per:3 alfred:1 write:1 express:1 four:3 threshold:1 nevertheless:1 kept:1 sum:3 respond:1 cqct:1 bit:5 entirely:1 layer:31 followed:3 activity:6 occur:1 constraint:1 encodes:1 loeve:2 argument:1 optimality:11 performing:1 department:1 according:1 combination:2 describes:1 primate:1 intuitively:1 rectification:6 previously:1 remains:1 available:1 multiplied:2 denker:1 convolved:4 original:4 substitute:1 top:1 maintaining:3 sanger:13 implied:1 receptive:8 diagonal:2 ssp:1 subspace:1 sci:1 me:2 polarity:2 relationship:1 preferentially:1 pmc:1 difficult:2 unfortunately:2 susceptible:1 negative:3 design:1 zt:1 perform:1 vertical:1 neuron:1 barrow:2 hinton:3 precise:1 gratitude:1 pair:4 pollen:2 learned:6 nip:1 able:1 pattern:6 summarize:1 program:1 bourlard:2 advanced:1 x8:5 auto:1 review:2 literature:1 acknowledgement:1 understanding:1 expect:1 interesting:1 suggestion:1 proportional:1 proven:3 foundation:3 principle:10 uncorrelated:6 row:7 elsewhere:1 supported:1 last:2 allow:1 depth:1 xn:1 contour:1 computes:1 rich:1 author:1 adaptive:1 san:1 approximate:1 lippmann:2 nato:1 wescon:1 un:1 learn:1 nature:1 ca:1 expansion:1 investigated:1 did:1 linearly:1 noise:1 profile:1 allowed:1 x16:1 cooper:1 lc:1 perceptual:1 third:2 xt:2 false:1 ci:2 sigmoids:1 karhunen:2 easier:1 boston:1 cx:1 lt:2 army:1 visual:3 ordered:2 ma:2 goal:1 identity:1 presentation:1 ann:2 change:2 infinite:1 except:3 determined:2 operates:1 principal:2 total:2 called:1 pas:1 svd:2 perceptrons:1 people:2 support:1 investigator:1 correlated:2 |
428 | 1,390 | Online learning from finite training sets
in nonlinear networks
David Barber t
Peter Sollich*
Department of Physics
University of Edinburgh
Edinburgh ERg 3JZ, U.K.
Department of Applied Mathematics
Aston University
Birmingham B4 7ET, U.K.
P.Sollich~ed.ac.uk
D.Barber~aston . ac.uk
Abstract
Online learning is one of the most common forms of neural network training. We present an analysis of online learning from finite
training sets for non-linear networks (namely, soft-committee machines), advancing the theory to more realistic learning scenarios.
Dynamical equations are derived for an appropriate set of order
parameters; these are exact in the limiting case of either linear
networks or infinite training sets. Preliminary comparisons with
simulations suggest that the theory captures some effects of finite
training sets, but may not yet account correctly for the presence of
local minima.
1
INTRODUCTION
The analysis of online gradient descent learning, as one of the most common forms
of supervised learning, has recently stimulated a great deal of interest [1, 5, 7, 3]. In
online learning, the weights of a network ('student') are updated immediately after
presentation of each training example (input-output pair) in order to reduce the
error that the network makes on that example. One of the primary goals of online
learning analysis is to track the resulting evolution of the generalization error - the
error that the student network makes on a novel test example, after a given number
of example presentations. In order to specify the learning problem, the training
outputs are assumed to be generated by a teacher network of known architecture.
Previous studies of online learning have often imposed somewhat restrictive and
? Royal Society Dorothy Hodgkin Research Fellow
tSupported by EPSRC grant GR/J75425: Novel Developments in Learning Theory for
Neural Networks
P. SolIich and D. Barber
358
unrealistic assumptions about the learning framework. These restrictions are, either
that the size of the training set is infinite, or that the learning rate is small[l, 5, 4].
Finite training sets present a significant analytical difficulty as successive weight
updates are correlated, giving rise to highly non-trivial generalization dynamics.
For linear networks, the difficulties encountered with finite training sets and noninfinitesimal learning rates can be overcome by extending the standard set of descriptive ('order') parameters to include the effects of weight update correlations[7].
In the present work, we extend our analysis to nonlinear networks. The particular
model we choose to study is the soft-committee machine which is capable of representing a rich variety of input-output mappings. Its online learning dynamics has
been studied comprehensively for infinite training sets[l, 5]. In order to carry out
our analysis, we adapt tools originally developed in the statistical mechanics literature which have found application, for example, in the study of Hopfield network
dynamics[2].
2
MODEL AND OUTLINE OF CALCULATION
For an N-dimensional input vector x, the output of the soft committee machine is
given by
(I)
where the nonlinear activation function g(hl ) = erf(hz/V2) acts on the activations
hi = wtxl.JFi (the factor 1/.JFi is for convenience only). This is a neural network
with L hidden units, input to hidden weight vectors WI, 1 = I..L, and all hidden to
output weights set to 1.
In online learning the student weights are adapted on a sequence of presented examples to better approximate the teacher mapping. The training examples are drawn,
with replacement, from a finite set, {(X/",yl-') ,j.t I..p}. This set remains fixed
piN.
during training. Its size relative to the input dimension is denoted by a
We take the input vectors xl-' as samples from an N dimensional Gaussian distribution with zero mean and unit variance. The training outputs y'" are assumed to
be generated by a teacher soft committee machine with hidden weight vectors w~,
m = I..M, with additive Gaussian noise corrupting its activations and output.
=
=
The discrepancy between the teacher and student on a particular training example (x, y), drawn from the training set, is given by the squared difference of their
corresponding outputs,
E=
H~9(hl) -yr = H~9(hl) - ~g(km +em) -eor
where the student and teacher activations are, respectively
h,
em,
and
m = I..M and
output respectively.
= {J;wtx
km
= {J;(w:n?x,
(2)
eo are noise variables corrupting the teacher activations and
Given a training example (x, y), the student weights are updated by a gradient
descent step with learning rate "I,
w; - W, = -"I\1wIE = - JNx8h E
l
(3)
359
On-line Learning from Finite Training Sets in Nonlinear Networks
The generalization error is defined to be the average error that the student makes on
a test example selected at random (and uncorrelated with the training set), which
we write as ?g = (E).
Although one could, in principle, model the student weight dynamics directly, this
will typically involve too many parameters, and we seek a more compact representation for the evolution of the generalization error. It is straightforward to show that
the generalization error depends, not on a detailed description of all the network
weights, but only on the overlap parameters Qll' = ~ W WI' and Rim = ~ W w':n
[1, 5, 7]. In the case of infinite 0, it is possible to obtain a closed set of equations
governing the overlap parameters Q, R [5]. For finite training sets, however, this is
no longer possible, due to the correlations between successive weight updates[7].
r
r
In order to overcome this difficulty, we use a technique developed originally to study
statistical physics systems [2] . Initially, consider the dynamics of a general vector of
order parameters, denoted by 0, which are functions of the network weights w. If
the weight updates are described by a transition probability T(w -+ w'), then an
approximate update equation for 0 is
0' - 0 = IfdW' (O(w') - O(w)) T(w -+
\
W'))
(4)
P(w)oc6(O(w)-O)
Intuitively, the integral in the above equation expresses the average change l of 0
caused by a weight update w -+ w', starting from (given) initial weights w. Since
our aim is to develop a closed set of equations for the order parameter dynamics, we
need to remove the dependency on the initial weights w. The only information we
have regarding w is contained in the chosen order parameters 0, and we therefore
average the result over the 'subshell' of all w which correspond to these values of
the order parameters. This is expressed as the 8-function constraint in equation(4).
It is clear that if the integral in (4) depends on w only through O(w), then the
average is unnecessary and the resulting dynamical equations are exact. This is in
fact the case for 0 -+ 00 and 0 = {Q, R}, the standard order parameters mentioned
above[5]. If this cannot be achieved, one should choose a set of order parameters to
obtain approximate equations which are as close as possible to the exact solution.
The motivation for our choice of order parameters is based on the linear perceptron
case where, in addition to the standard parameters Q and R, the overlaps projected
onto eigenspaces of the training input correlation matrix A = ~ E:=l xl' (xl') T are
required 2 . We therefore split the eigenvalues of A into r equal blocks ('Y = 1 ... r)
containing N' = N Ir eigenvalues each, ordering the eigenvalues such that they
increase with 'Y. We then define projectors p'Y onto the corresponding eigenspaces
and take as order parameters:
'Y
R1m
_
-
1 Tp'Y ..
N'w,
wm
UI.'Y
-
~
Nt W,Tp'Yb
II
(5)
where the b B are linear combinations of the noise variables and training inputs,
(6)
1 Here we assume that the system size N is large enough that the mean values of the
parameters alone describe the dynamics sufficiently well (i. e., self-averaging holds).
2The order parameters actually used in our calculation for the linear perceptron[7] are
Laplace transforms of these projected order parameters.
P. Sollich and D. Barber
360
As
r
-+
00,
these order parameters become functionals of a continuous variable3 .
The updates for the order parameters (5) due to the weight updates (3) can be
found by taking the scalar products of (3) with either projected student or teacher
weights, as appropriate. This then introduces the following activation 'components',
k'Y
m
= VNi
ff(w* )Tp'"Yx
m
=
so that the student and teacher activations are h, = ~ E'"Y hi and km ~ E'"Y k~,
respectively. For the linear perceptron, the chosen order parameters form a complete
set - the dynamical equations close, without need for the average in (4).
For the nonlinear case, we now sketch the calculation of the order parameter update
equations (4). Taken together, the integral over Wi (a sum of p discrete terms in
our case, one for each training example) and the subshell average in (4), define
an average over the activations (2), their components (7), and the noise variables
~m, ~o. These variables turn out to be Gaussian distributed with zero mean, and
therefore only their covariances need to be worked out. One finds that these are in
fact given by the naive training set averages. For example,
=
(8)
where we have used p'"Y A = a'"YP'"Y with a'"Y 'the' eigenvalue of A in the ,-th
eigenspace; this is well defined for r -+ 00 (see [6] for details of the eigenvalue
spectrum). The correlations of the activations and noise variables explicitly appearing in the error in (3) are calculated similarly to give,
(h,h,,) =
~
L:; Q~,
'"Y
(h,km) =
~L
:; Rim
(9)
'"Y
(h,~s)
=
~ L ~U,~
'"Y
where the final equation defines the noise variances. The T~m' are projected overlaps between teacher weight vectors, T~m' = ~ (w~)Tp'"Yw:n,. We will assume that
the teacher weights and training inputs are uncorrelated, so that T~m' is independent of ,. The required covariances of the 'component' activations are
a'"YR'"Y
(kinh,)
(c] h,)
(hi h" )
a
'm
-
a'"YU'"Y
-
a'"YQ'"Y
a
a
ls
II'
(k~km')
=
a'"YT'"Y
(c]k m, )
-
0
-
a'"YR'"Y
(hJkm,)
a
a
mm'
'm
(k~~s)
-
0
(C]~8' )
-
a'"Y 2
-(7s588 ,
=
.!.U'"Y
(hJ~s)
a
a
's
(10)
3Note that the limit r -+ 00 is taken after the thermodynamic limit, i.e., r ~ N. This
ensures that the number of order parameters is always negligible compared to N (otherwise
self-averaging would break down).
On-line Learning from Finite Training Sets in Nonlinear Networks
0.03 r f I I : - - - - -........- - - - - - - ,
0.025 I
(a)
0.25
(b)
o
00
OOOOOOOOC
0.2
000000000000000000000000
0000000 00
0.02
L.. ...o~ooo
~
I
0.15
0.01
361
'------~-----~
o
t
50
\
0000
'NNNoaa oa
aaaoaaaaaaaaaaaaaaaac
,,------------
o
100
50
t
100
Figure 1: fg vs t for student and teacher with one hidden unit (L = M = 1);
a = 2, 3, 4 from above, learning rate "I = 1. Noise of equal variance was added to
both activations and output (a) O'~ = 0'5 = 0.01, (b) O'~ = 0'5= 0.1. Simulations
for N = 100 are shown by circles; standard errors are of the order of the symbol
size. The bottom dashed lines show the infinite training set result for comparison.
r = 10 was used for calculating the theoretical predictions; the curved marked "+"
in (b), with r = 20 (and a = 2), shows that this is large enough to be effectively in
the r -+ 00 limit.
Using equation (3) and the definitions (7), we can now write down the dynamical
equations, replacing the number of updates n by the continuous variable t = n/ N
in the limit N -+ 00:
-"I (k-:nOh,E)
OtRim
OtU?s
-"I (c~oh,E)
OtQIz,
-"I (h7 Oh" E) - "I
(h~ Oh, E) + "12 a-y (Oh,Eoh" E)
(11)
a
where the averages are over zero mean Gaussian variables, with covariances (9,10).
Using the explicit form of the error E, we have
oh,E = g'(h,) [L9(hl') - Lg(km
I'
+ em) -
eo]
(12)
m
which, together with the equations (11) completes the description of the dynamics.
The Gaussian averages in (11) can be straightforwardly evaluated in a manner
similar to the infinite training set case[5], and we omit the rather cumbersome
explicit form of the resulting equations.
We note that, in contrast to the infinite training set case, the student activations
hI and the noise variables C and
are now correlated through equation (10).
Intuitively, this is reasonable as the weights become correlated, during training,
with the examples in the training set. In calculating the generalization error, on the
other hand, such correlations are absent, and one has the same result as for infinite
training sets. The dynamical equations (11), together with (9,10) constitute our
main result. They are exact for the limits of either a linear network (R, Q, T -+ 0,
so that g(x) ex: x) or a -+ 00, and can be integrated numerically in a straightforward
way. In principle, the limit r -+ 00 should be taken but, as shown below, relatively
small values of r can be taken in practice.
s
3
es
RESULTS AND DISCUSSION
We now discuss the main consequences of our result (11), comparing the resulting
predictions for the generalization dynamics, fg(t), to the infinite training set theory
P. Sollich and D. Barber
362
k
(a)
0.4 ..----------~--~----,
0.25
02
. 100000000000000000000000
1
______________
0.3
0.15 ,
0.2
0.1
\
0.05
...
,'--
~
--- ---
0.1
O~--~------~----~~~
o
(b)
10
20
30
40
t
50
~ooooooooooooooooooo
o
W
100
1W t 200
OL---~----------~----~
Figure 2: ?g VS t for two hidden units (L = M = 2). Left: a = 0.5, with a = 00
shown by dashed line for comparison; no noise. Right: a = 4, no noise (bottom)
and noise on teacher activations and outputs of variance 0.1 (top). Simulations for
N = 100 are shown by small circles; standard errors are less than the symbol size.
Learning rate fJ = 2 throughout.
and to simulations. Throughout, the teacher overlap matrix is set to
(orthogonal teacher weight vectors of length V'ii).
Tij
=
c5ij
In figure(l), we study the accuracy of our method as a function of the training
set size for a nonlinear network with one hidden unit at two different noise levels.
The learning rate was set to fJ = 1 for both (a) and (b). For small activation
and output noise (0'2 = 0.01), figure(la) , there is good agreement with the simulations for a down to a = 3, below which the theory begins to underestimate
the generalization error, compared to simulations. Our finite a theory, however,
is still considerably more accurate than the infinite a predictions. For larger noise
(0'2 = 0.1, figure(lb?, our theory provides a reasonable quantitative estimate of the
generalization dynamics for a > 3. Below this value there is significant disagreement, although the qualitative behaviour of the dynamics is predicted quite well,
including the overfitting phenomenon beyond t ~ 10. The infinite a theory in this
case is qualitatively incorrect.
In the two hidden unit case, figure(2), our theory captures the initial evolution of
?g(t) very well, but diverges significantly from the simulations at larger t; nevertheless, it provides a considerable improvement on the infinite a theory. One reason for
the discrepancy at large t is that the theory predicts that different student hidden
units will always specialize to individual teacher hidden units for t --+ 00, whatever
the value of a. This leads to a decay of ?g from a plateau value at intermediate times
t. In the simulations, on the other hand, this specialization (or symmetry breaking)
appears to be inhibited or at least delayed until very large t. This can happen even
for zero noise and a 2:: L, where the training data should should contain enough
information to force student and teacher weights to be equal asymptotically. The
reason for this is not clear to us, and deserves further study. Our initial investigations, however, suggest that symmetry breaking may be strongly delayed due to the
presence of saddle points in the training error surface with very 'shallow' unstable
directions.
When our theory fails, which of its assumptions are violated? It is conceivable
that multiple local minima in the training error surface could cause self-averaging
to break down; however, we have found no evidence for this, see figure(3a). On
the other hand, the simulation results in figure(3b) clearly show that the implicit
assumption of Gaussian student activations - as discussed before eq. (8) - can be
violated.
On-line Learning from Finite Training Sets in Nonlinear Networks
(a)
363
(b)
/
Variance over training histories
10"'" ' - - - - - - - - - - - - - - - '
102
N
Figure 3: (a) Variance of fg(t = 20) vs input dimension N for student and teacher
with two hidden units (L = M = 2), a = 0.5, 'fJ = 2, and zero noise. The bottom
curve shows the variance due to different random choices of training examples from
a fixed training set ('training history'); the top curve also includes the variance due
to different training sets. Both are compatible with the liN decay expected if selfaveraging holds (dotted line). (b) Distribution (over training set) of the activation
hI of the first hidden unit of the student. Histogram from simulations for N = 1000,
all other parameter values as in (a).
In summary, the main theoretical contribution of this paper is the extension of online
learning analysis for finite training sets to nonlinear networks. Our approximate
theory does not require the use of replicas and yields ordinary first order differential
equations for the time evolution of a set of order parameters. Its central implicit
assumption (and its Achilles' heel) is that the student activations are Gaussian
distributed. In comparison with simulations, we have found that it is more accurate
than the infinite training set analysis at predicting the generalization dynamics for
finite training sets, both qualitatively and also quantitatively for small learning
times t. Future work will have to show whether the theory can be extended to cope
with non-Gaussian student activations without incurring the technical difficulties
of dynamical replica theory [2], and whether this will help to capture the effects of
local minima and, more generally, 'rough' training error surfaces.
Acknowledgments: We would like to thank Ansgar West for helpful discussions.
References
[1] M. Biehl and H. Schwarze. Journal of Physics A, 28:643-656, 1995.
[2] A. C. C. Coolen, S. N. Laughton, and D. Sherrington. In NIPS 8, pp. 253-259,
MIT Press, 1996; S.N. Laughton, A.C.C. Coolen, and D. Sherrington. Journal
of Physics A, 29:763-786, 1996.
[3] See for example: The dynamics of online learning. Workshop at NIPS'95.
[4] T. Heskes and B. Kappen. Physical Review A, 44:2718-2762, 1994.
[5] D. Saad and S. A. Solla Physical Review E, 52:4225, 1995.
[6] P. Sollich. Journal of Physics A, 27:7771-7784, 1994.
[7] P. Sollich and D. Barber. In NIPS 9, pp.274-280, MIT Press, 1997; Europhysics
Letters, 38:477-482, 1997.
| 1390 |@word eor:1 km:6 seek:1 simulation:11 covariance:3 carry:1 kappen:1 initial:4 comparing:1 nt:1 activation:18 yet:1 additive:1 realistic:1 happen:1 remove:1 update:10 v:3 alone:1 selected:1 yr:3 provides:2 successive:2 kinh:1 become:2 differential:1 qualitative:1 incorrect:1 specialize:1 manner:1 expected:1 mechanic:1 ol:1 begin:1 eigenspace:1 developed:2 fellow:1 quantitative:1 act:1 uk:2 whatever:1 unit:10 grant:1 omit:1 before:1 negligible:1 local:3 vni:1 limit:6 consequence:1 studied:1 acknowledgment:1 practice:1 block:1 significantly:1 suggest:2 convenience:1 cannot:1 close:2 onto:2 restriction:1 imposed:1 projector:1 yt:1 straightforward:2 starting:1 l:1 immediately:1 wie:1 erg:1 oh:5 laplace:1 limiting:1 updated:2 exact:4 agreement:1 jfi:2 predicts:1 qll:1 bottom:3 epsrc:1 capture:3 ensures:1 ordering:1 solla:1 mentioned:1 ui:1 dynamic:13 hopfield:1 describe:1 quite:1 larger:2 biehl:1 otherwise:1 erf:1 final:1 online:11 descriptive:1 sequence:1 eigenvalue:5 analytical:1 product:1 description:2 extending:1 diverges:1 help:1 develop:1 ac:2 ansgar:1 eq:1 wtx:1 predicted:1 direction:1 j75425:1 require:1 behaviour:1 generalization:10 preliminary:1 investigation:1 achilles:1 extension:1 noninfinitesimal:1 hold:2 mm:1 sufficiently:1 great:1 mapping:2 birmingham:1 coolen:2 tool:1 rough:1 clearly:1 mit:2 gaussian:8 always:2 aim:1 c5ij:1 rather:1 hj:1 derived:1 improvement:1 contrast:1 helpful:1 typically:1 integrated:1 initially:1 hidden:12 noh:1 denoted:2 development:1 equal:3 yu:1 discrepancy:2 future:1 quantitatively:1 inhibited:1 individual:1 delayed:2 replacement:1 interest:1 highly:1 introduces:1 accurate:2 integral:3 capable:1 eigenspaces:2 orthogonal:1 circle:2 theoretical:2 soft:4 tp:4 deserves:1 ordinary:1 gr:1 too:1 straightforwardly:1 dependency:1 teacher:17 considerably:1 physic:5 yl:1 together:3 squared:1 central:1 containing:1 choose:2 yp:1 account:1 student:19 includes:1 caused:1 explicitly:1 depends:2 break:2 closed:2 wm:1 contribution:1 ir:1 accuracy:1 variance:8 correspond:1 yield:1 history:2 plateau:1 cumbersome:1 ed:1 definition:1 underestimate:1 pp:2 selfaveraging:1 rim:2 actually:1 appears:1 originally:2 supervised:1 specify:1 yb:1 ooo:1 evaluated:1 strongly:1 governing:1 implicit:2 correlation:5 until:1 sketch:1 hand:3 replacing:1 nonlinear:9 defines:1 schwarze:1 effect:3 contain:1 evolution:4 deal:1 during:2 self:3 outline:1 complete:1 sherrington:2 fj:3 novel:2 recently:1 common:2 physical:2 b4:1 extend:1 discussed:1 numerically:1 significant:2 mathematics:1 similarly:1 heskes:1 longer:1 surface:3 scenario:1 minimum:3 somewhat:1 eo:2 dashed:2 ii:3 thermodynamic:1 multiple:1 technical:1 adapt:1 calculation:3 h7:1 lin:1 europhysics:1 prediction:3 histogram:1 achieved:1 addition:1 completes:1 dorothy:1 saad:1 hz:1 presence:2 intermediate:1 split:1 enough:3 variety:1 architecture:1 reduce:1 regarding:1 absent:1 whether:2 specialization:1 peter:1 cause:1 constitute:1 tij:1 generally:1 detailed:1 involve:1 clear:2 yw:1 transforms:1 dotted:1 correctly:1 track:1 write:2 discrete:1 express:1 nevertheless:1 drawn:2 replica:2 advancing:1 asymptotically:1 sum:1 letter:1 hodgkin:1 throughout:2 reasonable:2 hi:5 encountered:1 adapted:1 constraint:1 worked:1 l9:1 relatively:1 department:2 combination:1 sollich:6 em:3 wi:3 shallow:1 heel:1 hl:4 intuitively:2 taken:4 equation:18 remains:1 pin:1 turn:1 committee:4 discus:1 otu:1 incurring:1 v2:1 appropriate:2 disagreement:1 r1m:1 appearing:1 top:2 include:1 yx:1 calculating:2 giving:1 restrictive:1 society:1 added:1 primary:1 gradient:2 conceivable:1 thank:1 oa:1 barber:6 unstable:1 trivial:1 reason:2 length:1 lg:1 rise:1 finite:13 descent:2 curved:1 extended:1 lb:1 david:1 namely:1 pair:1 required:2 nip:3 beyond:1 dynamical:6 below:3 royal:1 including:1 unrealistic:1 overlap:5 difficulty:4 force:1 predicting:1 representing:1 aston:2 yq:1 naive:1 review:2 literature:1 laughton:2 relative:1 principle:2 corrupting:2 uncorrelated:2 compatible:1 summary:1 perceptron:3 comprehensively:1 taking:1 fg:3 edinburgh:2 distributed:2 overcome:2 dimension:2 calculated:1 transition:1 curve:2 rich:1 qualitatively:2 projected:4 cope:1 functionals:1 approximate:4 compact:1 overfitting:1 assumed:2 unnecessary:1 spectrum:1 continuous:2 stimulated:1 jz:1 correlated:3 symmetry:2 main:3 motivation:1 noise:16 west:1 ff:1 fails:1 explicit:2 xl:3 breaking:2 down:4 symbol:2 decay:2 evidence:1 workshop:1 effectively:1 saddle:1 expressed:1 contained:1 scalar:1 goal:1 presentation:2 marked:1 considerable:1 change:1 infinite:13 averaging:3 e:1 la:1 violated:2 phenomenon:1 ex:1 |
429 | 1,391 | Effects of Spike Timing Underlying
Binocular Integration and Rivalry in a
Neural Model of Early Visual Cortex
Erik D. Lumer
Wellcome department of Cognitive Neurology
Institute of Neurology, University College of London
12 Queen Square, London, WC1N 3BG, UK
Abstract
In normal vision, the inputs from the two eyes are integrated into a single percept. When dissimilar images are
presented to the two eyes, however, perceptual integration gives way to alternation between monocular inputs,
a phenomenon called binocular rivalry. Although recent
evidence indicates that binocular rivalry involves a modulation of neuronal responses in extrastriate cortex, the
basic mechanisms responsible for differential processing of
con:6.icting and congruent stimuli remain unclear. Using a
neural network that models the mammalian early visual
system, I demonstrate here that the desynchronized firing of cortical-like neurons that first receive inputs from
the two eyes results in rivalrous activity patterns at later
stages in the visual pathway. By contrast, synchronization
of firing among these cells prevents such competition. The
temporal coordination of cortical activity and its effects
on neural competition emerge naturally from the network
connectivity and from its dynamics. These results suggest
that input-related differences in relative spike timing at
an early stage of visual processing may give rise to the
phenomena both of perceptual integration and rivalry in
binocular vision.
1
Introduction
The neural determinants of visual perception can be probed by subjecting the visual
system to ambiguous viewing conditions - stimulus configurations that admit more
E. D.Lumer
188
than one perceptual interpretation. For example, when a left-tilted grating is shown
to the left eye and a right-tilted grating to the right eye, the two stimuli are momentarily perceived together as a plaid pattern, but soon only one line grating becomes
visible, while the other is suppressed. This phenomenon, known as binocular rivalry,
has long been thought to involve competition between monocular neurons within
the primary visual cortex (VI), leading to the suppression of information from one
eye (Lehky, 1988; Blake, 1989). It has recently been shown, however, that neurons
whose activity covaries with perception during rivalry are found mainly in higher
cortical areas and respond to inputs from both eyes, thus suggesting that rivalry
arises instead through competition between alternative stimulus interpretations in
extrastriate cortex (Leopold and Logothetis, 1996). Because eye-specific information appears to be lost at this stage, it remains unclear how the stimulus conditions
(i.e. conflicting monocular stimuli) yielding binocular rivalry are distinguished from
the conditions (i.e. matched monocular inputs) that produce stable single vision.
I propose here that the degree of similarity between the images presented to the
two eyes is registered by the temporal coordination of neuronal activity in VI, and
that changes in relative spike timing within this area can instigate the differential
responses in higher cortical areas to conflicting or congruent visual stimuli. Stimulus and eye-specific synchronous activity has been described previously both in
the lateral geniculate nucleus (LGN) and in the striate cortex (Gray et al., 1989;
Sillito et al., 1994; Neuenschwander and Singer, 1996). It has been suggested that
such synchrony may serve to bind together spatially distributed neural events into
coherent representations (Milner, 1974; von der Malsburg, 1981; Singer, 1993). In
addition, reduced synchronization of striate cortical responses in strabismic cats has
been correlated with their perceptual inability to combine signals from the two eyes
or to incorporate signals from an amblyopic eye (Konig et al., 1993; Roelfsema et
al., 1994). However, the specific influences of interocular input-similarity on spike
coordination in the striate cortex, and of spike coordination on competition in other
cortical areas, remain unclear.
To examine these influences, a simplified neural model of an early visual pathway
is simulated. In what follows, I first describe the anatomical and physiological constraints incorporated in the model, and then show that a temporal patterning of
neuronal activity in its primary cortical area emerges naturally. By manipulating
the relative spike timing of neuronal discharges in this area, I demonstrate its role
in inducing differential responses in higher visual areas to conflicting or congruent visual stimulation. Finally, I discuss possible implications of these results for
understanding the neural basis of normal and ambiguous perception in vivo.
2
Model overview
The model has four stages based on the organization of the mammalian visual pathway (Gilbert, 1993). These stages represent: (i) sectors of an ipsilateral ('left eye')
and a contralateral ('right eye') lamina of the LGN, which relay visual inputs to
the cortex; (ii) two corresponding monocular regions in layer 4 of VI with different
ocular dominance; (iii) a primary cortical sector in which the monocular inputs are
first combined (called Vp in the model); and (iv) a secondary visual area of cortex
in which higher-order features are extracted (Vs in the model; Fig. 1). Each stage
consists of 'standard' integrate-and-fire neurons that are incorporated in synaptic
networks. At the cortical stages, these units are grouped in local recurrent circuits
that are similar to those used in previous modeling studies (Douglas et al., 1995;
Somers et al., 1995). Synaptic interactions in these circuits are both excitatory and
inhibitory between cells with similar orientation selectivity, but are restricted to in-
Spike Timing Effects in Binocular Integration and Rivalry
189
(L.L ___ , _.. ___ , ( )__ ... ___ -.. ,
:9
: : Exp' . ~1 ; ; ElI2p'
Iaver 4 : Ex1
t-.~...J
. :?~...J :
- - ~ In~~~; :.~~~ ; : Inhl..1j :,~~ J
(L)
Figure 1: Architecture of the model. Excitatory and inhibitory connections are represented by lines with arrowheads and round terminals , respectively, Each lamina
in the LGN consists of 100 excitatory units (Ex) and 100 inhibitory units (Inh) ,
coupled via local inhibition. Cortical units are grouped into local recurrent circuits
(stippled boxes) , each comprising 200 Ex units and 100 Inh units. In each monocular patch of layer 4, one cell group (Exl and Inh1) responds to left-tilted lines
(orientation 1), whereas a second group (Ex2 and Inh2) is selective for right-tilted
lines (orientation 2) . The same orientation selectivities are remapped onto Vp and
Vs, although cells in these areas respond to inputs from both eyes. In addition,
convergent inputs from Vp to Vs establish a third selectivity in Vs, namely for line
crossings (Ex+ and Inh+) .
hibition only between cell groups with orthogonal orientation preference (Kisvarday
and Eysel, 1993) . Two orthogonal orientations (orientation 1 and 2) are mapped
in each monocular sector of layer 4, and in Vp. To account for the emergence of
more complex response properties at higher levels in the visual system (Van Essen
and Gallant, 1994), forward connectivity patterns from Vp to Vs are organized to
support three feature selectivities in Vs , one for orientation 1, one for orientation 2,
and one for the conjunction of these two orientations, i.e. for line crossings. These
forward projections are reciprocated by weaker backward projections from Vs to
Vp. As a general rule , connections are established at random within and between
interconnected populations of cells, with connection probabilities between pairs of
cells ranging from 1 to 10 %, consistent with experimental estimates (Thomson et
al., 1988; Mason et al., 1991) . Visual stimulation is achieved by applying a stochastic synaptic excitation independently to activated cells in the LGN . A quantitative
description of the model parameters will be reported elsewhere.
190
3
E. D. Lumer
Results
In a first series of simulations, the responses of the model to conflicting and congruent visual stimuli are compared. When the left input consists of left-tilted lines
(orientation 1) and the right input of right-tilted lines (orientation 2), rivalrous response suppression occurs in the secondary visual area. At any moment, only one
of the three feature-selective cell groups in Vs can maintain elevated firing rates
(Fig. 2a). By contrast, when congruent plaid patterns are used to stimulate the
two monocular channels, these cell groups are forced in a regime in which they all
sustain elevated firing rates (Fig. 2b). This concurrent activation of cells selective
for orthogonal orientations and for line crossings can be interpreted as a distributed
representation of the plaid pattern in Vs 1. A quantitative assessment of the degree of competition in Vs is shown in Figure 2c. The rivalry index of two groups
of neurons is defined as the mean absolute value of the difference between their
instantaneous group-averaged firing rates divided by the highest instantaneous firing rate among the two cell groups. This index varies between 0 for nonrivalrous
groups of neurons and 1 for groups of neurons with mutually exclusive patterns of
activity. Groups of cells with different selectivity in Vs have a significantly higher
rivalry index when stimulated by conflicting rather than by congruent visual inputs
(p < 0.0001) (Fig. 2c).
Note that, in the example shown in Figure 2a, the differential responses to conflicting
inputs develop from about 200 ms after stimulus onset and are maintained over
the remainder of the stimulation epoch. In other simulations, alternation between
dominant and suppressed responses was also observed over the same epoch as a
result of fluctuations in the overall network dynamics. A detailed analysis of the
dynamics of perceptual alternation during rivalry, however, is beyond the scope of
this report.
Although Vp exhibits a similar distribution of firing rates during rivalrous and nonrivalrous stimulation, synchronization between the two cell groups in Vp is more
pronounced in the nonrivalrous than in the rivalrous case (Fig. 2d, upper plots).
Subtraction of the shift predictor demonstrates that the units are not phase-locked
to the stimuli. The changes in spike coordination among Vp units reflects the
temporal patterning of their layer 4 inputs. During rivalry, Vp cells with different
orientation selectivity are driven by layer 4 units that belong to separate monocular
pathways, and hence, are uncorrelated (Fig. 2d, lower left). By contrast, cells in Vp
receive convergent inputs from both eyes during nonrivalrous stimulation. Because
of the synchronization of discharges among cells responsive to the same eye within
layer 4 (Fig. 2d, lower right), the paired activities from the two monocular channels are also synchronized, a.nd provide synchronous inputs to cells with different
orienta.tion selectivity in Vp.
To establish unequivocally that changes in spike coordination within Vp are sufficient to trigger differential responses in Vs to conflicting and congruent stimuli,
the model can be modified as follows. A single group of cells in layer 4 is used to
drive with equal strength both orientation-selective populations of neurons in Vp.
The outputs from layer 4, however, is relayed to these two target populations with
average transmission delays that differ by either 10 ms or by 0 ms. In the first case,
competition prevails among cells in the secondary visual area. This contrasts with
the nonrivalrous activity in this area when similar transmission delays are used at
an earlier stage (data not shown). This test confirms that changes in relative spike
lTo discount possible effects of binocular snmmmation, synaptic strengths from layer 4
to Vp are reduced during congruent stimulation so as to produce a feedforward activation
of Vp comparable to that elicited by conflicting monocular inputs.
Spike Timing Effects in Binocular Integration and Rivalry
conflicting visual inputs
a
191
congruent visual inputs
b
40~
.S
,
30
_6
N
3
2
-::~
4
N
l:
J:
40
2
3
C)2
-Vs+
C)2
2
.s '
I-
3
o
1
2
<;'--t
" -" - : - "
3
4
L41
-R42
-ee- L42
R41
2
3
..
Time (ms)
Time (ms)
c
-Yp1
-Yp2
-3..-------:4
ir.
:[4
:u~
at-.. . t.~ " oC"_ ~'
'b'---~
1 -~2-
-Ys1
-Ys2
..
4
C
'C
d
congruent
conflicting
fHdNkd
1m gAd~50
><
CI)
"c
~
l
a:
1
I-
11.I"'\Jo/"1IfV!1'Y
I-
u::
'b
~eo~vp
1U
G)
1&
.
CJ
0.1
o
Vs1:Vs2
max(Vs}:Vs+
-80
0
80 -80
0
80
Time lag (ms)
Figure 2: A, Instantaneous firing rates in response to conflicting inputs for cell
groups in layer 4, in Vp, and in Vs (stimulus onset at t = 250ms). Discharge rates
of layer 4 cells driven by different 'eyes' are similar (lower plot). By contrast, Vs
exhi bi ts com peti ti ve firing pat terns soon after stimulus onset (upper plot). Feed back
influence from Vs to Vp results in comparatively weaker competition in Vp (middle
plot). B, Responses to congruent inputs. All cell groups in layer 4 are activated by
visual inputs. Nonrivalrous firing patterns ensue in Vp and Vs. C, Rivalry indices
during conflicting and congruent stimulation, are calculated for the two orientationselective cell groups and for the dominant and cross-selective cell group in Vs. D,
Interocular responses are uncorrelated in layer 4 (lower left), whereas intraocular
activities are synchronous at this stage (lower right). Enhanced synchronization of
discharges ensues between cell groups in Vp during congruent stimulation (upper
right), relative to the degree of coherence during conflicting stimulation (upper left).
E. D.Lumer
192
timing are sufficient to switch the outcome of neural network interactions involving
strong mutual inhibition from competitive to cooperative.
4
Conclusion
In the present study, a simplified model of a visual pathway was used to gain
insight into the neural mechanisms operating during binocular vision. Simulations of
neuronal responses to visual inputs revealed a stimulus-related patterning of relative
spike timing at an early stage of cortical processing. This patterning reflected
the degree of similarity between the images presented to the two 'eyes', and, in
turn, it altered the outcome of competitive interactions at later stages along the
visual pathway. These effects can help explaining how the same cortical networks
can exhibit both rivalrous and nonrivalrous activity, depending on the temporal
coordination of their synaptic inputs.
These results bear on the interpretation of recent empirical findings about the neuronal correlates of rivalrous perception. In experiments with awake monkeys, Logothetis and colleagues (Sheinberg et al., 1995; Leopold and Logothetis, 1996) have
shown that neurons whose firing rate correlates with perception during rivalry are
distributed at several levels along the primate visual pathways, including Vl/V2,
V4, and IT. Importantly, the fraction of modulated responses is lower in VI than in
extrastriate areas, and it increases with the level in the visual hierarchy. Simulations
of the present model exhibit a behavior that is consistent with these observations.
However, these simulations also predict that both rivalrous and nonrivalrous perception may have a clear neurophysiological correlate in VI, i.e. at the earliest
stage of visual cortical processing. Accordingly, congruent stimulation of both eyes
will synchronize the firing of binocular cells with overlapping receptive fields in Vl.
By contrast, conflicting inputs to the two eyes will cause a desynchronization between their corresponding neural events in Vl. Because this temporal registration
of stimulus dissimilarity instigates competition among binocular cells in higher visual areas and not between monocular pathways, the ensuing pattern of response
suppression and dominance is independent of the eyes through which the stimuli are
presented. Thus, the model can in principle account for the psychophysical finding that a single phase of perceptual dominance during rivalry can span multiple
interocular exchanges of the rival stimuli (Logothetis et al., 1996).
The present results also reveal a novel property of canonical cortical-like circuits
interacting through mutual inhibition, i.e. the degree of competition among such
circuits exhibits a remarkable sensitivity to the relative timing of neuronal action
potentials. This suggests that the temporal patterning of cortical activity may be
a fundamental mechanism for selecting among stimuli competing for the control of
attention and motor action.
Acknowledgements
This work was supported in part by an IRSIA visiting fellowship at the Center
for Nonlinear Phenomena and Complex Systems, Universite Libre de Bruxelles. I
thank Professor Gregoire Nicolis for his hospitality during my stay in Brussels; and
David Leopold and Daniele Piomelli for helpful discussions and comments on an
earlier version of the manuscript.
References
Blake R (1989) A neural theory of binocular vision. Psychol Rev 96:145-167.
Spike Timing Effects in Binocular Integration andRivalry
193
Douglas RJ, Koch C, Mahowald M, Martin K, Suarez H (1995) Recurrent excitation
in neocortical circuits. Science 269:981-985.
Gilbert C (1993) Circuitry, architecture, and functional dynamics of visual cortex.
Cereb Cortex 3:373-386.
Gray CM, Konig P, Engel AK, Singer, W (1989) Oscillatory responses in cat visual cortex exhibit inter-columnar synchronization which reflects global stimulus
properties. Nature 338 :334-337.
Kisvarday ZF, Eysel UT (1993) Functional and structural topography of horizontal
inhibitory connections in cat visual cortex. Europ J Neurosci 5:1558-1572.
Konig P, Engel AK, Lowel S, Singer, W (1993) Squint affects synchronization of
oscillatory responses in cat visual cortex. Eur J Neurosci 5:501-508.
Lehky SR (1988) An astable multivibrator model of binocular rivalry. Perception
17: 215- 228.
Leopold DA, Logothetis NK (1996) Activity changes in early visual cortex reflect
monkeys percepts during binocular rivalry. Nature 379:549-553.
Logothetis NK, Leopold DA, Sheinberg DL (1996) What is rivalling during rivalry?
Nature 380:621-624.
Neuenschwander S, Singer W (1996) Long-range synchronization of oscillatory light
responses in the cat retina and lateral geniculate nucleus. Nature 379:728-733.
Milner PM (1974) A model of visual shape recognition. Psychol Rev 81:521-535.
Roelfsema PR, Konig P, Engel AK, Sireteanu R , Singer W (1994) Reduced synchronization in the visual cortex of cats with strabismic amblyopia. Eur J Neurosci
6:1645-1655.
Sheinberg DL, Leopold DA, Logothetis NK (1995) Effects of binocular rivalry on
face cell activity in monkey temporal cortex. Soc Neurosci Abstr 21:15.12.
Sillito AM, Jones HE, Gerstein GL, West DC (1994) Feature-linked synchronization
of thalamic relay cell firing induced by feedback from the visual cortex. Nature
369:479-482.
Singer W (1993) Synchronization of cortical activity and its putative role in information processing. Annu Rev Physiol 55:349-374.
Somers D, Nelson S, Sur M (1995) An emergent model of orientation selectivity in
cat visual cortical simple cells. J Neurosci 15:5448-5465.
Van Essen DC, Gallant JL (1994) Neural mechanisms ofform and motion processing
in the primate visual system . Neuron 13:1-10.
von der Malsburg C (1981) The correlation theory of the brain. Internal Report
81-2, Max Planck Institute for Biophysical Chemistry, Gottingen.
| 1391 |@word determinant:1 version:1 middle:1 nd:1 confirms:1 simulation:5 moment:1 extrastriate:3 configuration:1 series:1 selecting:1 com:1 activation:2 physiol:1 tilted:6 visible:1 shape:1 motor:1 plot:4 v:20 exl:1 patterning:5 accordingly:1 preference:1 relayed:1 along:2 differential:5 consists:3 pathway:8 combine:1 ex2:1 inter:1 behavior:1 examine:1 brain:1 terminal:1 becomes:1 underlying:1 matched:1 circuit:6 what:2 cm:1 interpreted:1 monkey:3 finding:2 gottingen:1 temporal:8 quantitative:2 ti:1 demonstrates:1 uk:1 control:1 unit:9 planck:1 timing:10 bind:1 local:3 ak:3 firing:13 modulation:1 fluctuation:1 suggests:1 locked:1 bi:1 lumer:4 averaged:1 range:1 responsible:1 lost:1 area:14 empirical:1 thought:1 significantly:1 projection:2 suggest:1 onto:1 influence:3 applying:1 gilbert:2 center:1 attention:1 independently:1 rule:1 insight:1 importantly:1 his:1 population:3 discharge:4 target:1 logothetis:7 milner:2 trigger:1 enhanced:1 hierarchy:1 crossing:3 recognition:1 mammalian:2 cooperative:1 observed:1 role:2 suarez:1 region:1 momentarily:1 highest:1 dynamic:4 serve:1 lto:1 basis:1 emergent:1 cat:7 represented:1 forced:1 describe:1 london:2 outcome:2 whose:2 lag:1 emergence:1 ifv:1 biophysical:1 propose:1 interaction:3 interconnected:1 remainder:1 eysel:2 gregoire:1 description:1 inducing:1 pronounced:1 competition:10 konig:4 abstr:1 transmission:2 congruent:14 produce:2 lamina:2 help:1 recurrent:3 develop:1 depending:1 strong:1 soc:1 grating:3 europ:1 involves:1 synchronized:1 differ:1 plaid:3 stochastic:1 viewing:1 exchange:1 koch:1 blake:2 normal:2 exp:1 scope:1 predict:1 circuitry:1 early:6 relay:2 perceived:1 geniculate:2 coordination:7 grouped:2 concurrent:1 engel:3 reflects:2 hospitality:1 modified:1 rather:1 conjunction:1 earliest:1 indicates:1 mainly:1 contrast:6 suppression:3 am:1 helpful:1 vl:3 integrated:1 manipulating:1 selective:5 lgn:4 comprising:1 overall:1 among:8 orientation:16 integration:6 mutual:2 equal:1 l42:1 field:1 jones:1 subjecting:1 report:2 stimulus:20 retina:1 ve:1 phase:2 fire:1 maintain:1 organization:1 essen:2 yielding:1 light:1 activated:2 wc1n:1 implication:1 orthogonal:3 arrowhead:1 iv:1 modeling:1 earlier:2 queen:1 mahowald:1 contralateral:1 predictor:1 delay:2 reported:1 ensues:1 varies:1 my:1 combined:1 eur:2 fundamental:1 sensitivity:1 stay:1 v4:1 together:2 neuenschwander:2 connectivity:2 von:2 jo:1 reflect:1 prevails:1 cognitive:1 admit:1 leading:1 suggesting:1 account:2 potential:1 de:1 chemistry:1 bg:1 vi:5 onset:3 later:2 tion:1 vs2:1 linked:1 competitive:2 thalamic:1 elicited:1 synchrony:1 vivo:1 square:1 ir:1 percept:2 ensue:1 vp:22 interocular:3 drive:1 oscillatory:3 synaptic:5 colleague:1 ocular:1 naturally:2 universite:1 con:1 gain:1 yp1:1 emerges:1 ut:1 organized:1 cj:1 back:1 appears:1 feed:1 manuscript:1 higher:7 reflected:1 response:19 sustain:1 box:1 binocular:17 stage:12 correlation:1 horizontal:1 nonlinear:1 assessment:1 overlapping:1 reveal:1 gray:2 stimulate:1 effect:8 hence:1 spatially:1 ex1:1 round:1 during:15 ambiguous:2 maintained:1 excitation:2 daniele:1 oc:1 m:7 thomson:1 demonstrate:2 neocortical:1 cereb:1 motion:1 image:3 ranging:1 instantaneous:3 novel:1 recently:1 stimulation:10 functional:2 overview:1 jl:1 belong:1 interpretation:3 he:1 elevated:2 pm:1 strabismic:2 stable:1 cortex:17 similarity:3 inhibition:3 operating:1 dominant:2 recent:2 driven:2 selectivity:8 alternation:3 der:2 eo:1 subtraction:1 signal:2 ii:1 multiple:1 rj:1 cross:1 long:2 divided:1 paired:1 involving:1 basic:1 vision:5 represent:1 achieved:1 cell:31 receive:2 addition:2 whereas:2 fellowship:1 desynchronized:1 sr:1 comment:1 induced:1 ee:1 structural:1 feedforward:1 iii:1 revealed:1 switch:1 affect:1 architecture:2 competing:1 shift:1 synchronous:3 cause:1 action:2 orientationselective:1 detailed:1 involve:1 clear:1 discount:1 rival:1 lehky:2 reduced:3 canonical:1 inhibitory:4 ipsilateral:1 anatomical:1 probed:1 dominance:3 group:18 four:1 douglas:2 registration:1 backward:1 fraction:1 respond:2 somers:2 roelfsema:2 patch:1 putative:1 gerstein:1 coherence:1 comparable:1 layer:13 convergent:2 activity:15 strength:2 constraint:1 awake:1 yp2:1 span:1 martin:1 department:1 brussels:1 remain:2 suppressed:2 rev:3 primate:2 restricted:1 pr:1 wellcome:1 monocular:13 mutually:1 remains:1 previously:1 discus:1 vs1:1 mechanism:4 turn:1 singer:7 v2:1 distinguished:1 responsive:1 alternative:1 malsburg:2 establish:2 comparatively:1 psychophysical:1 spike:13 occurs:1 receptive:1 primary:3 exclusive:1 striate:3 responds:1 ys1:1 unclear:3 exhibit:5 visiting:1 bruxelles:1 lowel:1 separate:1 mapped:1 lateral:2 simulated:1 kisvarday:2 ensuing:1 thank:1 nelson:1 erik:1 sur:1 index:4 sector:3 rise:1 squint:1 gallant:2 upper:4 zf:1 neuron:10 observation:1 t:1 pat:1 sheinberg:3 incorporated:2 inh:3 dc:2 interacting:1 david:1 namely:1 pair:1 connection:4 leopold:6 coherent:1 registered:1 conflicting:14 established:1 beyond:1 suggested:1 pattern:8 perception:7 remapped:1 regime:1 max:2 including:1 event:2 synchronize:1 altered:1 eye:22 psychol:2 coupled:1 covaries:1 epoch:2 understanding:1 acknowledgement:1 relative:7 synchronization:11 bear:1 topography:1 remarkable:1 nucleus:2 integrate:1 degree:5 sufficient:2 consistent:2 principle:1 unequivocally:1 uncorrelated:2 excitatory:3 elsewhere:1 supported:1 gl:1 soon:2 weaker:2 institute:2 explaining:1 face:1 emerge:1 rivalry:21 absolute:1 distributed:3 van:2 feedback:1 calculated:1 cortical:17 forward:2 simplified:2 correlate:3 global:1 neurology:2 sillito:2 stimulated:1 channel:2 reciprocated:1 nature:5 correlated:1 complex:2 da:3 neurosci:5 libre:1 rivalrous:7 neuronal:7 fig:7 west:1 gad:1 perceptual:6 third:1 annu:1 specific:3 desynchronization:1 mason:1 physiological:1 evidence:1 dl:2 ci:1 dissimilarity:1 nk:3 columnar:1 neurophysiological:1 visual:40 prevents:1 extracted:1 professor:1 change:5 called:2 secondary:3 experimental:1 college:1 internal:1 support:1 tern:1 arises:1 inability:1 dissimilar:1 r41:1 modulated:1 incorporate:1 phenomenon:4 ex:3 |
430 | 1,392 | Approximating Posterior Distributions
in Belief Networks using Mixtures
Christopher M. Bishop
Neil Lawrence
Neural Computing Research Group
Dept. Computer Science & Applied Mathematics
Aston University
Binningham, B4 7ET, U.K.
Tommi Jaakkola
Michael I. Jordan
Center for Biological and Computational Learning
Massachusetts Institute of Technology
79 Amherst Street, ElO-243
Cambridge, MA 02139, U.S.A.
Abstract
Exact inference in densely connected Bayesian networks is computationally intractable, and so there is considerable interest in developing effective approximation schemes. One approach which has been adopted is to
bound the log likelihood using a mean-field approximating distribution.
While this leads to a tractable algorithm, the mean field distribution is assumed to be factorial and hence unimodal. In this paper we demonstrate
the feasibility of using a richer class of approximating distributions based
on mixtures of mean field distributions. We derive an efficient algorithm
for updating the mixture parameters and apply it to the problem of learning in sigmoid belief networks. Our results demonstrate a systematic
improvement over simple mean field theory as the number of mixture
components is increased.
1 Introduction
Bayesian belief networks can be regarded as a fully probabilistic interpretation of feedforward neural networks. Maximum likelihood learning for Bayesian networks requires
the evaluation of the likelihood function P(VIO) where V denotes the set of instantiated
(visible) variables, and 0 represents the set of parameters (weights and biases) in the network. Evaluation of P(VIO) requires summing over exponentially many configurations of
Approximating Posterior Distributions in Belief Networks Using Mixtures
417
the hidden variables H, and is computationally intractable except for networks with very
sparse connectivity, such as trees. One approach is to consider a rigorous lower bound on
the log likelihood, which is chosen to be computationally tractable, and to optimize the
model parameters so as to maximize this bound instead.
If we introduce a distribution Q(H), which we regard as an approximation to the true
posterior distribution, then it is easily seen that the log likelihood is bounded below by
F[Q]
= L Q(H) In P(V~H).
Q(
{H}
(1)
)
The difference between the true log likelihood and the bound given by (1) is equal to the
Kullback-Leibler divergence between the true posterior distribution P(HIV) and the approximation Q(H). Thus the correct log likelihood is reached when Q(H) exactly equals
the true posterior. The aim of this approach is therefore to choose an approximating distribution which leads to computationally tractable algorithms and yet which is also flexible so
as to permit a good representation of the true posterior. In practice it is convenient to consider parametrized distributions, and then to adapt the parameters to maximize the bound.
This gives the best approximating distribution within the particular parametric family.
1.1
Mean Field Theory
Considerable simplification results if the model distribution is chosen to be factorial over
the individual variables, so that Q(H) =
Q(hd, which gives meanfieid theory. Saul
et al. (1996) have applied mean field theory to the problem of learning in sigmoid belief
networks (Neal. 1992). These are Bayesian belief networks with binary variables in which
the probability of a particular variable Si being on is given by
ni
P(S,
= Ilpa(S,? =" ( ~ ],;S; + b)
(2)
where u(z) == (1 + e-Z)-l is the logistic sigmoid function, pa(Si) denote the parents of
Si in the network. and Jij and bi represent the adaptive parameters (weights and biases)
in the model. Here we briefly review the framework of Saul et ai. (1996) since this forms
the basis for the illustration of mixture modelling discussed in Section 3. The mean field
distribution is chosen to be a product of Bernoulli distributions of the form
Q(H) =
II p,~i (1 _ p,;)l-h
i
(3)
in which we have introduced mean-field parameters J.Li. Although this leads to considerable
simplification of the lower bound, the expectation over the log of the sigmoid function.
arising from the use of the conditional distribution (2) in the lower bound (I), remains
intractable. This can be resolved by using variational methods (Jaakkola, 1997) to find a
lower bound on F(Q), which is therefore itself a lower bound on the true log likelihood.
In particular, Saul et al. (1996) make use of the following inequality
(In[l
+ e Zi ])
::;
ei(Zi)
+ In(e-~iZi + e(1-~;)Zi)
(4)
where Zi is the argument of the sigmoid function in (2), and ( ) denotes the expectation with
respect to the mean field distribution. Again, the quality of the bound can be improved by
adjusting the variational parameter ei. Finally, the derivatives of the lower bound with
respect to the J ij and bi can be evaluated for use in learning. In summary. the algorithm
involves presenting training patterns to the network. and for each pattern adapting the P,i
and
to give the best approximation to the true posterior within the class of separable
distributions of the form (3). The gradients of the log likelihood bound with respect to the
model parameters Jij and bi can then be evaluated for this pattern and used to adapt the
parameters by taking a step in the gradient direction.
ei
418
2
C. M. Bishop, N. LAwrence, T. Jaakkola and M I. Jordan
Mixtures
Although mean field theory leads to a tractable algorithm, the assumption of a completely
factorized distribution is a very strong one. In particular, such representations can only effectively model posterior distributions which are uni-modal. Since we expect multi-modal
distributions to be common, we therefore seek a richer class of approximating distributions which nevertheless remain computationally tractable. One approach (Saul and Jordan,
1996) is to identify a tractable substructure within the model (for example a chain) and then
to use mean field techniques to approximate the remaining interactions. This can be effective where the additional interactions are weak or are few in number, but will again prove
to be restrictive for more general, densely connected networks. We therefore consider an
alternative approach I based on mixture representations of the form
M
Qmix(H) = L
(5)
amQ(Hlm)
m=l
in which each of the components Q(Hlm) is itself given by a mean-field distribution, for
example of the form (3) in the case of sigmoid belief networks. Substituting (5) into the
lower bound (1) we obtain
F[Qmix] = L amF[Q(Hlm)]
+ f(m, H)
(6)
m
where f(m, H) is the mutual information between the component label m and the set of
hidden variables H, and is given by
f(m,H) = L
L
m {H}
Q(Hlm)
amQ(Hlm) In Q . (H)'
(7)
mix
The first tenn in (6) is simply a convex combination of standard mean-field bounds and
hence is no greater than the largest of these and so gives no useful improvement over a
single mean-field distribution. It is the second term, i.e. the mutual infonnation, which
characterises the gain in using mixtures. Since f(m, H) ~ 0, the mutual information
increases the value of the bound and hence improves the approximation to the true posterior.
2.1
Smoothing Distributions
As it stands, the mutual infonnation itself involves a summation over the configurations
of hidden variables, and so is computationally intractable. In order to be able to treat it
efficiently we first introduce a set of 'smoothing' distributions R(Hlm), and rewrite the
mutual infonnation (7) in the form
f(m, H)
-
LLamQ(Hlm)lnR(Hlm) - LamInam
{H}
m
- L
m
m
L amQ(Hlm) In {R(H1m) Qmix(H) } .
am Q(Hlm)
(8)
{H}
It is easily verified that (8) is equivalent to (7) for arbitrary R(Hlm). We next make use of
the following inequality
(9)
- In x ~ - ~x + In ~ + 1
lHere we outline the key steps. A more detailed discussion can be found in Jaakkola and Jordan
(1997).
Approximating Posterior Distributions in Belief Networks Using Mixtures
419
to replace the logarithm in the third term in (8) with a linear function (conditionally on
the component label m). This yields a lower bound on the mutual information given by
J(m,H) ~ J),(m,H) where
h(m,H)
I:I:amQ(Hlm)lnR(Hlm)- I: am In am
m {H}
m
- I: Am I: R(Hlm)Qmix(H)
m
{H}
+ I: am InAm + 1.
(10)
m
With J),(m, H) substituted for J(m, H) in (6) we again obtain a rigorous lower bound on
the true log likelihood given by
F),[Qmix(H)] = I: amF[Q(Hlm)]
+ h(m, H).
(11)
m
The summations over hidden configurations {H} in (10) can be performed analytically if
we assume that the smoothing distributions R(Hlm) factorize. In particular, we have to
consider the following two summations over hidden variable configurations
II I: R(hilm)Q(hilk) ~ 7rR,Q(m, k)
I: R(Hlm)Q(Hlk)
{H}
i
I: I: Q(hilm) InR(hilm) ~f H(QIIRlm).
I: Q(Hlm) InR(Hlm)
{H}
(12)
h.
(13)
h.
We note that the left hand sides of (12) and (13) represent sums over exponentially many
hidden configurations, while on the right hand sides these have been re-expressed in terms
of expressions requiring only polynomial time to evaluate by making use of the factorization of R(Hlm).
It should be stressed that the introduction of a factorized form for the smoothing distributions still yields an improvement over standard mean field theory. To see this, we note that
if R(Hlm) = const. for all {H, m} then J(m, H) = 0, and so optimization over R(Hlm)
can only improve the bound.
2.2
Optimizing the Mixture Distribution
In order to obtain the tightest bound within the class of approximating distributions, we can
maximize the bound with respect to the component mean-field distributions Q(Hlm), the
mixing coefficients am, the smoothing distributions R(Hlm) and the variational parameters Am' and we consider each of these in turn.
We will assume that the choice of a single mean field distribution leads to a tractable lower
bound, so that the equations
8F[Q]
8Q(h j ) = const
(14)
can be solved efficiently2. Since h(m, H) in (10) is linear in the marginals Q(hjlm), it
follows that its derivative with respect to Q(hj 1m) is independent of Q(hjlm), although it
will be a function of the other marginals, and so the optimization of (11) with respect to
individual marginals again takes the form (14) and by assumption is therefore soluble.
Next we consider the optimization with respect to the mixing coefficients am. Since all of
the terms in (11) are linear in am, except for the entropy term, we can write
F),[Qmix(H)] = I:am(-Em) - I:amlnam + 1
m
(15)
m
2In standard mean field theory the constant would be zero, but for many models of interest the
slightly more general equations given by (14) will again be soluble.
C. M. Bishop, N. Lawrence, T. Jaakkola and M. L Jordan
420
where we have used (10) and defined
F[Q(Hlm)]
+L
Q(Hlm) InR(Hlm)
{H}
+ LAk LR(Hlk)Q(Hlm) +lnA m .
k
(16)
{H}
Maximizing (15) with respect to am, subject to the constraints 0 ~ am ~ 1 and Lm am =
1, we see that the mixing coefficients which maximize the lower bound are given by the
Boltzmann distribution
exp(-Em)
am
= Lk exp(-Ek )'
(17)
We next maximize the bound (11) with respect to the smoothing marginals R(h j 1m). Some
manipulation leads to the solution
R(hilm) =
amQA~ilm) [~>'~'Q(m'k)Q(hi'k)l-1
(18)
in which 7r~,Q(m, k) denotes the expression defined in (12) but with the j term omitted
from the product.
The optimization of the JLmj takes the form of a re-estimation formula given by an extension
of the result obtained for mean-field theory by Saul et al. (1996). For simplicity we omit
the details here.
Finally, we optimize the bound with respect to the Am, to give
1
1
m
am
~= -
L 7rR,Q(m, k).
(19)
k
Since the various parameters are coupled, and we have optimized them individually keeping
the remainder constant, it will be necessary to maximize the lower bound iteratively until
some convergence criterion is satisfied. Having done this for a particular instantiation of
the visible nodes, we can then determine the gradients of the bound with respect to the
parameters governing the original belief network, and use these gradients for learning.
3
Application to Sigmoid Belief Networks
We illustrate the mixtures formalism by considering its application to sigmoid belief networks of the form (2). The components of the mixture distribution are given by factorized
Bernoulli distributions of the form (3) with parameters JLmi. Again we have to introduce
variational parameters ~mi for each component using (4). The parameters {JLmi, ~mi} are
optimized along with {am, R(hjlm), Am} for each pattern in the training set.
We first investigate the extent to which the use of a mixture distribution yields an improvement in the lower bound on the log likelihood compared with standard mean field theory.
To do this, we follow Saul et al. (1996) and consider layered networks having 2 units in
the first layer, 4 units in the second layer and 6 units in the third layer, with full connectivity between layers. In all cases the six final-layer units are considered to be visible and
have their states clamped at zero. We generate 5000 networks with parameters {Jij, bi }
chosen randomly with uniform distribution over (-1, 1). The number of hidden variable
configurations is 26 = 64 and is sufficiently small that the true log likelihood can be computed directly by summation over the hidden states. We can therefore compare the value of
421
Approximating Posterior Distributions in Belief Networks Using Mixtures
the lower bound F with the true log likelihood L, using the nonnalized error (L - F)/ L.
Figure 1 shows histograms of the relative log likelihood error for various numbers of mixture components, together with the mean values taken from the histograms. These show
a systematic improvement in the quality of the approximation as the number of mixture
components is increased.
5 components, mean: 0 .011394
4 components, mean: 0.012024
3~r-----~----~----~---.
3000r-----~----~----~--_.
0 .02
0.04
0.06
0.08
3 components, mean : 0.01288
0.02
0.04
0.06
0.08
2 components, mean: 0.013979
3000r-----~----~----~--_.
2000
0.04
0.02
0.04
0.06
0.08
1 component, mean: 0.015731
g0.014
o
III
o
c
as
~ 0.012
0.06
0.01
0.08
0.08
0.01 Gn..:-p----~----~----~----...,
~r-----~----~----~--~
0.04
0.06
o
'------~----~----~-----'
1
2
3
4
5
no. of components
Figure 1: Plots of histograms of the normalized error between the true log likelihood and the lower
bound. for various numbers of mixture components. Also shown is the mean values taken from the
histograms. plotted against the number of components.
Next we consider the impact of using mixture distributions on learning. To explore this
we use a small-scale problem introduced by Hinton et al. (1995) involving binary images
of size 4 x 4 in which each image contains either horizontal or vertical bars with equal
probability, with each of the four possible locations for a bar occupied with probability 0.5.
We trained networks having architecture 1-8-16 using distributions having between 1 and
5 components. Randomly generated patterns were presented to the network for a total of
500 presentations, and the J-tmi and ~mi were initialised from a unifonn distribution over
(0,1). Again the networks are sufficiently small that the exact log likelihood for the trained
models can be evaluated directly. A Hinton diagram of the hidden-to-output weights for the
eight units in a network trained with 5 mixture components is shown in Figure 2. Figure 3
shows a plot of the true log likelihood versus the number M of components in the mixture
for a set of experiments in which, for each value of M, the model was trained 10 times
starting from different random parameter initializations. These results indicate that, as the
number of mixture components is increased, the learning algorithm is able to find a set
of network parameters having a larger likelihood, and hence that the improved flexibility
of the approximating distribution is indeed translated into an improved training algorithm.
We are currently applying the mixture fonnalism to the large-scale problem of hand-written
digit classification.
422
C. M. Bishop, N. Lawrence, T. Jaakkola and M. I. Jordan
.? 0.
t? ? ? ?
????
? ??
o ? ?
o
j? ? ?
- 0
0.0
jO 0 . 0
10 0 . 0
???
10 ??.
0 ??
o
?
0
-0
Figure 2: Hinton diagrams of the hidden-to-output weights for each of the 8 hidden units in a
network trained on the 'bars' problem using a mixture distribution having 5 components.
-5
-6
o
-7
8
0
'80
:? -8
Q)
~
CI
.S! -9
Q)
o
o
~
8o
o
o
o
6
o
B
8
8
o
o
o
o
o
0
0
0
o
E
-10
-11
-12
0
234
5
6
no. of components
Figure 3: True log likelihood (divided by the number of patterns) versus the number M of mixture
components for the 'bars' problem indicating a systematic improvement in performance as M is
increased.
References
Hinton, G. E., P. Dayan, B. 1. Frey, and R. M. Neal (1995). The wake-sleep algorithm
for unsupervised neural networks. Science 268, 1158-1161.
Jaakkola, T. (1997). Variational Methods for Inference and Estimation in Graphical
Models. Ph.D. thesis, MIT.
Jaakkola, T. and M. I. Jordan (1997). Approximating posteriors via mixture models. To
appear in Proceedings NATO ASI Learning in Graphical Models, Ed. M. I. Jordan.
Kluwer.
Neal, R. (1992) . Connectionist learning of belief networks. Artificial Intelligence 56,
71-113.
Saul, L. K., T. Jaakkola, and M. I. Jordan (1996). Mean field theory for sigmoid belief
networks. Journal of Artificial Intelligence Research 4,61-76.
Saul, L. K. and M. I. Jordan (1996). Exploiting tractable substructures in intractable
networks. In D. S. Touretzky, M . C. Mozer, and M. E. Hasselmo (Eds.), Advances in
Neural Information Processing Systems, Volume 8, pp. 486-492. MIT Press.
| 1392 |@word briefly:1 polynomial:1 seek:1 configuration:6 contains:1 si:3 yet:1 written:1 visible:3 plot:2 tenn:1 intelligence:2 inam:1 lr:1 node:1 location:1 along:1 prove:1 introduce:3 indeed:1 multi:1 considering:1 bounded:1 factorized:3 exactly:1 unit:6 omit:1 appear:1 frey:1 treat:1 initialization:1 factorization:1 bi:4 practice:1 digit:1 asi:1 adapting:1 convenient:1 lnr:2 layered:1 applying:1 optimize:2 equivalent:1 center:1 maximizing:1 starting:1 convex:1 simplicity:1 regarded:1 hd:1 exact:2 pa:1 updating:1 hlm:28 solved:1 connected:2 mozer:1 trained:5 rewrite:1 basis:1 completely:1 translated:1 easily:2 resolved:1 various:3 instantiated:1 effective:2 artificial:2 hiv:1 richer:2 larger:1 neil:1 itself:3 final:1 rr:2 interaction:2 jij:3 product:2 remainder:1 mixing:3 flexibility:1 exploiting:1 parent:1 convergence:1 derive:1 illustrate:1 ij:1 strong:1 lna:1 involves:2 indicate:1 tommi:1 direction:1 correct:1 biological:1 summation:4 extension:1 sufficiently:2 considered:1 exp:2 lawrence:4 elo:1 lm:1 substituting:1 omitted:1 estimation:2 label:2 currently:1 infonnation:3 individually:1 largest:1 hasselmo:1 mit:2 aim:1 qmix:6 occupied:1 hj:1 jaakkola:9 improvement:6 modelling:1 likelihood:19 bernoulli:2 rigorous:2 am:18 inference:2 dayan:1 hidden:11 classification:1 flexible:1 smoothing:6 mutual:6 field:21 equal:3 having:6 represents:1 unsupervised:1 connectionist:1 few:1 randomly:2 densely:2 divergence:1 individual:2 amf:2 interest:2 investigate:1 inr:3 evaluation:2 mixture:26 chain:1 necessary:1 tree:1 logarithm:1 re:2 plotted:1 increased:4 formalism:1 gn:1 uniform:1 amherst:1 systematic:3 probabilistic:1 michael:1 together:1 connectivity:2 again:7 jo:1 soluble:2 satisfied:1 thesis:1 choose:1 ek:1 derivative:2 li:1 ilm:1 coefficient:3 performed:1 reached:1 tmi:1 substructure:2 ni:1 efficiently:1 yield:3 identify:1 weak:1 bayesian:4 touretzky:1 ed:2 against:1 initialised:1 pp:1 mi:3 gain:1 adjusting:1 massachusetts:1 improves:1 follow:1 modal:2 izi:1 improved:3 evaluated:3 done:1 governing:1 until:1 hand:3 horizontal:1 christopher:1 ei:3 logistic:1 quality:2 requiring:1 true:14 normalized:1 hence:4 analytically:1 leibler:1 iteratively:1 neal:3 conditionally:1 criterion:1 presenting:1 outline:1 demonstrate:2 image:2 variational:5 sigmoid:9 common:1 b4:1 exponentially:2 volume:1 discussed:1 interpretation:1 kluwer:1 marginals:4 amq:4 cambridge:1 ai:1 mathematics:1 posterior:12 optimizing:1 manipulation:1 inequality:2 binary:2 seen:1 additional:1 greater:1 determine:1 maximize:6 ii:2 full:1 unimodal:1 mix:1 adapt:2 dept:1 divided:1 feasibility:1 impact:1 involving:1 expectation:2 histogram:4 represent:2 diagram:2 wake:1 subject:1 jordan:10 feedforward:1 iii:1 zi:4 architecture:1 expression:2 six:1 useful:1 detailed:1 factorial:2 ph:1 generate:1 unifonn:1 arising:1 write:1 group:1 key:1 four:1 nevertheless:1 verified:1 sum:1 family:1 bound:29 hi:1 layer:5 simplification:2 sleep:1 constraint:1 argument:1 separable:1 developing:1 combination:1 remain:1 slightly:1 em:2 making:1 taken:2 computationally:6 equation:2 remains:1 turn:1 tractable:8 adopted:1 tightest:1 permit:1 apply:1 eight:1 alternative:1 original:1 denotes:3 remaining:1 graphical:2 const:2 restrictive:1 approximating:12 g0:1 parametric:1 gradient:4 h1m:1 street:1 parametrized:1 extent:1 fonnalism:1 illustration:1 boltzmann:1 vertical:1 hinton:4 nonnalized:1 arbitrary:1 introduced:2 optimized:2 able:2 bar:4 below:1 pattern:6 belief:14 scheme:1 improve:1 aston:1 technology:1 lk:1 coupled:1 review:1 characterises:1 relative:1 fully:1 expect:1 versus:2 summary:1 keeping:1 bias:2 side:2 institute:1 saul:8 taking:1 sparse:1 regard:1 stand:1 adaptive:1 approximate:1 uni:1 nato:1 kullback:1 vio:2 instantiation:1 summing:1 assumed:1 factorize:1 lak:1 substituted:1 clamped:1 third:2 formula:1 bishop:4 intractable:5 effectively:1 ci:1 entropy:1 hlk:2 simply:1 explore:1 expressed:1 ma:1 conditional:1 presentation:1 replace:1 considerable:3 except:2 total:1 indicating:1 stressed:1 evaluate:1 |
431 | 1,393 | Comparison of Human and Machine Word
Recognition
M. Schenkel
Dept of Electrical Eng.
University of Sydney
Sydney, NSW 2006, Australia
schenkel@sedal.usyd.edu.au
C. Latimer
Dept of Psychology
University of Sydney
Sydney, NSW 2006, AustTalia
M. Jabri
Dept of Electrical Eng.
University of Sydney
Sydney, NSW 2006, Australia
marwan@sedal. usyd.edu.au
Abstract
We present a study which is concerned with word recognition rates for
heavily degraded documents. We compare human with machine reading capabilities in a series of experiments, which explores the interaction
of word/non-word recognition, word frequency and legality of non-words
with degradation level. We also study the influence of character segmentation, and compare human performance with that of our artificial neural
network model for reading. We found that the proposed computer model
uses word context as efficiently as humans, but performs slightly worse
on the pure character recognition task.
1
Introduction
Optical Character Recognition (OCR) of machine-print document images ?has matured
considerably during the last decade. Recognition rates as high as 99.5% have been reported on good quality documents. However, for lower image resolutions (200 Dpl and
below), noisy images, images with blur or skew, the recognition rate declines considerably. In bad quality documents, character segmentation is as big a problem as the actual
character recognition. fu many cases, characters tend either to merge with neighbouring
characters (dark documents) or to break into several pieces (light documents) or both. We
have developed a reading system based on a combination of neural networks and hidden
Markov models (HMM), specifically for low resolution and degraded documents.
To assess the limits of the system and to see where possible improvements are still to be
Comparison of Human and Machine Word Recognition
95
expected, an obvious comparison is between its performance and that of the best reading
system known, the human reader. It has been argued, that humans use an extremely
wide range of context information, such as current topics, syntax and semantic analysis in
addition to simple lexical knowledge during reading. Such higher level context is very hard
to model and we decided to run a. first comparison on a word recognition task, excluding
any context beyond word knowledge.
The main questions asked for this study are: how does human performance compare with
our system when it comes to pure character recognition (no context at all) of bad quality
documents? How do they compare when word context can be used? Does character
segmentation information help in reading?
2
Data Preparation
We created as stimuli 36 data sets, each containing 144 character strings, 72 words and
72 non-words, all lower case. The data sets were generated from 6 original sets, each
COI!taining 144 unique wordsjnon-words. For each original set we used three ways to
divide the words into the different degradation levels such that each word appears once in
each degradation level. We also had two ways to pick segmented/non-segmented so that
each word is presented once segmented and once non-segmented. This counterbalancing
creates the 36 sets out of the six original ones. The order of presentation within a test set
was randomized with respect to degradation, segmentation and lexical status.
All character strings were printed in 'times roman 10 pt' font. Degradation was achieved
by photocopying and faxing the printed docJiment before scanning it at 200Dpl. Care was
taken to randomize the print position of the words such that as few systematic degradation
differences as possible were introduced.
Words were picked from a dictionary of the 44,000 most frequent words in the 'Sydney
Morning Herald'. The length of the words was restricted to be between 5 and 9 characters.
They were divided in a 3x3x2 mixed factorial model containing 3 word-frequency groups,
3 stimulus degradation levels and visually segmented/non-segmented words. The three
word-frequency groups were: 1 to 10 occurences/million (o/m) as low frequency, 11 to
40 ojm as medium frequency and 41 or more ojm as high frequency. Each participant
was presented with four examples per stimulus class (e.g. four high frequency words in
medium degradation level, not segmented).
The non-words conformed to a 2x3x2 model containing legal/illegal non-words, 3 stimulus
degradation levels and visually segmented/non-segmented strings. The illegal non-words
(e.g. 'ptvca') were generated by randomly selecting a word length between 5 and 9 characters (using the same word length frequencies as the dictionary has) and then randomly
picking characters (using the same character frequencies as the dictionary has) and keeping the unpronouncable sequences. The legal non-words (e.g. 'slunk') were generated by
using trigrams (using the dictionary to compute the trigram probabilities) and keeping
pronouncable sequences. Six examples per non-word stimulus class were used in each test
set. (e.g. six illegal non-words in high degradaton level, segmented).
3
Human Reading
There were 36 participants in the study. Participants were students and staff of the
University of Sydney, recruited by advertisement and paid for their service. They were
all native English speakers, aged between 19 and 52 with no reported uncorrected visual
deficits.
The participants viewed the images, one at a time, on a computer monitor and were asked
to type in the character string they thought would best fit the image. They had been
96
M. Schenkel, C. Latimer and M. Jabri
instructed that half of the character strings were English words and half non-words, and
they were informed about the degradation levels and the segmentation hints. Participants
were asked to be as fast and as accurate as possible. After an initial training session of 30
randomly picked character strings not from an independent training set, the participants
had a short break and were then presented with the test set, one string at a time. After a
Carriage Return was typed, time was recorded and the next word was displayed. Training
and testing took about one hour. The words were about 1-1.5cm large on the screen and
viewed at a distance of 60cm, which corresponds to a viewing angle of 1?.
4
Machine Reading
For the machine reading tests, we used our integrated segmentation/recognition system,
using a sliding window technique with a combination of a neural network and an HMM [6).
In the following we describe the basic workings without going into too much detail on the
specific algorithms. For more detailed description see (6].
A sliding window approach to word recognition performs no segmentation on the input
data of the recognizer. It consists basically of sweeping a window over the input word in
small steps. At each step the window is taken to be a tentative character and corresponding
character class scores are produced. Segmentation and recognition decisions are then made
on the basis of the sequence of character scores produced, possibly taking contextual
information into account.
In the preprocessing stage we normalize the word to a fixed height. The result is a
grey-normalized pixel map of the word. This pixel map is the input to a neural network
which estimates a posteriori probabilities of occurrence for each character given the input
in the sliding window whose length corresponds approximately to two characters. We use
a space displacement neural network (SDNN) which is a multi-layer feed-forward network
with local connections and shared weights, the layers of which perform successively higherlevel feature extraction. SDNN's are derived from Time Delay Neural Networks which have
been successfully used in speech recognition (2] and handwriting recognition (4, 1]. Thanks
to its convolutional structure the computational complexity of the sliding window approach
is kept tractable. Only about one eighth of the network connections are reevaluated for
each new input window. The outputs of the SDNN are processed by an HMM. In our
case the HMM implements character duration models. It tries to align the best scores of
the SDNN with the corresponding expected character durations. The Viterbi algorithm is
used for this alignment, determining simultaneously the segmentation and the recognition
of the word. Finding this state sequence is equivalent to finding the most probable path
through the graph which represents the HMM. Normally additive costs are used instead of
multiplicative probabilities. The HMM then selects the word causing the smallest costs.
Our best architecture contains 4 convolutional layers with a total of 50,000 parameters (6].
The training set consisted of a subset of 180,000 characters from the SEDAL database, a
low resuloution degraded document database which was collected earlier and is independent of any data used in this experiment.
4.1
The Dictionary. Model
A natural way of including a dictionary in this process, is to restrict the solution space
of the HMM to words given by the dictionary. Unfortunately this means calculating the
cost for each word in the dictionary, which becomes prohibitively slow with increasing
dictionary size (we use a combination of available dictionaries, with a total size of 98,000
words). We thus chose a two step process for the dictionary search: in a first step a list of
the most probable words is generated, using a fast-matcher technique. In the second step
the HMM costs are calculated for the words in the proposed list.
Comparison of Human and Machine Word Recognition
97
To generate the word list, we take the character string as found by the HMM without the
dictionary and calculate the edit-distance between that string and all the words in the
dictionary. The edit-distance measues how many edit operations (insertion, deletion and
substitution) are necessary to convert a given input string into a target word [3, 5]. We
now select all dictionary words that have the smallest edit-distance to the string recognized
without using the dictionary. The composed word list contains on average 10 words, and
its length varies considerably depending on the quality of the initial string.
For all words in the word list the HMM cost is now calculated and the word with the
smallest cost is the proposed dictionary word. As the calculation of the edit-distance is
much faster than the calculation of the HMM costs, the recognition speed is increased
substantially.
In a last step the difference in cost between the proposed dictionary word and the initial
string is calculated. If this difference is smaller than a threshold, the system will return the
dictionary word, otherwise the original string is returned. This allows for the recognition
of non-dictionary words. The value for the threshold determines the amount of reliance
on the dictionary. A high value will correct most words but will also force non-words to
be recognized as words. A low value, on the other hand, leaves the non-words unchanged
but doesn't help for words either. Thus the value of the threshold influences the difference
between word and non-word recognition. We chose the value such that the over-all error
rate is optimized.
4.2
The Case of Segmented data
When character segmentation is given, we know how many characters we have and where
to look for them. There is no need for an HMM and we just sum up the character
probabilities over the x-coordinate in the region corresponding to a segment. This leaves
a vector of 26 scores {the whole alphabet) for each character in the input string. With no
dictionary constraints, we simply pick the label corresponding to the highest probability
for each character. The dictionary is used in the same way, replacing the HMM scores by
calculating the word scores directly from the corresponding character probabilities.
Results
5
Recognition Performance
Machine Reading
Human Reading
0.6
0.6
-- ...
X
Non-Segmental
0
S.pncnled
-- -
~0.5
~::;0.4
1::
,..-
-------lC'
r?3
IX'!
:::::=---Non-Words
ao.2
0.1
0.1
Words
o~------------~2------------~
Degradation
Figure 1: Human Reading Performance.
Words
0?~------------~2------------~3~
Degradation
Figure 2: Machine Reading Performance.
M. Schenkel, C. Latimer and M. Jabri
98
Figure 1 depicts the recognition results for human readers. All results are per character
error rates counted by the edit-distance. All results reported as significant pass an F -test
with p < .01. As expected there was a significant interaction between error rate and
degradation and clearly non-words have higher error rates than words. Also character
segmentation has also an influence on the error rate. Segmentation seems to help slightly
more for higher degradations.
Figure 2 shows performance of the machine algorithm. Again greater degradation leads
to higher error rates and non-words have higher error rates than words. Segmentation
hints lead to significantly better recognition for all degradation levels; in fact there is no
interaction between degradation and segmentation for the machine algorithm. In general
the machine benefited more from segmentation than humans.
One would expect a smaller gain from lexical knowledge for higher en;or rates (i.e. higher
degradation) as in the limit of complete degradation all error rates will be 100%. Both
humans and machine show this 'closing of the gap .
Segmented Recognition
0.6
Non-Segmented Recognition
0.6
-Human
-Human
- - - ? Mac;binc
0.1
Words
o~------------~2------------~3~
Degradation
Figure 3: Segmented Data.
o~------------~2------------~3~
Degradation
Figure 4: Non-Segmented Data.
More interesting is the direct comparison between the error rates for humans and machine
as shown in figure 3 and figure 4. The difference for non-words reflects the difference in
ability to recognize the geometrical shape of characters without context. For degradation
levels 1 and 2, the machine has the same reading abilities as humans for segmented data
and looses only about 7% in the non-'segmented case. For degradation level 3; the machine
clearly performs worse than human readers.
The difference between word and non-word error rates reflects the ability of the participant
to use lexical knowledge. Note that the task contains word/non-word discrimination as
well as recognition. It is striking how similar the behaviour for humans and machine is
for degradation levels 1 and 2.
Timing Results
Figure 5 shows the word entry times for humans. As the main goal was to compare
recognition rates, we did not emphasize entry speed when instructing the participants.
However, we recorded the word entry time for each word (which includes inspection time
and typing). When analysing the timimg data the only interest was in relative difference
between word groups. Times were therefore? converted for each participant into a z-score
(zero mean with a standard deviation of one) and statistics were made over the z-scores
of all participants.
Non-words generally took longer to recognize than words and segmented data took longer
Comparison of Human and Machine Woni Recognition
o.s,..-------,---------.--,
?----
----
-------====---------------x
-?
Non-Segmented
0.5,..-------,.--------~
Humau Reading Times
___ . -------0
99
-Human
;..., 0.3
----
~
--------
------=~~~=
.!:!. 0 I
~
--
.
b
Jj -{).I
"E
~
-{)3
-{).S'-;-------2:-------~3:-'
-{).S'-:-1------2:-------~3--l
Degradation
Desradation
Figure 5: Human Reading Times.
Figure 6:
Times.
Non-Segmented Reading
to recognize than non-segmented for humans which we believe stems from participants not
being used to reading segmented data. When asked, participants reported difficulties in
using the segmentation lines. Interestingly this segmentation effect is significant only for
words but not for non-words.
As predicted there is also an interaction between time and degradation. Greater degradations take longer to recognize. Again, the degradation effect for time is only significant
for words but barely for non-words.
Our machine reading algorithm behaves differently in segmented and non-segmented mode
with respect to time consumption. In segmented mode, the time for evaluating the word
list in our system is very short compared to the base recognition time, as there is no
HMM involved. Accordingly we found no or very little effects on timing for our system for
segmented data. All the timing information for the machine refer to the non-segmented
case (see Figure 6).
Frequency and Legality
Table 5 shows word frequencies, legality of non-words and entry-time. Our experiment
confirmed the well known frequency and legality effect for humans in recognition rate as
well as time and respectively for frequency. The only exception is that there is no difference
in error rate for middle and low frequency words.
The machine shows (understandably) no frequency effect in error rate or time, as all
lexical words had the same prior probability. Interestingly even when using the correct
prior probabilities we could not produce a strong word frequency? effect for the machine.
Also no legality effect was observed for the error rate. One way to incorporate legality
effects would be the use of Markov chains such as n-grams.
Note however, how the recognition time for non-words is higher than for words and the
legality effect for the recognition time. Recognition times for our system in non-segmented
mode depend mainly on the time it takes to evaluate the word list. Non-words generally
produce longer word lists than words, because there are no good quality matches for a
non-word in the dictionary (on average a word list length of 8.6 words was found for words
and of 14.5 for non-words). Also illegal non-words produce longer word lists than legal
ones, again because the match quality for illegal non-words is worse than for legal ones
(average length for illegal non-words 15.9 and for legal non-words 13.2). The z-scores for
the word list length parallel nicely the recognition time scores.
In segmented mode, the time for evaluating the word list is very short compared to the
M. Schenkel, C. Latimer and M. Jabri
100
base recognition time, as there is no HMM involved. Accordingly we found no or very
little effects on timing for our system in the segmented case.
Error
l%J
Words 41+
Words 11-40
Words 1-10
Legal Non-W.
lllegal Non-W.
Humans
Error z-Time
0.22
-0.37
-0.13
0.27
0.26
-0.06
0.36
0.07
0.46
0.31
Machine
Error z-Time
0.36
-0.14
0.34
-0.19
-0.22
0.33
0.47
0.09
0.49
0.28
Table 1: Human and Machine Error rates for the different word and non-word
classes. The z-times for the machine are for the non-segmented data only.
6
Discussion
.The ability to recognize the geometrical shape of characters without the possibility to use
any sort of context information is reflected in the error rate of illegal non-words. The
difference between the error rate for illegal non-words and the one for words reflects the
ability to use lexical knowledge. To our surprise the behavior of humans and machine
is very similar for both tasks, indicating a near to optimal machine recognition system.
Clearly this does not mean our system is a good model for human reading. Many effects
such as semantic and repetition priming are not reproduced and call for a system which
is able to build semantic classes and memorize the stimuli presented. Nevertheless, we
believe that our experiment validates empirically the verification model we implemented,
using real world data.
Acknowledgments
This research is supported by a grant from the Australian Research Council (grant No
A49530190).
References
[1] I. Guyon, P. Albrecht, Y. Le Cun, J. Denker, and W. Hubbard. Design of a neural
network character recognizer for a touch terminal. Pattern Recognition, 24(2):105-119,
1991.
[2] K. J. Lang and G. E. Hinton. A Time Delay Neural Network architecture for speech
recognition. Technical Report CMU-cs-88-152, Carnegie-Mellon University, Pittsburgh
PA, 1988.
[3] V.I. Levenshtein. Binary codes capable of correcting deletions, insertions and reversals.
Soviet Physics-Doklady, 10(8):707-710, 1966.
[4] 0. Matan, C. J. C. Burges, Y. Le Cun, and J. Denker. Multi-digit recognition using
a Space Dispacement Neural Network. In J. E. Moody, editor, Advances in Neural
Information Processing Systems 4, pages 488-495, Denver, 1992. Morgan Kaufmann.
f5] T. Okuda, E. Tanaka, and K. Tamotsu. A method for the correction of garbled words
based on the Levenshtein metric. IEEE Transactions on Computers, c-25(2):172-177,
1976.
[6] M. Schenkel and M. Jabri. Degraded printed document recognition using convolutional
neural networks and hidden markov models. In Proceedings of the A CNN, Melbourne,
1997.
| 1393 |@word cnn:1 middle:1 seems:1 grey:1 eng:2 nsw:3 pick:2 paid:1 initial:3 substitution:1 series:1 score:10 selecting:1 contains:3 document:10 legality:7 interestingly:2 current:1 contextual:1 lang:1 additive:1 matured:1 blur:1 shape:2 discrimination:1 half:2 leaf:2 accordingly:2 inspection:1 short:3 height:1 direct:1 consists:1 expected:3 behavior:1 multi:2 terminal:1 actual:1 little:2 window:7 increasing:1 becomes:1 medium:2 cm:2 string:15 substantially:1 developed:1 informed:1 loos:1 finding:2 prohibitively:1 doklady:1 normally:1 grant:2 before:1 service:1 local:1 timing:4 limit:2 path:1 merge:1 approximately:1 chose:2 au:2 range:1 decided:1 unique:1 acknowledgment:1 testing:1 implement:1 digit:1 displacement:1 thought:1 printed:3 illegal:8 significantly:1 word:126 context:8 influence:3 equivalent:1 map:2 lexical:6 duration:2 resolution:2 pure:2 occurences:1 correcting:1 coordinate:1 pt:1 target:1 heavily:1 neighbouring:1 us:1 sdnn:4 pa:1 recognition:41 native:1 database:2 observed:1 electrical:2 calculate:1 region:1 highest:1 complexity:1 insertion:2 asked:4 depend:1 segment:1 creates:1 basis:1 differently:1 soviet:1 alphabet:1 fast:2 describe:1 artificial:1 matan:1 whose:1 otherwise:1 ability:5 statistic:1 noisy:1 conformed:1 validates:1 reproduced:1 higherlevel:1 sequence:4 took:3 interaction:4 frequent:1 causing:1 description:1 normalize:1 produce:3 help:3 depending:1 strong:1 sydney:8 implemented:1 uncorrected:1 predicted:1 come:1 memorize:1 australian:1 c:1 correct:2 coi:1 human:31 australia:2 viewing:1 argued:1 behaviour:1 ao:1 probable:2 correction:1 visually:2 viterbi:1 trigram:2 dictionary:23 smallest:3 recognizer:2 label:1 council:1 edit:6 hubbard:1 repetition:1 successfully:1 reflects:3 clearly:3 derived:1 improvement:1 mainly:1 posteriori:1 integrated:1 hidden:2 going:1 selects:1 pixel:2 once:3 extraction:1 nicely:1 represents:1 look:1 report:1 stimulus:6 roman:1 few:1 hint:2 randomly:3 composed:1 simultaneously:1 recognize:5 interest:1 possibility:1 alignment:1 light:1 chain:1 accurate:1 fu:1 capable:1 necessary:1 divide:1 melbourne:1 increased:1 earlier:1 cost:8 mac:1 deviation:1 subset:1 entry:4 delay:2 too:1 reported:4 scanning:1 varies:1 considerably:3 thanks:1 explores:1 randomized:1 systematic:1 physic:1 picking:1 moody:1 again:3 recorded:2 successively:1 latimer:4 containing:3 possibly:1 f5:1 worse:3 return:2 albrecht:1 account:1 converted:1 student:1 includes:1 piece:1 multiplicative:1 break:2 picked:2 try:1 sort:1 participant:12 capability:1 parallel:1 ass:1 degraded:4 convolutional:3 kaufmann:1 efficiently:1 produced:2 basically:1 reevaluated:1 confirmed:1 frequency:16 typed:1 involved:2 obvious:1 handwriting:1 gain:1 knowledge:5 segmentation:17 appears:1 feed:1 higher:8 reflected:1 just:1 stage:1 working:1 hand:1 replacing:1 touch:1 morning:1 mode:4 quality:6 believe:2 effect:11 normalized:1 consisted:1 counterbalancing:1 semantic:3 during:2 speaker:1 syntax:1 complete:1 performs:3 geometrical:2 image:6 behaves:1 empirically:1 denver:1 million:1 sedal:3 significant:4 refer:1 mellon:1 session:1 closing:1 had:4 longer:5 align:1 segmental:1 base:2 binary:1 morgan:1 greater:2 care:1 staff:1 recognized:2 sliding:4 stem:1 segmented:31 technical:1 faster:1 match:2 calculation:2 dept:3 divided:1 basic:1 cmu:1 metric:1 achieved:1 addition:1 aged:1 tend:1 recruited:1 call:1 near:1 concerned:1 fit:1 psychology:1 architecture:2 restrict:1 decline:1 six:3 returned:1 speech:2 jj:1 generally:2 detailed:1 factorial:1 amount:1 dark:1 processed:1 generate:1 per:3 carnegie:1 group:3 x3x2:2 four:2 threshold:3 reliance:1 monitor:1 nevertheless:1 kept:1 graph:1 convert:1 sum:1 run:1 angle:1 striking:1 reader:3 guyon:1 schenkel:6 decision:1 layer:3 constraint:1 speed:2 extremely:1 optical:1 combination:3 smaller:2 slightly:2 character:38 cun:2 restricted:1 taken:2 legal:6 skew:1 know:1 tractable:1 reversal:1 available:1 operation:1 denker:2 ocr:1 occurrence:1 original:4 calculating:2 build:1 unchanged:1 print:2 question:1 font:1 randomize:1 usyd:2 distance:6 deficit:1 hmm:15 consumption:1 topic:1 collected:1 barely:1 length:8 code:1 unfortunately:1 design:1 perform:1 markov:3 displayed:1 hinton:1 excluding:1 incorporate:1 sweeping:1 introduced:1 connection:2 optimized:1 tentative:1 instructing:1 deletion:2 hour:1 tanaka:1 beyond:1 able:1 below:1 pattern:1 eighth:1 reading:20 including:1 natural:1 force:1 typing:1 difficulty:1 dispacement:1 created:1 prior:2 determining:1 relative:1 expect:1 mixed:1 interesting:1 verification:1 editor:1 supported:1 last:2 keeping:2 english:2 burges:1 wide:1 taking:1 calculated:3 evaluating:2 gram:1 world:1 doesn:1 instructed:1 made:2 forward:1 preprocessing:1 taining:1 dpl:2 counted:1 transaction:1 emphasize:1 status:1 pittsburgh:1 marwan:1 search:1 decade:1 table:2 priming:1 jabri:5 understandably:1 did:1 main:2 big:1 whole:1 benefited:1 en:1 screen:1 depicts:1 slow:1 lc:1 position:1 advertisement:1 ix:1 bad:2 specific:1 list:12 gap:1 surprise:1 simply:1 visual:1 corresponds:2 determines:1 viewed:2 presentation:1 goal:1 shared:1 hard:1 analysing:1 specifically:1 degradation:28 total:2 pas:1 matcher:1 exception:1 select:1 indicating:1 preparation:1 levenshtein:2 evaluate:1 |
432 | 1,394 | Combining Classifiers Using
Correspondence Analysis
Christopher J. Merz
Dept. of Information and Computer Science
University of California, Irvine, CA 92697-3425 U.S.A.
cmerz@ics.uci.edu
Category: Algorithms and Architectures.
Abstract
Several effective methods for improving the performance of a single learning algorithm have been developed recently. The general
approach is to create a set of learned models by repeatedly applying the algorithm to different versions of the training data, and
then combine the learned models' predictions according to a prescribed voting scheme. Little work has been done in combining the
predictions of a collection of models generated by many learning
algorithms having different representation and/or search strategies.
This paper describes a method which uses the strategies of stacking and correspondence analysis to model the relationship between
the learning examples and the way in which they are classified by
a collection of learned models. A nearest neighbor method is then
applied within the resulting representation to classify previously
unseen examples. The new algorithm consistently performs as well
or better than other combining techniques on a suite of data sets.
1
Introduction
Combining the predictions of a set of learned models! to improve classification
and regression estimates has been an area of much research in machine learning and neural networks [Wolpert, 1992, Merz and Pazzani, 1997, Perrone, 1994,
Breiman, 1996, Meir, 1995]. The challenge of this problem is to decide which models
to rely on for prediction and how much weight to give each. The goal of combining
learned models is to obtain a more accurate prediction than can be obtained from
any single source alone.
1
A learned model may be anything from a decision/regression tree to a neural network.
C. l Men
592
Recently, several effective methods have been developed for improving the performance of a single learning algorithm by combining multiple learned models generated using the algorithm. Some examples include bagging [Breiman, 1996], boosting
[Freund, 1995], and error correcting output codes [Kong and Dietterich, 1995]. The
general approach is to use a particular learning algorithm and a model generation technique to create a set of learned models and then combine their predictions according to a prescribed voting scheme. The models are typically generated
by varying the training data using resampling techniques such as bootstrapping
[Efron and Tibshirani, 1993J or data partitioning [Meir, 1995] . Though these methods are effective, they are limited to a single learning algorithm by either their
model generation technique or their method of combining.
Little work has been done in combining the predictions of a collection of models
generated by many learning algorithms each having different representation and/or
search strategies. Existing approaches typically place more emphasis on the model
generation phase rather than the combining phase [Opitz and Shavlik, 1996]. As a
result, the combining method is rather limited. The focus of this work is to present
a more elaborate combining scheme, called SCANN, capable of handling any set
of learned models , and evaluate it on some real-world data sets. A more detailed
analytical and empirical study of the SCANN algorithm is presented in [Merz, 1997] .
This paper describes a combining method applicable to model sets that are homogeneous or heterogeneous in their representation and/or search techniques. Section 2
describes the problem and explains some of the caveats of solving it. The SCANN
algorithm (Section 3), uses the strategies of stacking [Wolpert, 1992J and correspondence analysis (Greenacre, 1984] to model the relationship between the learning examples and the way in which they are classified by a collection of learned models . A
nearest neighbor method is then applied to the resulting representation to classify
previously unseen examples.
In an empirical evaluation on a suite of data sets (Section 4), the naive approach of
taking the plurality vote (PV) frequently exceeds the performance of the constituent
learners. SCANN, in turn, matches or exceeds the performance of PV and several
other stacking-based approaches. The analysis reveals that SCANN is not sensitive
to having many poor constituent learned models , and it is not prone to overfit by
reacting to insignificant fluctuations in the predictions of the learned models.
2
Problem Definition and Motivation
The problem of generating a set of learned models is defined as follows. Suppose
two sets of data are given: a learning set C = {(Xi, Yi), i = 1, .. . ,I} and a test set
T = {(Xt, yd, t = 1, .. . , T}. Xi is a vector of input values which are either nominal
or numeric values, and Yi E {Cl , ... , Cc} where C is the number of classes. Now
suppose C is used to build a set of N functions, :F {fn (x)}, each element of which
approximates f(x) , the underlying function.
=
The goal here is to combine the predictions of the members of :F so as to find
the best approximation of f(x). Previous work [Perrone, 1994] has indicated that
the ideal conditions for combining occur when the errors of the learned models are
uncorrelated. The approaches taken thus far attempt to generate learned models
which make uncorrelated errors by using the same algorithm and presenting different
samples of the training data [Breiman, 1996, Meir, 1995], or by adjusting the search
heuristic slightly [Opitz and Shavlik, 1996, Ali and Pazzani, 1996J.
No single learning algorithm has the right bias for a broad selection of problems.
Combining Classifiers Using Correspondence Analysis
593
Therefore, another way to achieve diversity in the errors of the learned models
generated is to use completely different learning algorithms which vary in their
method of search and/or representation. The intuition is that the learned models
generated would be more likely to make errors in different ways. Though it is not
a requirement of the combining method described in the next section, the group of
learning algorithms used to generate :F will be heterogeneous in their search and/or
representation methods (i.e., neural networks, decision lists, Bayesian classifiers,
decision trees with and without pruning, etc .). In spite of efforts to diversify the
errors committed, it is still likely that some of the errors will be correlated because
the learning algorithms have the same goal of approximating f, and they may use
similar search strategies and representations. A robust combining method must
take this into consideration.
Approach
3
The approach taken consists of three major components: Stacking, Correspondence
Analysis, and Nearest Neighbor (SCANN) . Sections 3.1-3.3 give a detailed description of each component, and section 3.4 explains how they are integrated to form
the SCANN algorithm.
3.1
Stacking
Once a diverse set of models has been generated, the issue of how to combine them
arises. Wolpert (Wolpert, 1992] provided a general framework for doing so called
stacked genemiization or stacking. The goal of stacking is to combine the members
of:F based on information learned about their particular biases with respect to ?2 .
The basic premise of stacking is that this problem can be cast as another induction
problem where the input space is the (approximated) outputs of the learned models,
and the output space is the same as before, i.e.,
The approximated outputs of each learned model, represented as jn(Xi), are generated using the following in-sample/out-of-sample approach:
1. Divide the ?0 data up into V partitions.
2. For each partition, v,
? Train each algorithm on all but partition v to get {j;V}.
? Test each learned model in {j;V} on partition v.
? Pair the predictions on each example in partition v (i.e., the new input
space) with the corresponding output, and append the new examples
to ?1
3. Return ?1
3.2
Correspondence Analysis
Correspondence Analysis (CA) (Greenacre, 1984] is a method for geometrically exploring the relationship between the rows and columns of a matrix whose entries
are categorical. The goal here is to explore the relationship between the training
2Henceforth ? will be referred to as ?0 for clarity.
c. 1. Men
594
Stage
1
Symbol
N
n
r
2
3
c
P
Dc
Dr
A
A
F
G
Table 1? Correspondence Analysis calculations.
Description
Definition
Records
votes of learned models .
(I x J) indicator matrix
Grand total of table N.
i=1 jJ=1 nij
Row masses.
ri = ni+/n
Column masses .
cj=n+j/n
Correspondence matrix.
(1/n)N
(J x J) diagonal matrix Masses c on diagonal.
Masses r on diagonal.
(I X I) diagonal matrix
Dr -1/2(p _ rcT)Dc -1/2 Standardized residuals.
urv'!
SVD of A.
Dr -1/2ur
Principal coordinates of rows.
Dc -1/2vr
Principal coordinates of columns.
2:1 2:
examples and how they are classified by the learned models. To do this, the prediction matrix, M, is explored where min = in (xd (1 ::; i ::; I, and 1 ::; n ::; N). It is
also important to see how the predictions for the training examples relate to their
true class labels, so the class labels are appended to form M' , an (I x J) matrix
(where J = N + 1). For proper application of correspondence analysis, M' must be
converted to an (I x (J . C)) indicator matrix, N, where ni,(joJ+e) is a one exactly
when mij = ee, and zero otherwise.
The calculations of CA may be broken down into three stages (see Table 1). Stage
one consists of some preprocessing calculations performed on N which lead to the
standardized residual matrix, A . In the second stage, a singular value decomposition
(SVD) is performed on A to redefine it in terms ofthree matrices: U(lXK), r(KxK) '
and V(KXJ), where K = min(I - 1, J - 1) . These matrices are used in the third
stage to determine F(lXK) and G(JxK) , the coordinates of the rows and columns
of N, respectively, in the new space . It should be noted that not all K dimensions
are necessary. Section 3.4, describes how the final number of dimensions, K *, is
determined .
Intuitively, in the new geometric representation , two rows, f p* and fq*, will lie close
to one another when examples p and q receive similar predictions from the collection
of learned models. Likewise, rows gr* and gu will lie close to to one another when
the learned models corresponding to r and s make similar predictions for the set
of examples. Finally, each column, r, has a learned model, j', and a class label, c',
with which it is associated; f p * will lie closer to gr* when model j' predicts class c'.
3.3
Nearest Neighbor
The nearest neighbor algorithm is used to classify points in a weighted Euclidean
space. In this scenario, each possible class will be assigned coordinates in the space
derived by correspondence analysis. Unclassified examples will be mapped into the
new space (as described below) , and the class label corresponding to the closest
class point is assigned to the example .
Since the actual class assignments for each example reside in the last C columns of
N, their coordinates in the new space can be found by looking in the last Crows
of G. For convenience, these class points will be called Class!, . .. , Classc .
To classify an unseen example, XTest, the predictions of the learned models on XTest
must be converted to a row profile, rT , oflength J . C, where r& oJ+e) is 1/ J exactly
Combining Classifiers Using Correspondence Analysis
Data set
abalone
bal
breast
credit
dementia
glass
heart
ionosphere
..
1flS
krk
liver
lymphography
musk
retardation
sonar
vote
wave
wdbc
Table 2: Experimental results.
PV
SCANN S-BP
S-BAYES
vs PV
vs PV vs PV
ratio
ratio
ratio
error
80.35 .490
.499
.487
13.81 .900
.859
.992
4.31
.886
.881
.920
13.99 .999
1.012
1.001
32.78
.989
.932
1.037
1.158 1.215
31.44 1.008
18.17 .964
.998
.972
3.05
.691
1.289 1.299
4.44
1.017
1.467
1.033
1.030
1.080
1.149
6.67
1.024
1.035
1.077
29.33
1.162 1.100
17.78 1.017
.812
.889
.835
13.51
32.64 .970
.960
.990
23.02
1.079
.990
1.007
.903
.908
.893
5.24
21.94
1.008
1.109 1.008
4.27
1.000
1.103
1.007
= ee,
595
Best Ind.
vs PV
ratio
.535 111'
.911 BP
.938BP
1.054 BP
1.048c4 . 5
1.155 0C1
.962BP
2.175 c4 .5
1.150 oc1
1. 159 NN
1.138 cN2
.983Pebl8
1. 113Peb13
.936Baye3
1.048 BP
.927 c4 .5
1.200Pebl8
1. 164NN
when mij
and zero otherwise. However, since the example is unclassified,
XTe3t is of length (J - 1) and can only be used to fill the first (( J - 1) .
entries
in iT. For this reason , C different versions are generated, i.e., iT, . .. , i c , where
each one "hypothesizes" that XTe3t belongs to one of the C classes (by putting 1/ J
in t~e appropr~ate col~~) .. Loc~ting thes=l.rofiles in the scale~ sp~ce is a matter
of s1mple matflx multIphcatIOn, 1.e., f'[ = re Gr- 1. The f'[ wh1ch lies closest to a
class point, say Classc') is considered the "correct" hypothesized class, and XTe3t
is assigned the class label c' .
3.4
C)
The SCANN Algorithm
Now that the three main parts of the approach have been described, a summary of
the SCANN algorithm can be given as a function of Co and the constituent learning
algorithms, A. The first step is to use Co and A to generate the stacking data, C 1 ,
capturing the approximated predictions of each learned model. Next, C1 is used
to form the indicator matrix, N. A correspondence analysis is performed on N to
derive the scaled space, A = urvT. The number of dimensions retained from
this new representation, K *, is the value which optimizes classification on C 1 . The
resulting scaled space is used to derive the row/column coordinates F and G, thus
geometrically capturing the relationships between the examples, the way in which
they are classified, and their position relative to the true class labels. Finally, the
nearest neighbor strategy exploits the new representation by predicting which class
is most likely according to the predictions made on a novel example.
596
4
C. J Merz
Experimental Results
The constituent learning algorithms, A, spanned a variety of search and/or
representation techniques: Backpropagation (BP) [Rumelhart et al., 1986], CN2
[Clark and Niblett, 1989], C4.5 [Quinlan, 1993], OC1 [Salzberg; and Beigel, 1993],
PEBLS [Cost, 1993], nearest neighbor (NN), and naive Bayes. Depending on the
data set, anywhere from five to eight instantiations of algorithms were applied. The
combining strategies evaluated were PV, SCANN, and two other learners trained
on ?1: S-BP, and S-Bayes.
The data sets used were taken from the UCI Machine Learning Database Repository
[Merz and Murphy, 1996], except for the unreleased medical data sets: retardation
and dementia. Thirty runs per data set were conducted using a training/test partition of 70/30 percent. The results are reported in Table 2. The first column gives
the mean error rate over the 30 runs of the baseline combiner, PV. The next three
columns ("SCANN vs PV", "S-BP vs PV", and "S-Bayes vs PV") report the ratio
of the other combining strategies to the error rate of PV. The column labeled "Best
Ind . vs PV" reports the ratio with respect to the model with the best average error
rate. The superscript of each entry in this column denotes the winning algorithm.
A value less than 1 in the "a vs b" columns represents an improvement by method a
over method b. Ratios reported in boldface indicate the difference between method
a and method b is significant at a level better than 1 percent using a two-tailed sign
test.
It is clear that, over the 18 data sets, SCANN holds a statistically significant advantage on 7 sets improving upon PV's classification error by 3-50 percent. Unlike the
other combiners, SCANN posts no statistically significant losses to PV (i.e., there
were 4 losses each for S-BP and S-Bayes). With the exception of the retardation
data set, SCANN consistently performs as well or better than the best individual
learned model. In the direct comparison of SCANN with the S-BP and S-Bayes,
SCANN posts 5 and 4 significant wins, respectively, and no losses.
The most dramatic improvement of the combiners over PV came in the abalone
data set. A closer look at the results revealed that 7 of the 8 learned models
were very poor classifiers with error rates around 80 percent, and the errors of the
poor models were highly correlated. This empirically demonstrates PV's known
sensitivity to learned models with highly correlated errors. On the other hand, PV
performs well on the glass and wave data sets where the errors of the learned models
are measured to be fairly uncorrelated. Here, SCANN performs similarly to PV,
but S-BP and S-Bayes appear to be overfitting by making erroneous predictions
based on insignificant variations on the predictions of the learned models.
5
Conclusion
A novel method has been introduced for combining the predictions of heterogeneous
or homogeneous classifiers. It draws upon the methods of stacking, correspondence
analysis and nearest neighbor. In an empirical analysis, the method proves to be
insensitive to poor learned models and matches the performance of plurality voting
as the errors of the learned models become less correlated.
References
[Ali and Pazzani, 1996) Ali, K. and Pazzani, M. (1996). Error reduction through
learning multiple descriptions. Machine Learning, 24:173.
Combining Classifiers Using Correspondence Analysis
[Breiman, 1996] Breiman, L. (1996).
24(2):123-40.
Bagging predictors.
597
Machine Learning,
[Clark and Niblett, 1989] Clark, P. and Niblett, T. (1989). The CN2 induction
algorithm. Machine Learning, 3(4):261-283.
[Cost, 1993] Cost, S.; Salzberg, S. (1993). A weighted nearest neighbor algorithm
for learning with symbolic features. Machine Learning, 10(1):57-78.
[Efron and Tibshirani, 1993] Efron, B. and Tibshirani, R. (1993). An Introduction
to the Bootstrap. Chapman and Hall, London and New York.
[Freund, 1995] Freund, Y. (1995). Boosting a weak learning algorithm by majority.
Information and Computation, 121(2):256-285. Also appeared in COLT90.
[Greenacre, 1984] Greenacre, M. J. (1984). Theory and Application of Correspondence Analysis. Academic Press, London.
[Kong and Dietterich, 1995] Kong, E. B. and Dietterich, T. G. (1995). Errorcorrecting output coding corrects bias and variance. In Proceedings of the 12th
International Conference on Machine Learning, pages 313-321. Morgan Kaufmann.
[Meir, 1995] Meir, R. (1995). Bias, variance and the combination ofleast squares estimators. In Tesauro, G., Touretzky, D., and Leen, T., editors, Advances in Neural
Information Processing Systems, volume 7, pages 295-302. The MIT Press.
[Merz, 1997] Merz, C. (1997). Using correspondence analysis to combine classifiers.
Submitted to Machine Learning.
[Merz and Murphy, 1996] Merz, C. and Murphy, P. (1996). UCI repository of machine learning databases.
[Merz and Pazzani, 1997] Merz, C. J. and Pazzani, M. J. (1997). Combining neural network regression estimates with regularized linear weights. In Mozer, M.,
Jordan, M., and Petsche, T., editors, Advances in Neural Information Processing
Systems, volume 9. The MIT Press.
[Opitz and Shavlik, 1996] Opitz, D. W. and Shavlik, J. W. (1996). Generating
accurate and diverse members of a neural-network ensemble. In Touretzky, D. S.,
Mozer, M. C., and Hasselmo, M. E., editors, Advances in Neural Information
Processing Systems, volume 8, pages 535-541. The MIT Press.
[Perrone, 1994] Perrone, M. P. (1994). Putting it all together: Methods for combining neural networks. In Cowan, J. D., Tesauro, G., and Alspector, J., editors,
Advances in Neural Information Processing Systems, volume 6, pages 1188-1189.
Morgan Kaufmann Publishers, Inc.
[Quinlan, 1993] Quinlan, R. (1993). G..4-5 Programs for Machine Learning. Morgan
Kaufmann, San Mateo, CA.
[Rumelhart et al., 1986] Rumelhart, D. E., Hinton, G. E., and Williams, R. J.
(1986). Learning internal representations by error propagation. In Rumelhart,
D. E., McClelland, J. 1., and the PDP research group., editors, Parallel distributed processing: Explorations in the microstructure of cognition, Volume 1:
Foundations. MIT Press.
[Salzberg; and Beigel, 1993] Salzberg;, S. M. S. K. S. and Beigel, R. (1993). OC1:
Randomized induction of oblique decision trees. In Proceedings of AAAI-93.
AAAI Pres.
[Wolpert, 1992] Wolpert, D. H. (1992). Stacked generalization. Neural Networks,
5:241-259.
| 1394 |@word kong:3 repository:2 version:2 decomposition:1 xtest:2 dramatic:1 reduction:1 loc:1 existing:1 must:3 fn:1 partition:6 resampling:1 alone:1 v:9 oblique:1 record:1 caveat:1 boosting:2 five:1 direct:1 become:1 consists:2 combine:6 redefine:1 alspector:1 frequently:1 little:2 actual:1 provided:1 underlying:1 mass:4 developed:2 bootstrapping:1 suite:2 voting:3 xd:1 exactly:2 classifier:8 scaled:2 demonstrates:1 partitioning:1 medical:1 appear:1 before:1 thes:1 reacting:1 fluctuation:1 yd:1 emphasis:1 mateo:1 co:2 limited:2 statistically:2 thirty:1 backpropagation:1 bootstrap:1 area:1 empirical:3 spite:1 symbolic:1 get:1 convenience:1 close:2 selection:1 applying:1 williams:1 correcting:1 estimator:1 fill:1 spanned:1 coordinate:6 variation:1 suppose:2 nominal:1 homogeneous:2 us:2 element:1 rumelhart:4 approximated:3 predicts:1 database:2 labeled:1 intuition:1 mozer:2 broken:1 trained:1 solving:1 ali:3 upon:2 learner:2 completely:1 gu:1 kxj:1 represented:1 stacked:2 train:1 effective:3 london:2 whose:1 heuristic:1 say:1 otherwise:2 unseen:3 final:1 superscript:1 advantage:1 analytical:1 uci:3 combining:23 achieve:1 description:3 constituent:4 requirement:1 generating:2 derive:2 depending:1 liver:1 measured:1 nearest:9 indicate:1 correct:1 exploration:1 explains:2 premise:1 microstructure:1 generalization:1 plurality:2 exploring:1 hold:1 around:1 credit:1 ic:1 considered:1 hall:1 cognition:1 major:1 vary:1 applicable:1 label:6 sensitive:1 hasselmo:1 create:2 pres:1 weighted:2 mit:4 rather:2 breiman:5 varying:1 combiner:1 derived:1 focus:1 improvement:2 consistently:2 fq:1 baseline:1 glass:2 nn:3 typically:2 integrated:1 issue:1 classification:3 musk:1 fairly:1 once:1 having:3 chapman:1 represents:1 broad:1 look:1 report:2 individual:1 murphy:3 phase:2 attempt:1 highly:2 evaluation:1 accurate:2 capable:1 closer:2 necessary:1 tree:3 divide:1 euclidean:1 re:1 nij:1 classify:4 column:12 salzberg:4 assignment:1 stacking:10 cost:3 entry:3 predictor:1 conducted:1 gr:3 reported:2 grand:1 sensitivity:1 international:1 randomized:1 corrects:1 together:1 aaai:2 henceforth:1 dr:3 return:1 converted:2 diversity:1 coding:1 matter:1 hypothesizes:1 inc:1 performed:3 doing:1 wave:2 bayes:7 parallel:1 appended:1 square:1 ni:2 variance:2 kaufmann:3 likewise:1 ensemble:1 weak:1 bayesian:1 cc:1 classified:4 submitted:1 touretzky:2 definition:2 associated:1 irvine:1 adjusting:1 efron:3 cj:1 leen:1 done:2 though:2 evaluated:1 anywhere:1 stage:5 overfit:1 hand:1 christopher:1 propagation:1 indicated:1 dietterich:3 hypothesized:1 true:2 assigned:3 ind:2 noted:1 anything:1 abalone:2 bal:1 presenting:1 performs:4 percent:4 consideration:1 novel:2 recently:2 empirically:1 insensitive:1 volume:5 approximates:1 oflength:1 significant:4 diversify:1 similarly:1 etc:1 closest:2 belongs:1 optimizes:1 tesauro:2 scenario:1 came:1 yi:2 morgan:3 determine:1 multiple:2 exceeds:2 match:2 academic:1 calculation:3 post:2 prediction:20 regression:3 basic:1 breast:1 heterogeneous:3 c1:2 receive:1 singular:1 source:1 publisher:1 unlike:1 cowan:1 member:3 jordan:1 ee:2 ideal:1 revealed:1 variety:1 architecture:1 jxk:1 effort:1 york:1 jj:1 repeatedly:1 detailed:2 clear:1 category:1 mcclelland:1 generate:3 meir:5 ofthree:1 sign:1 tibshirani:3 per:1 diverse:2 group:2 putting:2 clarity:1 ce:1 geometrically:2 run:2 unclassified:2 place:1 decide:1 draw:1 decision:4 capturing:2 fl:1 correspondence:17 occur:1 bp:12 ri:1 prescribed:2 min:2 according:3 combination:1 perrone:4 poor:4 describes:4 slightly:1 ate:1 ur:1 making:1 intuitively:1 errorcorrecting:1 taken:3 heart:1 previously:2 turn:1 rct:1 eight:1 petsche:1 jn:1 bagging:2 standardized:2 denotes:1 include:1 quinlan:3 retardation:3 exploit:1 ting:1 build:1 prof:1 approximating:1 opitz:4 strategy:8 rt:1 diagonal:4 win:1 mapped:1 majority:1 evaluate:1 reason:1 induction:3 boldface:1 code:1 length:1 retained:1 relationship:5 ratio:7 relate:1 append:1 proper:1 hinton:1 looking:1 committed:1 dc:3 pdp:1 introduced:1 cast:1 pair:1 c4:4 california:1 learned:35 below:1 appeared:1 challenge:1 oc1:3 program:1 oj:1 rely:1 regularized:1 predicting:1 indicator:3 residual:2 greenacre:4 scheme:3 improve:1 categorical:1 naive:2 geometric:1 relative:1 freund:3 loss:3 men:2 generation:3 clark:3 foundation:1 cn2:3 editor:5 uncorrelated:3 row:8 prone:1 summary:1 last:2 bias:4 combiners:2 shavlik:4 neighbor:9 taking:1 distributed:1 dimension:3 world:1 numeric:1 reside:1 collection:5 made:1 preprocessing:1 san:1 far:1 lymphography:1 pruning:1 overfitting:1 reveals:1 instantiation:1 xi:3 search:8 sonar:1 tailed:1 table:5 classc:2 crow:1 pazzani:6 ca:4 correlated:4 robust:1 improving:3 cl:1 krk:1 sp:1 main:1 motivation:1 profile:1 referred:1 elaborate:1 vr:1 position:1 pv:20 winning:1 col:1 lie:4 third:1 down:1 erroneous:1 xt:1 dementia:2 symbol:1 list:1 insignificant:2 explored:1 ionosphere:1 wdbc:1 wolpert:6 likely:3 explore:1 kxk:1 mij:2 goal:5 determined:1 except:1 principal:2 called:3 total:1 svd:2 merz:11 experimental:2 vote:3 exception:1 internal:1 arises:1 beigel:3 dept:1 lxk:2 handling:1 |
433 | 1,395 | A 1,OOO-Neuron System with One
Million 7-bit Physical Interconnections
Yuzo Hirai
Institute of Information Sciences and Electronics
University of Tsukuba
1-1-1 Ten-nodai, Tsukuba, Ibaraki 305, Japan
e-mail: hirai@is.tsukuba.ac.jp
Abstract
An asynchronous PDM (Pulse-Density-Modulating) digital neural
network system has been developed in our laboratory. It consists
of one thousand neurons that are physically interconnected via one
million 7-bit synapses. It can solve one thousand simultaneous
nonlinear first-order differential equations in a fully parallel and
continuous fashion. The performance of this system was measured
by a winner-take-all network with one thousand neurons. Although
the magnitude of the input and network parameters were identical for each competing neuron, one of them won in 6 milliseconds.
This processing speed amounts to 360 billion connections per second. A broad range of neural networks including spatiotemporal
filtering, feedforward, and feedback networks can be run by loading
appropriate network parameters from a host system.
1
INTRODUCTION
The hardware implementation of neural networks is crucial in order to realize the
real-time operation of neural functions such as spatiotemporal filtering, learning
and constraint processings. Since the mid eighties, many VLSI chips and systems
have been reported in the literature, e.g. [1] [2]. Most of the chips and the systems
including analog and digital implementations, however, have focused on feedforward
neural networks. Little attention has been paid to the dynamical aspect of feedback neural networks, which is especially important in order to realize constraint
processings, e.g. [3]. Although there were a small number of exceptions that used
analog circuits [4] [5], their network sizes were limited as compared to those of their
feedforward counterparts because of wiring problems that are inevitable in regard
to full and physical interconnections. To relax this problem, a pulse-stream system
has been used in analog [6] and digital implementations [7].
Y. Hirai
706
The author developed a fully interconnected 54-neuron system that uses an asynchronous PDM (Pulse-Density-Modulating) digital circuit system [8]. The present
paper describes a thousand-neuron system in which all of the neurons are physically interconnected via one million 7-bit synapses in order to create a fully parallel
feedback system. The outline of this project was described in [10]. In addition to
the enlargement of system size, synapse circuits were improved and time constant
of each neuron was made variable. The PDM system was used because it can accomplish faithful analog data transmission between neurons and can relax wiring
problems. An asynchronous digital circuit was used because it can solve scaling
problems, and we could also use it to connect more than one thousand VLSI chips,
as described below.
2
NEURON MODEL AND THE CIRCUITS
2.1
SINGLE NEURON MODEL
The behavior of each neuron in the system can be described by the following nonlinear first-order differential equation:
dyi(t)
Iti--;];t
Yi{ t)
<pta]
N
=
=
=
-viet) +
L WijYj{t) + li{t),
j=l
<p[yi{t)], and
if a > 0
{~
(1)
(2)
(3)
otherwise,
where Iti is a time constant of the i-th neuron, y;(t) is an internal potential of the
i-th neuron at time t, Wij is a synaptic weight from the j-th to the i-th neurons,
and li(t) is an external input to the i-th neuron. <pta] is an analog threshold output
function which becomes saturated at a given maximum val~e.
The system solves Eq.{l) in the following integral form:
yi{t) = (t {-VieT)
10
+
t
j=l
WijYj(T) + h(T)}
d~Itl + yi(O),
(4)
where y;(O) is an initial value. An analog output of a neuron is expressed by a
pulse stream whose frequency is proportional to the positive, instantaneous internal
potential.
2.2
2.2.1
SINGLE NEURON CmCUIT
Synapse circuits
The circuit diagrams for a single neuron are shown in Fig. 1. As shown in Fig.l(a),
it consists of synapse circuits, excitatory and inhibitory dendrite OR circuits, and
a cel~ body circuit. Each synapse circuit transforms the instantaneous frequency of
the input pulses to a frequency that is proportional to the synaptic weight. This
transformation is carried out by a 6-bit rate multiplier, as shown in Fig.l{b). The
behavior of a rate multiplier is illustrated in Fig.l(c) using a 3-bit case for brevity.
A rate multiplier is a counter and its state transits to the next state when an input
pulse occurs. Each binary bit of a given weight specifies at which states the output
pulses are generated. When the LSB is on, an output pulse is generated at the
fourth state. When the second bit is on, output pulses are generated at the second
A I,OOO-Neuron System with One Million 7-bit Physical Interconnections
Neuron circuit
Dendrite extension terminals
~
U
0_0
J
-
J JUL :
Synaptic
weight
(7-blt)
~~
U
Rate
multiplier
(&-bit)
wt
1
+
~
j ~ Synapse ~ j
I circuit I - a:
a:
nIl!..
'.:;g::.
-~
InJ)Ut pulses
JUL ~.?
tp@
Synapse circuit
-'L
r,:,.
~
@
(Circuit
sign
C
CI@l?rJUL
I.
II
Cell body circuit
f<:IrI
\51
I
@. JlJlJL
Output pulses
(a)
'w
1
1
1
1
1
1
1 1 1
1
1
1
1 1 1
11 1 1
1111111
(c)
sign
Cell body circuit
-
1
1
1
101
110
111
(b)
"""
states
01234567
100
J
~
weight
000
010
011
ia _.ISy~pse rI e ~ Lt . : : \ @ ,
,.=.
~
3-blt rate multiplier
001
(Osw<1)
i ..
707
r-=-~+'_:_--,
UDr-----.,
Sampling ~ Up-down f-- Rate
circuit
Down counter ~ ~ multiplier
2f Output
;. gate and
f~~~_(1~2-b-+t=)__I--(_12~-br....It)_
=~
Main clock =4fma (2OMHz)
-:-='112
I~
r=----,~-..~
Down
?
L-----:~---II X (-1)
r
Up'
Rate
multiplier
(&-bit)
I
If
f.J
for2f
H slgn=O
(d)
Figure 1: Circuit diagram of a single neuron. (a) Circuit diagram of a single neuron
and (b) that of a synapse circuit. (c) To illustrate the function of a rate multiplier,
the multiplication table for a 3-bit case is shown. (d) Circuit diagram of a cell body
circuit. See details in text.
and at the sixth states. When the MSB is on, they are generated at all of the odd
states. Therefore, the magnitude of synaptic weight that can be represented by
a rate multiplier is less than one. In our circuit, this limitation was overcome by
increasing the frequency of a neuron output by a factor of two, as described below.
2.2.2
Dendrite circuits
Output pulses from a synapse circuit are fed either to an excitatory dendrite OR circuit or to an inhibitory one, according to the synaptic weight sign. In each dendrite
OR circuit, the synaptic output pulses are summed by OR gates, as is shown along
the right side of Fig.1(a). Therefore, if these output pulses are synchronized, they
are counted as one pulse and linear summation cannot take place. In our circuit,
each neuron is driven by an individual clock oscillator. Therefore, they will tend to
become desynchronized. The summation characteristic was analysed in [9], and it
was shown to have a saturation characteristic that is similar to the positive part of
a hyperbolic tangent function.
2.2.3
Cell body circuit
A cell body circuit performs the integration given by Eq.( 4) as follows. As shown
in Fig.1(d), integration was performed by a 12-bit up-down counter. Input pulses
from an excitatory dendrite OR circuit are fed into the up-input of the counter and
those from an inhibitory one are fed into the down-input after conflicts between
Y. Hirai
708
excitatory and inhibitory pulses have been resolved by a sampling circuit. A 12-bit
rate multiplier produces internal pulses whose frequency is 21, where 1 is proportional to the absolute value of the counter. The rate multiplier is driven by a main
clock whose frequency is 4/max, Imax being the maximum output frequency. When
the counter value is positive, an output pulse train whose frequency is either 1 or
2/, according to the scale factor is transmitted from a cell body circuit.
The negative feedback term that appeared in the integrand of Eq.( 4) can be realized
by feeding the internal pulses into the down-input of the counter when the counter
value is positive and feeding them into the up-input when it is negative. The 6bit rate multiplier inserted in this feedback path changes the time constant of a
neuron. Let f3i be the rate value of the rate multiplier, where 0 ~ f3i < 26 . The
Eq.(4) becomes:
yi(t)
(5)
26
211
Therefore, the time constant changes to l!i?,
where Pi was given by -,seconds.
/3
max
It should be noted that, since the magnitUde of the total input was increased by a
factor of ~, the strength of the input should be decreased by the inverse of that
factor in order to maintain an appropriate output level. If it is not adjusted, we
can increase the input strength. Therefore, the system has both input and output
scaling functions. The time constant varies from about 416psec for f3 = 63 to
26.2msec for f3 = 1. When f3 = 0, the negative feedback path is interrupted and
the circuit operates as a simple integrator, and every feedforward network can be
run in this mode of operation.
3
3.1
THE 1,OOO-NEURON SYSTEM
VLSI CHIP
A single type of VLSI chip was fabricated using a 0.7 pm CMOS gate array with
250,000 gates. A single chip contains 18 neurons and 51 synapses for each neuron.
Therefore, each chip has a total of 918 synapses. About 85% of the gates in a gate
array could be used, which was an extremely efficient value. A chip was mounted
on a flat package with 256 pins. Among them, 216 pins were used for signals and
the others were used for twenty pairs of V cc( =3.3V) and GND.
3.2
THE SYSTEM
As illustrated in Fig.2(a), this system consists of 56 x 20 = 1,120 chips. 56 chips
are used for both cell bodies and synapses, and the others are used to extend
dendrite circuits and increase the number of synapses. In order to extend the
dendrites, the dendrite signals in a chip can be directly transmitted to the dendrite
extention terminals of another chip by bypassing the cell body circuits. There are
51 x 20 1,020 synapses per neuron. Among them, 1,008 synapses are used for fully
hard-wired interconnections and the other 12 synapses are used to receive external
signals. There are a total of 1,028,160 synapses in this system. It is controlled by
a personal computer. The synaptic weights, the contents of the up-down counters
=
A 1,OOO-Neuron System with One Million 7-bit Physical Interconnections
Host computer
709
High-speed bus
12 external
inputs
m Neural chips
o Neural chips
for neurons
for synapses
(a)
(b)
Figure 2: Structure of the system. (a) System configuration. The down arrows
emitted from the open squares designate signal lines that are extending dendrites.
The others designate neuron outputs. (b) Exterior of the system. It is controlled
by a personal computer.
and the control registers can be read and written by the host system. It takes about
6 seconds to set all the network parameters from the host system.
The exterior of this system is shown in Fig.2(b). Inside the cabinet, there are four
shelves. In each shelf, fourteen circuit boards were mounted and on each board 20
chips were mounted. One chip was used for 18 neurons and the other chips were
used to extend the dendrites. Each neuron is driven by an individual 20MHz clock
oscillator.
4
SYSTEM PERFORMANCE
In order to measure the performance of this system, one neuron was used as a signal
generator. By setting all the synaptic weights and the internal feedback gain of a
signal neuron to zero, and by setting the content of the up-down counter to a given
value, it can produce an output with a constant frequency that is proportional to
the counter value. The input strength of the other neurons can be adjusted by
changing the counter value of a signal neuron or the synaptic weights from it.
The step reponses of a neuron to different inputs are shown in Fig.3(a). As seen in
the figure, the responses exactly followed Eq.(l) and the time constant Was about
400psec. Figure 3(b) shows responses with different time constants. The inputs
were identical for all cases.
Figure 3(c) shows the response of a temporal filter that was obtained by the difference between a fast and a slow neuron. By combining two low-pass filters that had
different cutoff frequencies, a band-pass filter was created. A variety of spatiotem-
Y. Hirai
710
r: rc-------::
_ - - - - - --
'&.
::J
768
500
512
256
OL--~--~--~--~--~--~
o
s;.-~ /~.~-------
1792
...
,I
~ 1500
...?.
" 32. /
S
16
i.?
1
1000
/.'/:/.
1 /
I'
8
/~-
_____
/'
-----
"-- _.---. --
___ ----
- ---
~ ~ ~~-:::-::~~-----'--o
10000 20000 30000 40000 50000
10000
20000
30000
40000
50000
line (x 100 Il8IlO58conds)
time (x 100 nanoeeconds)
(b)
(a)
1~~-~~-~--~---~--.
2000
Bela_53
---::=--------:j
800
1&600
i 1500
S
!
1000
1
400
i ~P--"""IiiiiiiiOi;~
1000
%?200
500
-400
OIL....--~---'---~......::::.==--..-J
o
10000 20000 30000 40000 50000
:L_'---_'---_~~~
lime (x 100 nanoeeconds)
(c)
o
20000
40000
60000
time (x 100 nanoseconds)
80000
(d)
Figure 3: Responses obtained by the system. (a) Step responses to different input
levels. Parameters are the values that are set in the up-down counter of a signal
neuron. (b) Step responses for different time constants. Parameters are the values of
f3i in Eq.5. Inputs were identical in all cases. (c) Response of a temporal filter that
was ol>taind by the difference between a fast and a slow neuron. (d) Response of
a winner-take-all network among 1,007 neurons. The responses of a winner neuron
and 24 of the 1,006 defeated neurons are shown.
poral filters can be implemented in this way.
Figure 3(d) shows the responses of a winner-take-all network among 1,007 neurons.
The time courses of the responses of a winner neuron and 24 of the 1,006 defeated
neurons are shown in the figure. The strength of all of the inhibitory synaptic
weights between neurons was set to 2 x (- ::), where 2 is an output scale factor. The
synaptic weights from a signal neuron to the 1,007 competing ones were identical and
were ~;. Although the network parameters and the inputs to all competing neurons
were identical, one of them won in 6 msec. Since the system operates asynchronously
and the spatial summation of the synaptic output pulses is probabilistic, one of the
competing neurons can win in a stochastic manner.
In order to derive the processing speed in terms of connections per second, the
same winner-take-all network was solved by the Euler method on a latest workstation. Since it took about 76.2 seconds and 2,736 iterations to converge, the
processing speed of the workstation was about 36 million connections per second
l007xI0<17x2736)
.
( ::::::!
76 .2 !
? S'mce t h'1S system .IS 10000'
,
times f aster t han t h e wor kstatlOn,
A I,OOO-Neuron System with One MillioK7-bit Physical Interconnections
711
the processing speed amounts to 360 billion connections per second.
Various kinds of neural networks including spatiotemporal filtering, feedforward and
feedback neural networks can be run in this single system by loading appropriate
network parameters from the host system. The second version of this system, which
can be used via the Internet, will be completed by the end of March, 1998.
Acknowledgements
The author is grateful to Mr. Y. Kuwabara and Mr. T. Ochiai of Hitachi Microcomputer System Ltd. for their collaboration in developing this system and to Dr.
M. Yasunaga and Mr. M. Takahashi for their help in testing it. The author is also
grateful to Mr. H. Toda for his collaboration in measuring response data. This work
was supported by "Proposal-Based Advanced Industrial Technology R&D Program"
from NEDO in Japan.
References
[1] C. Mead: Analog VLSI and Neural Systems. Addison-Wesley Publishing Company, Massachusetts, 1989
[2] K.W.Przytula and V.K.Prasanna, Eds.: Parallel Digital Implementations of
Neural Networks. Prentice Hall, New Jersey, 1993
[3] J.J. Hopfield: Neurons with graded response have collective computational
properties like those of two-state neurons. Proc. Natl. Acad. Sci. U.S.A., 81,
pp.3088-3092, 1984
[4] P. Mueller, J. van der Spiegel, V. Agami, D. Blackman, P. Chance, C. Donham,
R. Etienne, J. Kim. M. Massa and S. Samarasekera: Design and performance
of a prototype analog neural computer. Proc. the 2nd International Conf. on
Microelectronics for Neural Networks, pp.347-357, 1991
[5] G. Cauwenberghs: A learning analog neural network chip with continuous-time
recurrent dynamics. In J. D. Cowan, G. Tesauro and J. Alspector, Eds., Advances in Neural Information Processing Systems 6, Morgan Kaufmann Publishers, San Mateo, CA, pp.858-865, 1994
[6] S. Churcher, D. J. Baxter, A. Hamilton, A. F. Murry, and H. M. Reekie:
Generic analog neural computation - The EPSILON chip. In S. J. Hanson, J.
D. Cowan and C. L. Giles, Eds., Advances in Neural Information Processing
Systems 6, Morgan Kaufmann Publishers, San Mateo, CA, pp.773-780, 1993
[7] H. Eguchi, T. Furuta, H. Horiguchi, S. Oteki and T. Kitaguchi: Neural network
LSI chip with on-chip learning. Proceedings of IJCNN'91 Seattle, Vol.I/453-456,
1991
[8] Y. Hirai, et al.: A digital neuro-chip with unlimited connectability for large
scale neural networks. Proc. International Joint Conf. on Neural Networks'89
Washington D.C., Vo1.11/163-169, 1989
[9] Y.Hirai, VLSI Neural Network Systems (Gordon and Breach Science Publishers, Birkshire, 1992)
[10] Y. Hirai and M. Yasunaga: A PDM digital neural network system with 1,000
neurons fully interconnected via 1,000,000 6-bit synapses. Proc. International
Conference on Neural Information Processings'96, Vo1.ll/1251, 1996
| 1395 |@word version:1 loading:2 nd:1 open:1 donham:1 pulse:21 paid:1 initial:1 configuration:1 contains:1 electronics:1 analysed:1 written:1 realize:2 interrupted:1 msb:1 rc:1 along:1 differential:2 become:1 consists:3 inside:1 manner:1 alspector:1 behavior:2 integrator:1 terminal:2 ol:2 company:1 little:1 increasing:1 becomes:2 project:1 circuit:38 kind:1 developed:2 microcomputer:1 transformation:1 fabricated:1 temporal:2 every:1 exactly:1 control:1 hamilton:1 positive:4 acad:1 mead:1 path:2 mateo:2 limited:1 range:1 faithful:1 testing:1 hyperbolic:1 cannot:1 prentice:1 przytula:1 cmcuit:1 iri:1 attention:1 latest:1 focused:1 array:2 imax:1 his:1 us:1 inserted:1 solved:1 thousand:5 counter:13 aster:1 pdm:4 dynamic:1 personal:2 grateful:2 resolved:1 joint:1 hopfield:1 chip:22 represented:1 various:1 jersey:1 train:1 fast:2 furuta:1 whose:4 solve:2 relax:2 interconnection:6 otherwise:1 spiegel:1 asynchronously:1 took:1 interconnected:4 wijyj:2 combining:1 billion:2 seattle:1 transmission:1 extending:1 produce:2 wired:1 cmos:1 help:1 illustrate:1 derive:1 ac:1 recurrent:1 measured:1 odd:1 eq:6 solves:1 implemented:1 synchronized:1 filter:5 stochastic:1 hitachi:1 feeding:2 summation:3 designate:2 adjusted:2 extension:1 bypassing:1 hall:1 kuwabara:1 proc:4 modulating:2 create:1 shelf:2 blackman:1 industrial:1 kim:1 mueller:1 vlsi:6 wij:1 among:4 spatial:1 summed:1 integration:2 wor:1 f3:3 washington:1 sampling:2 identical:5 extention:1 broad:1 poral:1 inevitable:1 pta:2 others:3 gordon:1 eighty:1 individual:2 maintain:1 saturated:1 udr:1 natl:1 dyi:1 integral:1 increased:1 giles:1 tp:1 mhz:1 measuring:1 euler:1 reported:1 connect:1 varies:1 spatiotemporal:3 accomplish:1 density:2 international:3 probabilistic:1 dr:1 external:3 conf:2 li:2 japan:2 takahashi:1 potential:2 register:1 stream:2 performed:1 cauwenberghs:1 parallel:3 jul:2 square:1 kaufmann:2 characteristic:2 massa:1 cc:1 simultaneous:1 synapsis:12 synaptic:12 ed:3 sixth:1 frequency:10 pp:4 cabinet:1 workstation:2 gain:1 massachusetts:1 ut:1 wesley:1 hirai:8 response:13 improved:1 synapse:8 ooo:5 clock:4 nonlinear:2 mode:1 oil:1 multiplier:13 counterpart:1 read:1 laboratory:1 illustrated:2 wiring:2 ll:1 noted:1 won:2 outline:1 enlargement:1 performs:1 instantaneous:2 nanosecond:1 blt:2 physical:5 fourteen:1 winner:6 jp:1 million:6 analog:10 extend:3 pm:1 had:1 han:1 driven:3 tesauro:1 binary:1 yi:5 der:1 transmitted:2 seen:1 morgan:2 mr:4 churcher:1 converge:1 signal:9 ii:2 full:1 isy:1 defeated:2 host:5 controlled:2 neuro:1 physically:2 iteration:1 cell:8 receive:1 addition:1 proposal:1 decreased:1 diagram:4 lsb:1 desynchronized:1 crucial:1 publisher:3 tend:1 cowan:2 emitted:1 feedforward:5 baxter:1 variety:1 ibaraki:1 competing:4 f3i:3 prototype:1 br:1 pse:1 ltd:1 amount:2 transforms:1 mid:1 band:1 ten:1 hardware:1 gnd:1 specifies:1 lsi:1 millisecond:1 inhibitory:5 sign:3 per:5 vol:1 four:1 threshold:1 changing:1 cutoff:1 yasunaga:2 run:3 inverse:1 package:1 fourth:1 place:1 scaling:2 lime:1 bit:17 internet:1 followed:1 tsukuba:3 strength:4 ijcnn:1 constraint:2 ri:1 flat:1 unlimited:1 aspect:1 speed:5 integrand:1 extremely:1 developing:1 according:2 march:1 describes:1 equation:2 bus:1 pin:2 addison:1 fed:3 end:1 operation:2 appropriate:3 generic:1 ochiai:1 gate:6 completed:1 publishing:1 etienne:1 epsilon:1 especially:1 toda:1 graded:1 realized:1 occurs:1 win:1 sci:1 transit:1 mail:1 negative:3 implementation:4 design:1 collective:1 twenty:1 neuron:57 iti:2 psec:2 pair:1 connection:4 hanson:1 conflict:1 dynamical:1 below:2 appeared:1 saturation:1 program:1 including:3 max:2 ia:1 advanced:1 technology:1 created:1 carried:1 breach:1 text:1 literature:1 acknowledgement:1 tangent:1 val:1 multiplication:1 fully:5 limitation:1 filtering:3 proportional:4 mounted:3 generator:1 digital:8 reekie:1 prasanna:1 pi:1 collaboration:2 excitatory:4 course:1 supported:1 asynchronous:3 l_:1 side:1 viet:2 institute:1 absolute:1 van:1 regard:1 feedback:8 overcome:1 author:3 made:1 san:2 counted:1 continuous:2 table:1 ca:2 exterior:2 dendrite:12 main:2 arrow:1 body:9 fig:9 board:2 fashion:1 slow:2 msec:2 down:10 microelectronics:1 ci:1 magnitude:3 mce:1 lt:1 expressed:1 chance:1 itl:1 oscillator:2 content:2 change:2 hard:1 operates:2 wt:1 vo1:2 nil:1 inj:1 total:3 pas:2 exception:1 eguchi:1 internal:5 brevity:1 |
434 | 1,396 | Multiple Threshold Neural Logic
Jehoshua Bruck
Vasken Bohossian
~mail:
California Institute of Technology
Mail Code 136-93
Pasadena, CA 91125
{vincent, bruck}~paradise.caltech.edu
Abstract
We introduce a new Boolean computing element related to the Linear Threshold element, which is the Boolean version of the neuron.
Instead of the sign function, it computes an arbitrary (with polynornialy many transitions) Boolean function of the weighted sum of
its inputs. We call the new computing element an LT M element,
which stands for Linear Threshold with Multiple transitions.
The paper consists of the following main contributions related to
our study of LTM circuits: (i) the creation of efficient designs of
LTM circuits for the addition of a multiple number of integers and
the product of two integers. In particular, we show how to compute
the addition of m integers with a single layer of LT M elements.
(ii) a proof that the area of the VLSI layout is reduced from O(n 2 )
in LT circuits to O(n) in LTM circuits, for n inputs symmetric
Boolean functions, and (iii) the characterization of the computing
power of LTM relative to LT circuits.
1
Introduction
Human brains are by far superior to computers in solving hard problems like combinatorial optimization and image and speech recognition, although their basic building blocks are several orders of magnitude slower. This observation has boosted
interest in the field of artificial neural networks [Hopfield 82], [Rumelhart 82]. The
latter are built by interconnecting artificial neurons whose behavior is inspired by
that of biological neurons. In this paper we consider the Boolean version of an artificial neuron, namely, a Linear Threshold (LT) element, which computes a neural-like
MUltiple Threshold Neural Logic
0
WI
-Wo
Wn
253
1
0
LT gate
0
t3
1
t2
WI
t3
1
t2
t1
Wn
t1
0
1
SYM gate
0
1
LTM gate
Figure 1: Schematic representation of LT, SYM and LTM computing elements.
Boolean function of n binary inputs [Muroga 71]. An LT element outputs the sign
of a weighted sum of its Boolean inputs. The main issues in the study of networks
(circuits) consisting of LT elements, called LT circuits, include the estimation of
their computational capabilities and limitations and the comparison of their properties with those of traditional Boolean logic circuits based on AND, OR and NOT
gates (called AON circuits). For example, there is a strong evidence that LT circuits are more efficient than AON circuits in implementing a number of important
functions including the addition, product and division of integers [Siu 94], [Siu 93].
Motivated by our recent work on the VLSI implementation of LT elements
[Bohossian 95b], we introduce in this paper a more powerful computing element,
a multiple threshold neuron, which we call LTM, which stands for Linear Threshold with Multiple transitions, see [Haring 66] and [Olafsson 88]. Instead of the
sign function in the LT element it computes an arbitrary (with polynomialy many
transitions) Boolean function of the weighted sum of its inputs.
The main issues in the study of LTM circuits (circuits consisting of LTM elements)
include the estimation of their computational capabilities and limitations and the
comparison of their properties to those of AON circuits. A natural approach in this
study is first to understand the relation between LT circuits and LT M circuits. Our
main contributions in this paper are:
? We demonstrate the power of LTM by deriving efficient designs of LTM
circuits for the addition of m integers and the product of two integers.
? We show that LT M circuits are more amenable in implementation than LT
circuits. In particular, the area of the VLSI layout is reduced from O(n 2 )
in LT circuits to O(n) in LTM circuits, for n input symmetric Boolean
functions.
? We characterize the computing power of LT M relative to LT circuits.
Next we describe the formal definitions of LT and LT M elements.
1.1
Definitions and Examples
Definition 1 (Linear Threshold Gate - LT)
A linear threshold gate computes a Boolean function of its binary inputs :
n
f(X) = sgn(wo
+ L WiXi)
i=l
V. Bohossian and J. Bruck
254
where the Wi are integers and sgn(.) outputs 1 if its argument is greater or equal to
0, and 0 otherwise.
Figure 1 shows an-input LT element; if L~ WiXi ~ -Wo the element outputs 1,
otherwise it outputs o. A single LT gate is unable to compute parity. The latter
belongs to the general class of symmetric functions - SY M.
Definition 2 (Symmetric Functions - SY M)
A Boolean function f is symmetric if its value depends only on the number of ones
in the input denoted by IX I.
Figure 1 shows an example of a symmetric function; it has three transitions, it
outputs 1 for IXI < tl and for t2 ~ IXI < t3, and 0 otherwise. AND, OR and
parity are examples of symmetric functions. A single LT element can implement
only a limited subset of symmetric functions. We define LT M as a generalization
of SY M. That is, we allow the weights to be arbitrary as in the case of LT, rather
than fixed to 1 (see Figure 1 ).
Definition 3 (Linear Threshold Gate with Multiple Transitions - LT M)
A function f is in LTM if there exists a set of weights Wi E Z, 1 ~ i ~ n and a
function h : Z ---+ {O, 1} such that
n
f(X) = h(L wixd for all X
E
{O,l}n
i=l
The only constraint on h is that it undergoes polynomialy many transitions as its
input scans [- L~=l IWi I, L~=l IWi I]?
Notice that without the constraint on the number of transitions, an LTM gate is
capable of computing any Boolean function. Indeed, given an arbitrary function f,
let Wi = 2i - 1 and h(L~ 2i - 1 xd = f(xt, .? ?, x n ).
Example 1 (XOR E LTM)
XOR(X) outputs 1 if lXI, the number of l's in X, is odd. Otherwise it outputs
O. To implement it choose Wi = 1 and h(k) = ~(1 - (_l)k) for 0 ~ k ~ n. Note
that h(k) needs not be defined for k < 0 and k > n, and has polynomialy many
transitions.
Another useful function that LTM can compute is ADD (X, Y), the sum of two
n-bit integers X and Y.
Example 2 (ADD E LT M)
To implement addition we set fl (X, Y) = h, (L~=l 2i (Xi + yd) where h, (k) = 1 for
k E [2 ' ,2 x 2' - 1] U [3 X 2 ' , +00). Defined thus, fl computes the m-th bit of X + Y.
1.2
Organization
The paper is organized as follows . In Section 2, we study a number of applications
as well as the VLSI implementations of LTM circuits. In particular, we show how
to compute the addition of m integers with a single layer of LT!vI elements. In
Section 3, we prove J..he characterization results of LT M - inclusion relations, in
particular LTM ~ LT2. In addition, we indicate which inclusions are proper and
exhibit functions to demonstrate the separations.
255
Multiple Threshold Neural Logic
2
LT M Constructions
The theoretical results about LTM can be applied to the VLSI implementation of
Boolean functions. The idea of a gate with multiple thresholds came to us as we
were looking for an efficient VLSI implementation of symmetric Boolean functions.
Even though a single LT gate is not powerful enough to implement any symmetric
function, a 2-layer LT circuit is. FUrthermore, it is well known that such a circuit
performs much better than the traditional logic circuit based on AND, OR and
NOT gates. The latter has exponential size (or unbounded depth) [Wegener 91].
Proposition 4 (LT2 versus LT M for symmetric function implementation)
The LT2 layout of a symmetric function requires area of O(n 2), while using LT M
one needs only area of O( n).
PROOF:
Implementing a generalized symmetric function in LT2 requires up to n LT gates in
the first layer. Those have the same weights Wi except for the threshold Woo Instead
of laying out n times the same linear sum E~ WiXi we do it once and compare the
result to n different thresholds. The resulting circuit corresponds to a single LTM
gate.
0
The LT2 layout is redundant, it has n copies of each weight, requiring area of at
least O(n 2). On the other hand, LTM performs a single weighted sum, its area
requirement is O(n).
A single LT M gate can compute the addition of m n-bit integers M ADD. The
only constraint is that m be polynomial in n.
Theorem 5 (MADD E LTM)
A single layer of LT M gates can compute the sum of m n-bit integers, provided
that m is at most polynomial in n.
PROOF:
MAD D returns an integer of at most n + log m bits. We need one LT M gate per
bit. The least significant bit is computed by a simple m-bit XOR. For all other
bits we use h(X(l), .. ,x(m?) = hl(E~=12i Ej=l x~j?) to compute the l-th bit ofthe
mm.
0
Corollary 6 (PRODUCT E PTM) A single layer of PTM (which is defined
below) gates, can compute the product of m n-bit integers, provided that m is at
most polynomial in n.
PROOF:
By analogy with PTb defined in [Bruck 90], in PT Ml (or simply PT M) we allow a
polynomial rather than a linear sum: f(X) = h(WIXl + ... +wnxn +W(1,2)XIX2+ ... )
However we restrict the sum to have polynomialy many terms (else, any Boolean
function could be realized with a single gate). The product of two n-bit integers
X and Y can be written as PRODUCT(X, Y) = E~=l XiY. We use the construction of MADD in order to implement PRODUCT. PRODUCT(X, Y) =
MADD(x 1Y,x2 Y , ... ,xnY). fleX, Y) = hi (LJj=l LJi=12i XjYi) b outputs the l-th
bit of the product.
0
"n "I
V. Bohossian and J Bruck
256
Figure 2: Relationship between Classes
3
Classification of LTM
.-
---
We me a hat to indicate small (polynomialy growing) weights, e.g. LT, LT M
[Bohossian 95a], [Siu 91], and a subscript to indicate the depth (number of layers)
of the circuit of more than a single layer. All the circuits we consider in this paper
are of polynomial size (number of elements) in n (number of inputs). For example,
the class fr2 consists of those B0...2!ean functions that can be implemented by a
depth-2 polynomial size circuit of LT elements.
Figure 2 depicts the membership relations between five classes of Boolean functions,
including, LT, ilr, LTM, LTM and ilr2, along with the functions used to establish
the separations.
In this section we will prove the relations illustrated by Figure 2 .
Theorem 7 (Classification of LTM )
The inclusions and separations shown in Figure 12 hold. That is,
.-
1. LT
~
LT ; LTM
12. LT
~
LTM ; LTM
---
.-
9. LTM; LT2
4? XOR
E
CTM
but XOR tJ. LT
5. CaMP E LT but CaMP tJ. LTM
6. ADD E LTM but ADD tJ. LTULTM
7. IPk E
fr2
but 1Pk tJ. LTM
PROOF:
We show only the outline of the proof. The complete version can be found in
[Bohossian 96]. Claims 1 and 2 follow from the definition. The first part of Claim 4
was shown in Example 1 and the second is well known. In Claim 5, CaMP stands
for the Comparison functio!lt the proof mes the pigeonhole principle and is related
to the proof of CaMP tJ. LT which can be found in [Siu 91]. In Claim 6 to show
that ADD tJ. LTM we use the same idea as for CaMP. Claim 3 is proved using a
result from [Goldman 93]: a single LT gate with arbitrary weights can be realized
by an LT2 circuit. Claim 7 introduces the function IPk(X, Y) = 1 iff L:~ XiYi ~ k,
257
Multiple Threshold Neural Logic
o otherwise.
.-
If IPk E LTM, using the result from [Goldman 93], we can construct
a LT2 circuit that computes IP2 (Inner Product mod 2) which is known to be false
0
[Hajnal 94].
What remains to be shown in order to complete the classification picture is
LT n LTM. We conjecture that this is true.
4
fr
=
Conclusions
Our original goal was to use theoretical results in order to efficiently layout a
generalized symmetric function. During that process we came to the conclusion
that the LT2 implementation is partially redundant, which lead to the definition
of LTM, a new, more powerful computing element. We characterized the power
of LTM relative to LT. We showed how it can be used to reduce the area of
VLSI layouts from O(n 2 ) to O(n) and derive efficient designs for multiple addition
and product. Interesting directions for future investigation are (i) to prove the
conjecture: fr = LT n LTM, (ii) to apply spectral techniques ([Bruck 90)) to the
analysis of LT M, in particular show how PT M fits into the classification picture
(Figure 2 ).
Another direction for future research consists in introducing the ideas described
above in the domain of VLSI. We have fabricated a programmable generalized
symmetric function on a 2J,L, analog chip using the model described above. Floating
gate technology is used to program the weights. We store a weight on a single
transistor by injecting and tunneling electrons on the floating gate [Hasler 95].
Acknowledgments
This work was supported in part by the NSF Young Investigator Award CCR9457811 and by the Sloan Research Fellowship.
References
[Bohossian 95a] V. Bohossian and J. Bruck. On Neural Networks with Minimal
Weights. In Advances in Neural Information Processing Systems 8, MIT Press,
Cambridge, MA, 1996, pp.246-252.
[Bohossian 95b] V. Bohossian, P. Hasler and J. Bruck. Programmable Neural Logic.
Proceedings of the second annual IEEE International Conference on Innovative
Systems in Silicon, pp. 13-21, October 1997.
[Bohossian 96] V. Bohossian and J. Bruck.
ral Logic.
Technical Report, ETR010,
http://paradise.caltech.edu/ETR.html)
Multiple Threshold N euJune 1996.
(available at
[Bruck 90] J. Bruck. Harmonic Analysis of Polynomial Threshold Functions. SIAM
J. Disc. Math, Vol. 3(No. 2)pp. 168- 177, May 1990.
[Goldman 93] M. Goldmann and M. Karpinski. Simulating threshold circuits by
majority circuits. In Proc. 25th ACM STOC, pages pp. 551-560, 1993.
258
V. Bohossian and J Bruck
[Hajnal 94] A. Hajnal, W. Maass, P. Pudlak, M. Szegedy, G. Turan. Threshold
Circuits of Bounded Depth. Journal of Computer and System Sciences, Vol.
46(No. 2):pp. 129-154, April 1993.
[Haring 66] D.R. Haring. Multi-Threshold Threshold Elements. IEEE Transactions
on Electronic Computers, Vol. EC-15, No.1, February 1966.
[Hasler 95] P. Hasler, C. Diorio, B.A. Minch and C.A. Mead. Single Transistor
Learning Synapses. Advances in Neural Information Processing Systems 7,
MIT Press, Cambridge, MA, 1995, pp.817-824.
[Hastad 94] J. Hastad. On the size of weights for threshold gates. SIAM. J. Disc.
Math., 7:484-492, 1994.
[Hofmeister 96] T. Hofmeister. A Note on the Simulation of Exponential Threshold
Weights. 1996 CONCOON conference.
[Hopfield 82] J. Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proc. of the USA National Academy of Sciences,
79:2554-2558, 1982.
[Muroga 71] M. Muroga. Threshold Logic and its Applications. Wiley-Interscience,
1971.
[Olafsson 88] S. Olafsson and Y.S. Abu-Mostafa. The Capacity of Multilevel
Threshold Functions. IEEE Transactions on Pattern Analysis and Machine
Intelligence, Vol.10, No.2, March 1988.
[Rumelhart 82] D. Rumelhart and J. McClelland. Parallel distributed processing:
Explorations in the microstructure of cognition. MIT Press, 1982.
[Siu 91] K. Siu and J. Bruck. On the power of threshold circuits with small weights.
SIAM J. Disc. Math., Vol. 4(No. 3):pp. 423-435, August 1991.
[Siu 93] K. Siu, J. Bruck, T. Kailath, and T. Hofmeister. Depth Efficient Neural
Networks for Division and Related Problems. IEEE Trans. on Information
Theory, Vol. 39(No. 3), May 1993.
[Siu 94] K. Siu and V.P. Roychowdhury. On Optimal Depth Threshold Circuits for
Multiplication and Related Problems. SIAM J. Disc. Math., Vol. 7(No. 2):pp.
284-292, May 94.
[Szegedy 89] M. Szegedy. Algebraic Methods in Lower Bounds for Computational
Models with Limited Communication. PhD Thesis, Dep. Computer Science,
Chicago Univ., December 1989.
[Wegener 91] 1. Wegener. The complexity of the parity function in unbounded fanin unbounded depth circuits. In Theoretical Computer Science, Vol. 85, pp.
155-170, 1991.
| 1396 |@word version:3 polynomial:7 simulation:1 xiy:1 xiyi:1 written:1 chicago:1 hajnal:3 intelligence:1 ptm:2 characterization:2 math:4 five:1 unbounded:3 along:1 consists:3 prove:3 interscience:1 introduce:2 indeed:1 behavior:1 growing:1 multi:1 brain:1 ptb:1 inspired:1 goldman:3 provided:2 bounded:1 circuit:38 what:1 turan:1 fabricated:1 xd:1 t1:2 mead:1 subscript:1 yd:1 limited:2 acknowledgment:1 flex:1 block:1 implement:5 pudlak:1 area:7 layout:6 deriving:1 xjyi:1 construction:2 pt:3 ixi:2 element:22 rumelhart:3 recognition:1 fr2:2 pigeonhole:1 diorio:1 lji:1 complexity:1 solving:1 creation:1 division:2 hopfield:3 chip:1 emergent:1 univ:1 describe:1 paradise:2 artificial:3 whose:1 otherwise:5 ability:1 haring:3 transistor:2 product:12 fr:2 iff:1 academy:1 requirement:1 derive:1 odd:1 b0:1 dep:1 strong:1 implemented:1 indicate:3 direction:2 exploration:1 human:1 sgn:2 implementing:2 multilevel:1 microstructure:1 generalization:1 investigation:1 proposition:1 biological:1 mm:1 hold:1 cognition:1 claim:6 electron:1 mostafa:1 ctm:1 estimation:2 proc:2 injecting:1 combinatorial:1 weighted:4 mit:3 rather:2 ej:1 boosted:1 corollary:1 ral:1 camp:5 membership:1 pasadena:1 vlsi:8 relation:4 issue:2 classification:4 html:1 denoted:1 field:1 equal:1 once:1 construct:1 muroga:3 future:2 t2:3 report:1 national:1 floating:2 consisting:2 organization:1 interest:1 introduces:1 tj:6 amenable:1 capable:1 theoretical:3 minimal:1 boolean:17 hastad:2 introducing:1 subset:1 siu:10 characterize:1 minch:1 international:1 siam:4 thesis:1 choose:1 return:1 szegedy:3 sloan:1 depends:1 vi:1 hofmeister:3 capability:2 parallel:1 iwi:2 contribution:2 xor:5 efficiently:1 sy:3 t3:3 ofthe:1 vincent:1 disc:4 synapsis:1 definition:7 pp:9 proof:8 proved:1 wixi:3 organized:1 follow:1 april:1 though:1 furthermore:1 hand:1 undergoes:1 usa:1 building:1 requiring:1 true:1 symmetric:15 maass:1 illustrated:1 during:1 generalized:3 outline:1 complete:2 demonstrate:2 performs:2 image:1 harmonic:1 superior:1 physical:1 analog:1 he:1 significant:1 silicon:1 cambridge:2 inclusion:3 add:6 recent:1 showed:1 belongs:1 store:1 binary:2 came:2 caltech:2 greater:1 bohossian:13 redundant:2 ii:2 multiple:12 technical:1 characterized:1 award:1 functio:1 schematic:1 basic:1 karpinski:1 addition:9 fellowship:1 else:1 december:1 mod:1 call:2 integer:14 wegener:3 iii:1 enough:1 wn:2 fit:1 restrict:1 inner:1 idea:3 reduce:1 motivated:1 wo:3 aon:3 algebraic:1 speech:1 programmable:2 useful:1 mcclelland:1 reduced:2 http:1 nsf:1 notice:1 roychowdhury:1 sign:3 per:1 vol:8 abu:1 threshold:27 hasler:4 sum:9 powerful:3 electronic:1 separation:3 tunneling:1 bit:13 layer:8 fl:2 hi:1 bound:1 annual:1 constraint:3 x2:1 argument:1 innovative:1 conjecture:2 jehoshua:1 ip2:1 march:1 wi:7 hl:1 lt2:9 remains:1 madd:3 available:1 goldmann:1 apply:1 spectral:1 simulating:1 lxi:1 slower:1 gate:23 hat:1 original:1 include:2 establish:1 february:1 realized:2 traditional:2 exhibit:1 unable:1 capacity:1 majority:1 me:2 mail:2 etr:1 mad:1 laying:1 code:1 relationship:1 october:1 ilr:1 stoc:1 design:3 implementation:7 proper:1 collective:1 neuron:5 observation:1 looking:1 communication:1 arbitrary:5 august:1 namely:1 california:1 trans:1 below:1 pattern:1 ipk:3 program:1 built:1 including:2 power:5 natural:1 bruck:14 technology:2 picture:2 woo:1 ljj:1 multiplication:1 relative:3 interesting:1 limitation:2 analogy:1 versus:1 principle:1 supported:1 parity:3 copy:1 sym:2 formal:1 allow:2 understand:1 institute:1 distributed:1 depth:7 transition:9 stand:3 computes:6 olafsson:3 far:1 ec:1 transaction:2 logic:9 ml:1 xi:1 ca:1 ean:1 domain:1 pk:1 main:4 tl:1 depicts:1 wiley:1 interconnecting:1 exponential:2 ix:1 young:1 theorem:2 xt:1 evidence:1 polynomialy:5 exists:1 false:1 ltm:39 phd:1 magnitude:1 fanin:1 lt:57 simply:1 partially:1 corresponds:1 acm:1 xix2:1 ma:2 goal:1 kailath:1 hard:1 except:1 called:2 latter:3 scan:1 investigator:1 |
435 | 1,397 | Function Approximat.ion with the
Sweeping Hinge Algorithm
Don R. Hush, Fernando Lozano
Dept. of Elec. and Compo Engg.
University of New Mexico
Albuquerque, NM 87131
Bill Horne
MakeWaves, Inc.
832 Valley Road
Watchung, NJ 07060
Abstract
We present a computationally efficient algorithm for function approximation with piecewise linear sigmoidal nodes. A one hidden
layer network is constructed one node at a time using the method of
fitting the residual. The task of fitting individual nodes is accomplished using a new algorithm that searchs for the best fit by solving
a sequence of Quadratic Programming problems. This approach offers significant advantages over derivative-based search algorithms
(e.g. backpropagation and its extensions). Unique characteristics
of this algorithm include: finite step convergence, a simple stopping criterion, a deterministic methodology for seeking "good" local
minima, good scaling properties and a robust numerical implementation.
1
Introduction
The learning algorithm developed in this paper is quite different from the traditional family of derivative-based descent methods used to train Multilayer Perceptrons (MLPs) for function approximation. First, a constructive approach is used,
which builds the network one node at a time. Second, and more importantly, we
use piecewise linear sigmoidal nodes instead of the more popular (continuously differentiable) logistic nodes. These two differences change the nature of the learning
problem entirely. It becomes a combinatorial problem in the sense that the number
of feasible solutions that must be considered in the search is finite. We show that
this number is exponential in the input dimension, and that the problem of finding the global optimum admits no polynomial-time solution. We then proceed to
develop a heuristic algorithm that produces good approximations with reasonable
efficiency. This algorithm has a simple stopping criterion, and very few user specified parameters. In addition, it produces solutions that are comparable to (and
sometimes better than) those produced by local descent methods, and it does so
D. R. Hush, R Lozano and B. Horne
536
using a deterministic methodology, so that the results are independent of initial
conditions.
2
Background and Motivation
We wish to approximate an unknown continuous function f(x) over a compact set
with a one-hidden layer network described by
n
f~(x)
= ao + L
(1)
aiU(x, Wi)
i=l
where n is the number of hidden layer nodes (basis functions), x E ~d is the input
vector, and {u(x, w)} are sigmoidal functions parameterized by a weight vector w.
A set of example data, S = {Xi, Yi}, with a total of N samples is available for
training and test.
The models in (1) have been shown to be universal approximators. More importantly, (Barron, 1993) has shown that for a special class of continuous functions,
r c, the generalization error satisfies
~
E[lIf - fn,NII2]
IIf -
fnll 2 + E[lIfn - fn,NII2]
= 0 (*)
+0
( nd ~g N
)
where 11?11 is the appropriate two-norm, f n is the the best n-node approximation to
f, and fn,N is the approximation that best fits the samples in S. In this equation
IIf - fnll 2 and E[lIfn - fn,NII2] correspond to the approximation and estimation error
respectively. Of particular interest is the O(l/n) bound on approximation error,
which for fixed basis functions is of the form O(1/n 2 / d ) (Barron, 1993). Barron's
result tells us that the (tunable) sigmoidal bases are able to avoid the curse of
dimensionality (for functions in rc). Further, it has been shown that the O(l/n)
bound can be achieved constructively (Jones, 1992), that is by designing the basis
functions (nodes) one at a time. The proof of this result is itself constructive,
and thus provides a framework for the development of an algorithm which can (in
principle) achieve this bound. One manifestation of this algorithm is shown in
Figure 1. We call this the iterative approximation algorithm (I1A) because it builds
the approximation by iterating on the residual (Le. the unexplained portion of the
function) at each step. This is the same algorithmic strategy used to form bases in
numerous other settings, e.g. Grahm-Schmidt, Conjugate Gradient, and Projection
Pursuit. The difficult part of the I1A algorithm is in the determination of the best
fitting basis function Un in step 2. This is the focus of the remainder of this paper.
3
Algorithmic Development
We begin by defining the hinging sigmoid (HS) node on which our algorithms are
based. An HS node performs the function
w+,
Uh(X,
w) =
{
- Tw,
X,
w_,
-T>
w,
X _ w+
- Tw_ ~ w,
X ~ w+
T<
w, x _ w_
(2)
where w T = [WI w+ w_] and x is an augmented input vector with a 1 in the first
component. An example of the surface formed by an HS node on a two-dimensional
input is shown in Figure 2. It is comprised of three hyperplanes joined pairwise
Function Approximation with the Sweeping Hinge Algorithm
537
Initialization: fo(x) = 0
for n
1 to nma:c do
1. Compute Residual:
2. Fit Residual:
=
en(x) = f(x) - fn-l (x)
un(x) = argminO"EE lIen(x) - u(x)11
3. Update Estimate:
fn(x) = o:fn-l (x) + f3u n (x)
where 0: and f3 are chosen to minimize IIf(x) - fn(x)1I
endloop
Figure 1: Iterative Approximation Algorithm (rIA).
Figure 2: A Sigmoid Hinge function in two dimensions .
continuously at two hinge locations. The upper and middle hyperplanes are joined
at "Hinge I" and the lower and middle hyperplanes are joined at "Hinge 2". These
hinges induce linear partitions on the input space that divide the space into three
regions, and the samples in 5 into three subsets,
5+ = {(Xi,Yi): Wr-Xi ~ w+}
5, = {(Xi,Yi): w_ ~ WTXi ~ w+}
5_ = {(Xi,Yi): WTXi ~ w_}
(3)
These subsets, and the corresponding regions of the input space, are referred to as
the PLUS, LINEAR and MINUS subsets/regions respectively. We refer to this type
of partition as a sigmoidal partition. A sigmoidal partition of 5 will be denoted
P = {5+, 5" 5_}, and the set of all such partitions will be denoted II = {Pd.
Input samples that fall on the boundary between two regions can be assigned to the
set on either side. These points are referred to as hinge samples and playa crucial
role in subsequent development. Note that once a weight vector w is specified, the
partition P is completely determined, but the reverse is not necessarily true. That
is, there are generally an infinite number of weight vectors that induce the same
partition.
We begin our quest for a learning algorithm with the development of an expression
for the empirical risk. The empirical risk (squared error over the sample set) is
defined
(4)
D. R Hush, F. Lozano and B. Horne
538
This expression can be expanded into three terms, one for each set in the partition,
Ep(w) =
~ :E(Yi -
W_)2
~
+ ~ :E(Yi -
W+)2
+ ~ 2)Yi -
~
WT Xi)2
~
After further expansion and rearrangement of terms we obtain
1
Ep(w) = 2wTRw - w T r
where
+
s;
"L:s, XiX; r, = "L: s, XiYi
s; = ! "L:s Y; st = "L:s+ Yi Sy = "L:s_ Yi
R, =
R=
(
R,
~
r=un
(5)
(6)
(7)
(8)
and N+ , N, and N_ are the number of samples in S+ , S, and S_ respectively. The
subscript P is used to emphasize that this criterion is dependent on the partition (i.e.
P is required to form Rand r). In fact, the nature of the partition plays a critical
role in determining the properties of the solution. When R is positive definite (i.e.
full rank), P is referred to as a stable partition, and when R has reduced rank P is
referred to as an unstable partition. A stable partition requires that R, > O. For
purposes of algorithm development we will assume that R, > 0 when ISti > Nmin,
where Nmin is a suitably chosen value greater than or equal to d + 1. With this, a
necessary condition for a stable partition is that there be at least one sample in S+
and S_ and N, ~ Nmin. When seeking a minimizing solution for Ep(w) we restrict
ourselves to stable partitions because of the potential nonuniqueness associated with
solutions to unstable partitions.
Determining a weight vector that simultaneously minimizes E p (w) and preserves
the current partition can be posed as a constrained optimization problem. This
problem takes on the form
- w Tr
min !wTRw
2
subject to Aw ~ 0
(9)
where the inequality constraints are designed to maintain the current partition defined by (3). This is a Quadratic Programming problem with inequality constraints,
and because R > 0 it has a unique global minimum. The general Quadratic Programming problem is N P-hard and also hard to approximate (Bellare and Rogaway,
1993). However, the convex case which we restrict ourselves to here (i.e. R > 0)
admits a polynomial time solution. In this paper we use the active set algorithm
(Luenberger, 1984) to solve (9). With the proper implementation, this algorithm
runs in O(k(~ + Nd)) time, where k is typically on the order of d or less.
The solution to the quadratic programming problem in (9) is only as good as the
current partition allows. The more challenging aspect of minimizing Ep(w) is in
the search for a good partition. Unfortunately there is no ordering or arrangement
of partitions that is convex in Ep(w), so the search for the optimal partition will
be a computationally challenging problem. An exhaustive search is usually out of
the question because of the prohibitively large number of partitions, as given by the
following lemma.
Lemma 1: Let S contain a total of N samples in Rd that lie in general position.
Then the number of sigmoidal partitions defined in (3) is 8(Nd+l).
Function Approximation with the Sweeping Hinge Algorithm
539
Proof: A detailed proof is beyond the scope of this paper, but an intuitive proof
follows. It is well-known that the number of linear dichotomies of N points in d
dimensions is 8(N d ) (Edelsbrunner, 1987). Each sigmoidal partition is comprised
of two linear dichotomies, one formed by Hinge 1 and the other by Hinge 2, and
these dichotomies are constrained to be simple translations of one another. Thus,
to enumerate all sigmoidal partitions we allow one of the hinges, say Hinge 1, can
take on 8(Nd) different positions. For each of these the other hinge can occupy
only'" N unique positions. The total is therefore 8 (Nd+l ).
The search algorithm developed here employs a Quadratic Programming (QP) algorithm at each new partition to determine the optimal weight vector for that
partition (Le. the optimal orientation for the separating hyperplanes). Transitions
are made from one partition to the next by allowing hinge samples to flip from one
side of the hinge boundary to the next. The search is terminated when a minimum
value of Ep(w) is found (Le. it can no longer be reduced by flipping hinge samples).
Such an algorithm is shown in Figure 3. We call this the HingeDescent algorithm
because it allows the hinges to "walk across" the data in a manner that descends
the Ep(w) criterion. Note that provisions are made within the algorithm to avoid
unstable partitions. Note also that it is easy to modify this algorithm to descend
only one hinge at a time, simply by omitting one of the blocks of code that flips
samples across the corresponding hinge boundary.
{This routine is invoked with a stable feasible solution W = {w, R, r, A, S+, SI, S_ }.}
procedure HingeDescent (W)
{ Allow hinges to walk across the data until a minimizing partition is found. }
E_1wTRw-wTr
- 2
do
Emin = E
{Flip Hinge 1 Samples.}
for each ?Xi, Yi) on Hinge 1) do
if ?Xi, Yi) E S+ and N+ > 1) then
Move (Xi,Yi) from S+ to S" and update R, r, and A
elseif ?Xi, Yi) E S, and N, > N min ) then
Move (Xi, Yi) from S, to S+, and update R, r, and A
endif
endloop
{Flip Hinge 2 Samples.}
for each ?Xi, Yi) on Hinge 2) do
if ?Xi,Yi) E S- and N_ > 1) then
Move (Xi,Yi) from S- to S" and update R, r, and A
elseif ?Xi, Yi) E S, and N, > Nmin) then
Move (Xi,Yi) from S, to S-, and update R, r, and A
endif
endloop
{Compute optimal solution for new partition.}
W = QPSolve(W};
E= ~wTRw-wTr
while (E < Emin) j
return(W)j
end;
{HingeDescent}
Figure 3: The HingeDescent Algorithm.
Lemma 2: When started at a stable partition, the HingeDescent algorithm will
D. R Hush, R Lozano and B. Horne
540
converge to a stable partition of Ep(w) in a finite number of steps.
Proof: First note that when R> 0, a QP solution can always be found in a finite
number of steps. The proof of this result is beyond the scope of this paper, but can
easily be found in the literature (Luenberger, 1984). Now, by design, HingeDescent
always moves from one stable partition to the next, maintaining the R > 0 property
at each step so that all QP solutions can be produced in a finite number of steps.
In addition, Ep(w) is reduced at each step (except the last one) so no partitions
are revisited, and since there are a finite number of partitions (see Lemma 1) this
algorithm must terminate in a finite number of steps. QED.
Assume that QPSol ve runs in O(k( cP +N d)) time as previously stated. Then the run
time of HingeDescent is given by O(Np((k+Nh)cP+kNd)), where Nh is the number
of samples flipped at each step and Np is the total number of partitions explored.
Typical values for k and Nh are on the order of d, simplifying this expression to
O(Np(d 3 + NcP)). Np can vary widely, but is often substantially less than N.
HingeDescent seeks a local minimum over II, and may produce a poor solution,
depending on the starting partition. One way to remedy this is to start from
several different initial partitions, and then retain the best solution overall. We
take a different approach here, that always starts with the same initial condition,
visits several local minima along the way, and always ends up with the same final
solution each time.
The SweepingHinge algorithm works as follows. It starts by placing one of the
hinges, say Hinge 1, at the outer boundary of the data. It then sweeps this hinge
across the data, M samples at a time (e.g. M = 1), allowing the other hinge (Hinge
2) to descend to an optimal position at each step. The initial hinge locations are
determined as follows. A linear fit is formed to the entire data set and the hinges are
positioned at opposite ends of the data so that the PLUS and MINUS regions meet
the LINEAR region at the two data samples on either end. After the initial linear
fit, the hinges are allowed to descend to a local minimum using HingeDescent. Then
Hinge 1 is swept across the data M samples at a time. Mechanically this is achieved
by moving M additional samples from S, to S+ at each step. Hinge 2 is allowed
to descend to an optimal position at each of these steps using the Hinge2Descent
algorithm. This algorithm is identical to HingeDescent except that the code that
flips samples across Hinge 1 is omitted. The best overall solution from the sweep is
retained and "fine-tuned" with one final pass through the HingeDescent algorithm
to produce the final solution.
The run time of SweepingHinge is no worse than N j M times that of HingeDescent.
Given this, an upper bound on the (typical) run time for this algorithm (with
M = 1) is O(NNp(d 3 + NcP)). Consequently, SweepingHinge scales reasonably
well in both Nand d, considering the nature of the problem it is designed to solve.
4
Empirical Results
The following experiment was adapted from (Breiman, 1993). The function lex) =
e- lIx ll 2 is sampled at 100d points {xd such that IIxll ~ 3 and IIxll is uniform on [0,3].
The dimension d is varied from 4 to 10 (in steps of 2) and models of size 1 to 20
nodes are trained using the I1AjSweepingHinge algorithm. The number of samples
traversed at each step of the sweep in SweepingHinge was set to M = 10. Nmin
was set equal to 3d throughout. A refitting pass was employed after each new node
was added in the I1A. The refitting algorithm used HingeDescent to "fine-tune"
each node each node before adding the next node. The average sum of squared
Function Approximation with the Sweeping Hinge Algorithm
3000
2500
2000
541
d=4
d=6
d=8
d=10
d=4
d=6
d=8
d=10
1500
1000
500
6
8
10
12
14
16
Number or Nodes
18
20
Figure 4: Upper (lower) curves are for training (test) data.
error, e-2 , was computed for both the training data and an independent set of test
data of size 200d. Plots of 1/e-2 versus the number of nodes are shown in Figure
4. The curves for the training data are clearly bounded below by a linear function
of n (as suggested by inverting the O(l/n) result of Barron's). More importantly
however, they show no significant dependence on the dimension d. The curves for
the test data show the effect of the estimation error as they start to "bend over"
around n = 10 nodes. Again however, they show no dependence on dimension.
Acknowledgements
This work was inspired by the theoretical results of (Barron, 1993) for Sigmoidal
networks as well as the "Hinging Hyperplanes" work of (Breiman, 1993) , and the
"Ramps" work of (Friedman and Breiman, 1994). This work was supported in part
by ONR grant number N00014-95-1-1315.
References
Barron, A.R. (1993) Universal approximation bounds for superpositions of a sigmoidal function. IEEE Transactions on Information Theory 39(3):930-945.
Bellare, M. & Rogaway, P. (1993) The complexity of approximating a nonlinear
program. In P.M. Pardalos (ed.), Complexity in numerical optimization, pp. 16-32,
World Scientific Pub. Co.
Breiman, L. (1993) Hinging hyperplanes for regression, classification and function
approximation. IEEE Transactions on Information Theory 39(3):999-1013.
Breiman, L. & Friedman, J.H. (1994) Function approximation using RAMPS. Snowbird Workshop on Machines that Learn.
Edelsbrunner, H. (1987) In EATCS Monographs on Theoretical Computer Science
V. 10, Algorithms in Combinatorial Geometry. Springer-Verlag.
Jones, L.K. (1992) A simple lemma on greedy approximation in Hilbert space and
convergence rates for projection pursuit regression and neural network training.
The Annals of Statistics, 20:608-613.
Luenberger, D.G. (1984) Introduction to Linear and Nonlinear Programming.
Addison-Wesley.
| 1397 |@word h:3 middle:2 knd:1 polynomial:2 norm:1 nd:5 suitably:1 seek:1 simplifying:1 minus:2 tr:1 initial:5 pub:1 tuned:1 current:3 xiyi:1 si:1 must:2 grahm:1 fn:8 numerical:2 partition:40 engg:1 subsequent:1 designed:2 plot:1 update:5 greedy:1 ria:1 compo:1 provides:1 node:20 location:2 revisited:1 hyperplanes:6 sigmoidal:11 nnp:1 rc:1 along:1 constructed:1 fitting:3 manner:1 pairwise:1 inspired:1 curse:1 considering:1 becomes:1 begin:2 horne:4 bounded:1 minimizes:1 substantially:1 developed:2 finding:1 nj:1 xd:1 prohibitively:1 grant:1 positive:1 before:1 local:5 modify:1 subscript:1 meet:1 plus:2 initialization:1 challenging:2 co:1 unique:3 block:1 definite:1 backpropagation:1 procedure:1 empirical:3 universal:2 projection:2 road:1 induce:2 valley:1 bend:1 risk:2 bill:1 deterministic:2 starting:1 convex:2 importantly:3 annals:1 play:1 user:1 programming:6 designing:1 ep:9 role:2 lix:1 descend:4 region:6 ordering:1 monograph:1 pd:1 complexity:2 trained:1 solving:1 efficiency:1 basis:4 uh:1 completely:1 easily:1 wtrw:3 train:1 elec:1 dichotomy:3 tell:1 exhaustive:1 quite:1 heuristic:1 posed:1 solve:2 widely:1 say:2 ramp:2 statistic:1 f3u:1 itself:1 final:3 sequence:1 advantage:1 differentiable:1 remainder:1 achieve:1 intuitive:1 convergence:2 optimum:1 produce:4 aiu:1 develop:1 n_:2 depending:1 snowbird:1 descends:1 pardalos:1 ao:1 generalization:1 traversed:1 extension:1 around:1 considered:1 algorithmic:2 scope:2 vary:1 omitted:1 purpose:1 estimation:2 combinatorial:2 unexplained:1 superposition:1 clearly:1 always:4 avoid:2 breiman:5 focus:1 rank:2 sense:1 dependent:1 stopping:2 typically:1 entire:1 nand:1 hidden:3 eatcs:1 lien:1 overall:2 classification:1 orientation:1 denoted:2 development:5 constrained:2 special:1 lif:1 equal:2 once:1 f3:1 iif:3 flipped:1 placing:1 jones:2 identical:1 np:4 piecewise:2 few:1 employ:1 simultaneously:1 preserve:1 ve:1 individual:1 geometry:1 ourselves:2 maintain:1 friedman:2 rearrangement:1 interest:1 necessary:1 divide:1 rogaway:2 walk:2 theoretical:2 subset:3 uniform:1 comprised:2 aw:1 st:1 retain:1 refitting:2 nma:1 mechanically:1 continuously:2 squared:2 again:1 nm:1 worse:1 iixll:2 derivative:2 return:1 potential:1 w_:6 inc:1 portion:1 i1a:3 start:4 minimize:1 mlps:1 formed:3 characteristic:1 sy:1 correspond:1 albuquerque:1 produced:2 fo:1 ed:1 pp:1 proof:6 associated:1 sampled:1 tunable:1 popular:1 dimensionality:1 provision:1 hilbert:1 routine:1 positioned:1 wesley:1 methodology:2 emin:2 rand:1 nmin:5 until:1 approximat:1 nonlinear:2 logistic:1 scientific:1 omitting:1 effect:1 contain:1 true:1 remedy:1 lozano:4 assigned:1 ll:1 criterion:4 manifestation:1 performs:1 cp:2 invoked:1 sigmoid:2 qp:3 nh:3 significant:2 refer:1 rd:1 moving:1 stable:8 longer:1 surface:1 base:2 playa:1 edelsbrunner:2 reverse:1 n00014:1 verlag:1 inequality:2 onr:1 approximators:1 accomplished:1 yi:19 swept:1 minimum:6 greater:1 additional:1 employed:1 converge:1 determine:1 fernando:1 ii:2 full:1 determination:1 offer:1 wtr:2 visit:1 regression:2 multilayer:1 sometimes:1 achieved:2 ion:1 qed:1 background:1 addition:2 fine:2 crucial:1 subject:1 call:2 ee:1 easy:1 fit:5 restrict:2 opposite:1 expression:3 proceed:1 enumerate:1 generally:1 iterating:1 detailed:1 tune:1 bellare:2 reduced:3 occupy:1 wr:1 endif:2 sum:1 run:5 parameterized:1 family:1 reasonable:1 throughout:1 scaling:1 comparable:1 entirely:1 layer:3 bound:5 quadratic:5 adapted:1 constraint:2 aspect:1 min:2 expanded:1 poor:1 conjugate:1 across:6 wi:2 tw:1 computationally:2 equation:1 previously:1 addison:1 flip:5 end:4 luenberger:3 available:1 pursuit:2 barron:6 appropriate:1 schmidt:1 include:1 hinge:37 maintaining:1 build:2 approximating:1 seeking:2 move:5 sweep:3 arrangement:1 question:1 flipping:1 lex:1 strategy:1 added:1 dependence:2 traditional:1 gradient:1 separating:1 outer:1 s_:4 unstable:3 code:2 retained:1 minimizing:3 mexico:1 difficult:1 unfortunately:1 stated:1 constructively:1 implementation:2 design:1 proper:1 unknown:1 allowing:2 upper:3 finite:7 descent:2 defining:1 varied:1 sweeping:4 inverting:1 required:1 specified:2 hush:4 able:1 beyond:2 suggested:1 usually:1 below:1 program:1 critical:1 residual:4 numerous:1 started:1 literature:1 acknowledgement:1 determining:2 versus:1 principle:1 argmino:1 translation:1 supported:1 last:1 side:2 allow:2 fall:1 xix:1 boundary:4 dimension:6 curve:3 transition:1 world:1 made:2 transaction:2 approximate:2 compact:1 emphasize:1 global:2 active:1 xi:16 don:1 search:8 continuous:2 iterative:2 un:3 nature:3 terminate:1 robust:1 ncp:2 reasonably:1 learn:1 expansion:1 necessarily:1 terminated:1 motivation:1 allowed:2 augmented:1 referred:4 en:1 position:5 wish:1 exponential:1 lie:1 elseif:2 explored:1 admits:2 workshop:1 adding:1 hinging:3 simply:1 joined:3 wtxi:2 springer:1 satisfies:1 consequently:1 feasible:2 change:1 hard:2 determined:2 infinite:1 except:2 typical:2 wt:1 lemma:5 total:4 pas:2 perceptrons:1 quest:1 constructive:2 dept:1 |
436 | 1,398 | EM Algorithms for PCA and SPCA
Sam Roweis?
Abstract
I present an expectation-maximization (EM) algorithm for principal
component analysis (PCA). The algorithm allows a few eigenvectors and
eigenvalues to be extracted from large collections of high dimensional
data. It is computationally very efficient in space and time. It also naturally accommodates missing infonnation. I also introduce a new variant
of PC A called sensible principal component analysis (SPCA) which defines a proper density model in the data space. Learning for SPCA is also
done with an EM algorithm. I report results on synthetic and real data
showing that these EM algorithms correctly and efficiently find the leading eigenvectors of the covariance of datasets in a few iterations using up
to hundreds of thousands of datapoints in thousands of dimensions.
1
Why EM for peA?
Principal component analysis (PCA) is a widely used dimensionality reduction technique in
data analysis. Its popularity comes from three important properties. First, it is the optimal
(in tenns of mean squared error) linear scheme for compressing a set of high dimensional
vectors into a set of lower dimensional vectors and then reconstructing. Second, the model
parameters can be computed directly from the data - for example by diagonalizing the
sample covariance. Third, compression and decompression are easy operations to perfonn
given the model parameters - they require only matrix multiplications.
Despite these attractive features however, PCA models have several shortcomings. One is
that naive methods for finding the principal component directions have trouble with high
dimensional data or large numbers of datapoints. Consider attempting to diagonalize the
sample covariance matrix of n vectors in a space of p dimensions when n and p are several
hundred or several thousand. Difficulties can arise both in the fonn of computational complexity and also data scarcity. I Even computing the sample covariance itself is very costly,
requiring 0 (np2) operations. In general it is best to avoid altogether computing the sample
? rowei s@cns . cal tech. edu; Computation & Neural Systems, California Institute of Tech.
IOn the data scarcity front, we often do not have enough data in high dimensions for the sample
covariance to be of full rank and so we must be careful to employ techniques which do not require full
rank matrices. On the complexity front, direct diagonalization of a symmetric matrix thousands of
rows in size can be extremely costly since this operation is O(P3) for p x p inputs. Fortunately, several
techniques exist for efficient matrix diagonalization when only the first few leading eigenvectors and
eigerivalues are required (for example the power method [10] which is only O(p2?.
EM Algorithms for PCA and SPCA
627
covariance explicitly. Methods such as the snap-shot algorithm [7] do this by assuming that
the eigenvectors being searched for are linear combinations of the datapoints; their complexity is O(n 3 ). In this note, I present a version of the expectation-maximization (EM)
algorithm [1] for learning the principal components of a dataset. The algorithm does not require computing the sample covariance and has a complexity limited by 0 (knp) operations
where k is the number of leading eigenvectors to be learned.
Another shortcoming of standard approaches to PCA is that it is not obvious how to deal
properly with missing data. Most of the methods discussed above cannot accommodate
missing values and so incomplete points must either be discarded or completed using a
variety of ad-hoc interpolation methods. On the other hand, the EM algorithm for PCA
enjoys all the benefits [4] of other EM algorithms in tenns of estimating the maximum
likelihood values for missing infonnation directly at each iteration.
Finally, the PCA model itself suffers from a critical flaw which is independent of the technique used to compute its parameters: it does not define a proper probability model in the
space of inputs. This is because the density is not nonnalized within the principal subspace.
In other words, if we perfonn PCA on some data and then ask how well new data are fit
by the model, the only criterion used is the squared distance of the new data from their
projections into the principal subspace. A datapoint far away from the training data but
nonetheless near the principal subspace will be assigned a high "pseudo-likelihood" or low
error. Similarly, it is not possible to generate "fantasy" data from a PCA model. In this note
I introduce a new model called sensible principal component analysis (SPCA), an obvious
modification of PC A, which does define a proper covariance structure in the data space. Its
parameters can also be learned with an EM algorithm, given below.
In summary, the methods developed in this paper provide three advantages. They allow
simple and efficient computation of a few eigenvectors and eigenvalues when working with
many datapoints in high dimensions. They permit this computation even in the presence of
missing data. On a real vision problem with missing infonnation, I have computed the 10
leading eigenvectors and eigenvalues of 217 points in 212 dimensions in a few hours using
MATLAB on a modest workstation. Finally, through a small variation, these methods allow
the computation not only of the principal subspace but of a complete Gaussian probabilistic
model which allows one to generate data and compute true likelihoods.
2 Whence EM for peA?
Principal component analysis can be viewed as a limiting case of a particular class of linearGaussian models. The goal of such models is to capture the covariance structure of an observed p-dimensional variable y using fewer than the p{p+ 1) /2 free parameters required in
a full covariance matrix. Linear-Gaussian models do this by assuming that y was produced
as a linear transfonnation of some k-dimensionallatent variable x plus additive Gaussian
noise. Denoting the transfonnation by the p x k matrix C, and the ~dimensional) noise
by v (with covariance matrix R) the generative model can be written as
y = Cx+v
x-N{O,I)
v-N(O,R)
(la)
The latent or cause variables x are assumed to be independent and identically distributed
according to a unit variance spherical Gaussian. Since v are also independent and nonnal
distributed (and assumed independent of x), the model reduces to a single Gaussian model
2 All vectors are column vectors. To denote the transpose of a vector or matrix I use the notation
x T . The determinant of a matrix is denoted by IAI and matrix inversion by A-1. The zero matrix
is 0 and the identity matrix is I. The symbol", means "distributed according to". A multivariate
normal (Gaussian) distribution with mean JL and covariance matrix 1:: is written as N (JL, 1::). The
same Gaussian evaluated at the point x is denoted N (JL, 1::) Ix-
628
S. Roweis
for y which we can write explicitly:
+ R)
(lb)
In order to save parameters over the direct covariance representation in p-space, it is necessary to choose k < p and also to restrict the covariance structure of the Gaussian noise v by
constraining the matrix R.3 For example, if the shape of the noise distribution is restricted
to be axis aligned (its covariance matrix is diagonal) the model is known asfactor analysis.
y ",N (O,CC T
2.1 Inference and learning
There are two central problems of interest when working with the linear-Gaussian models
described above. The first problem is that of state inference or compression which asks:
given fixed model parameters C and R, what can be said about the unknown hidden states
x given some observations y? Since the datapoints are independent, we are interested in
the posterior probability P (xly) over a single hidden state given the corresponding single
observation. This can be easily computed by linear matrix projection and the resulting
density is itself Gaussian:
P(
I ) = P(Ylx)P(x) = N(Cx,R)lyN(O,I)lx
N(O,CCT+R)ly
(2a)
P (xly) = N ((3y,I - (3C) Ix ,
(3 = CT(CC T + R)-l
(2b)
from which we obtain not only the expected value (3y of the unknown state but also an
estimate of the uncertainty in this value in the form of the covariance 1- (3C. Computing
y from x (reconstruction) is also straightforward: P (ylx) = N (Cx, R) Iy. Finally,
xy
P(y)
computing the likelihood of any datapoint y is merely an evaluation under (1 b).
The second problem is that of learning, or parameter fitting which consists of identifying
the matrices C and R that make the model assign the highest likelihood to the observed
data. There are a family of EM algorithms to do this for the various cases of restrictions to
R but all follow a similar structure: they use the inference formula (2b) above in the e-step
to estimate the unknown state and then choose C and the restricted R in the m-step so as
to maximize the expected joint likelihood of the estimated x and the observed y.
2.2 Zero noise limit
Principal component analysis is a limiting case of the linear-Gaussian model as the covariance of the noise v becomes infinitesimally small and equal in all directions. Mathematically, PCA is obtained by taking the limit R = limf~O d. This has the effect of making
the likelihood of a point y dominated solely by the squared distance between it and its reconstruction Cx. The directions of the columns of C which minimize this error are known
as the principal components. Inference now reduces t04 simple least squares projection:
P (xIY)
=N
((3y,I - (3C) Ix ,
(3
= lim C T (CC T + d)-l
f~O
(3a)
P (xly) = N ((CTC)-lC T y, 0) Ix = 6(x - (CTC)-lC T y)
(3b)
Since the noise has become infinitesimal, the posterior over states collapses to a single
point and the covariance becomes zero.
3This restriction on R is not merely to save on parameters: the covariance of the observation noise
must be restricted in some way for the model to capture any interesting or informative projections in
the state x. If R were not restricted, the learning algorithm could simply choose C = 0 and then
set R to be the covariance of the data thus trivially achieving the maximum likelihood model by
explaining all of the structure in the data as noise. (Remember that since the model has reduced to a
single Gaussian distribution for y we can do no better than having the covariance of our model equal
the sample covariance of our data.)
4Recall that if C is p x k with p > k and is rank k then left multiplication by C T (CC T )-l
(which appears not to be well defined because (CC T ) is not invertible) is exactly eqUivalent to left
multiplication by (C T C) -1 CT. The intuition is that even though CC T truly is not invertible, the
directions along which it is not invertible are exactly those which C T is about to project out.
EM Algorithms for PCA and SPCA
3
629
An EM algorithm for peA
The key observation of this note is that even though the principal components can be computed explicitly, there is still an EM algorithm for learning them. It can be easily derived as
the zero noise limit of the standard algorithms (see for example [3, 2] and section 4 below)
by replacing the usual e-step with the projection above. The algorithm is:
? e-step:
? m-step:
x
= (CTC)-lCTy
cnew = YXT(XXT)-l
where Y is a p x n matrix of all the observed data and X is a k x n matrix of the unknown
states. The columns of C will span the space of the first k principal components. (To compute the corresponding eigenvectors and eigenvalues explicitly, the data can be projected
into this k-dimensional subspace and an ordered orthogonal basis for the covariance in the
subspace can be constructed.) Notice that the algorithm can be performed online using
only a single datapoint at a time and so its storage requirements are only O(kp) + O(k2).
The workings of the algorithm are illustrated graphically in figure 1 below.
Gaussian Input Data
'
..
Non-Gaussian Input Data
-.
,.'"
.,
'
"
'.
l.?.?
~
, I ',
~ 0
.
~ 0
-I
',:',
-I
'.
';'
. ~ , ":
-2
:
-2
.'.'
"
(
,'I"
,,'
-~3L--_'7'2-'-'--_~
I -~o----c~--:------:
xl
,
I
.
.
~3~---~2-----~I--~O~--~--~---
xl
Figure 1: Examples of iterations of the algorithm. The left panel shows the learning of the first
principal component of data drawn from a Gaussian distribution, while the right panel shows learning
on data from a non-Gaussian distribution. The dashed lines indicate the direction of the leading
eigenvector of the sample covariance. The dashed ellipse is the one standard deviation contour of
the sample covariance. The progress of the algorithm is indicated by the solid lines whose directions
indicate the guess of the eigenvector and whose lengths indicate the guess of the eigenvalue at each
iteration. The iterations are numbered; number 0 is the initial condition. Notice that the difficult
learning on the right does not get stuck in a local minimum, although it does take more than 20
iterations to converge which is unusual for Gaussian data (see figure 2).
The intuition behind the algorithm is as follows: guess an orientation for the principal
subspace. Fix the guessed subspace and project the data y into it to give the values of the
hidden states x. Nowfix the values ofthe hidden states and choose the subspace orientation
which minimizes the squared reconstruction errors of the datapoints. For the simple twodimensional example above, I can give a physical analogy. Imagine that we have a rod
pinned at the origin which is free to rotate. Pick an orientation for the rod. Holding the
rod still, project every datapoint onto the rod, and attach each projected point to its original
point with a spring. Now release the rod. Repeat. The direction of the rod represents our
guess of the principal component of the dataset. The energy stored in the springs is the
reconstruction error we are trying to minimize.
3.1 Convergence and Complexity
The EM learning algorithm for peA amounts to an iterative procedure for finding the subspace spanned by the k leading eigenvectors without explicit computation of the sample
S. Roweis
630
covariance. It is attractive for small k because its complexity is limited by 0 (knp) per iteration and so depends only linearly on both the dimensionality of the data and the number
of points. Methods that explicitly compute the sample covariance matrix have complexities
limited by 0 (np2), while methods like the snap-shot method that form linear combinations
of the data must compute and diagonalize a matrix of all possible inner products between
points and thus are limited by O(n 2 p) complexity. The complexity scaling of the algorithm
compared to these methods is shown in figure 2 below. For each dimensionality, a random covariance matrix E was generated5 and then lOp points were drawn from N (0, E).
The number of floating point operations required to find the first principal component was
recorded using MATLAB'S flops function. As expected, the EM algorithm scales more
favourably in cases where k is small and both p and n are large. If k ~ p ~ n (we want all
the eigenvectors) then all methods are O(p3).
The standard convergence proofs for EM [I] apply to this algorithm as well, so we can be
sure that it will always reach a local maximum of likelihood. Furthennore, Tipping and
Bishop have shown [8, 9] that the only stable local extremum is the global maximum at
which the true principal subspace is found; so it converges to the correct result. Another
possible concern is that the number of iterations required for convergence may scale with
p or n. To investigate this question, I have explicitly computed the leading eigenvector for
synthetic data sets (as above, with n = lOp) of varying dimension and recorded the number
of iterations of the EM algorithm required for the inner product of the eigendirection with
the current guess of the algorithm to be 0.999 or greater. Up to 450 dimensions (4500
datapoints), the number of iterations remains roughly constant with a mean of 3.6. The
ratios of the first k eigenvalues seem to be the critical parameters controlling the number of
iterations until convergence (For example, in figure I b this ratio was 1.0001.)
Convergence Behaviour
~~metbod
Sompli Covariance + Dill?
Smtple Covariance only
Figure 2: Time complexity and convergence behaviour of the algorithm. In all cases, the number
of datapoints n is 10 times the dimensionality p. For the left panel, the number of floating point
operations to find the leading eigenvector and eigenvalue were recorded. The EM algorithm was
always run for exactly 20 iterations. The cost shown for diagonalization of the sample covariance
uses the MATLAB functions cov and eigs. The snap-shot method is show to indicate scaling only;
one would not normally use it when n > p . In the right hand panel, convergence was investigated
by explicitly computing the leading eigenvector and then running the EM algorithm until the dot
product of its guess and the true eigendirection was 0.999 or more. The error bars show ? one
standard deviation across many runs. The dashed line shows the number of iterations used to produce
the EM algorithm curve ('+') in the left panel.
5First, an axis-aligned covariance is created with the p eigenvalues drawn at random from a uniform distribution in some positive range. Then (p - 1) points are drawn from a p-dimensional zero
mean spherical Gaussian and the axes are aligned in space using these points.
EM Algorithms for PCA and SPCA
631
3.2 Missing data
In the complete data setting, the values of the projections or hidden states x are viewed as
the "missing information" for EM. During the e-step we compute these values by projecting
the observed data into the current subspace. This minimizes the model error given the
observed data and the model parameters. However, if some of the input points are missing
certain coordinate values, we can easily estimate those values in the same fashion. Instead
of estimating only x as the value which minimizes the squared distance between the point
and its reconstruction we can generalize the e-step to:
? generalized e-step: For each (possibly incomplete) point y find the unique pair of
points x? and y. (such that x? lies in the current principal subspace and y. lies in
the subspace defined by the known information about y) which minimize the norm
IICx? - y?lI. Set the corresponding column of X to x* and the corresponding
column ofY to y ?.
If y is complete, then y* = y and x* is found exactly as before. If not, then x* and y* are
the solution to a least squares problem and can be found by, for example, QR factorization
of a particular constraint matrix. Using this generalized e-step I have found the leading
principal components for datasets in which every point is missing some coordinates.
4 Sensible Principal Component Analysis
If we require R to be a multiple ?I of the identity matrix (in other words the covariance
ellipsoid of v is spherical) but do not take the limit as E --t 0 then we have a model which
I shall call sensible principal component analysis or SPCA. The columns of C are still
known as the principal components (it can be shown that they are the same as in regular
PC A) and we will call the scalar value E on the diagonal of R the global noise level. Note
that SPCA uses 1 + pk - k(k - 1)/2 free parameters to model the covariance. Once
again, inference is done with equation (2b). Notice however, that even though the principal
components found by SPCA are the same as those for PCA, the mean of the posterior is
not in general the same as the point given by the PCA projection (3b). Learning for SPCA
also uses an EM algorithm (given below).
Because it has afinite noise level E, SPCA defines a proper generative model and probability
distribution in the data space:
(4)
which makes it possible to generate data from or to evaluate the actual likelihood of new test
data under an SPCA model. Furthermore, this likelihood will be much lower for data far
from the training set even if they are near the principal subspace, unlike the reconstruction
error reported by a PCA model.
The EM algorithm for learning an SPCA model is:
? e-step: {3
? m-step:
= CT(CCT + d)-l
cnew
= Y J..L~:E-l
J..Lx
Enew
= (3Y
= trace[XX T
:Ex
-
= nI -
n{3C
+ J..LxJ..L~
CJ..Lx yT]/n2
Two subtle points about complexity6 are important to notice; they show that learning for
SPCA also enjoys a complexity limited by 0 (knp) and not worse.
6 First, since d is diagonal, the inversion in the e-step can be performed efficiently using the
matrix inversion lemma: {CC T + d)-l = (I/f - C(I + CTC/f)-ICT /f 2 ). Second, since we
are only taking the trace of the matrix in the m-step, we do not need to compute the fu\1 sample
covariance XXT but instead can compute only the variance along each coordinate.
632
S. Roweis
5 Relationships to previous methods
The EM algorithm for PCA, derived above using probabilistic arguments, is closely related
to two well know sets of algorithms. The first are power iteration methods for solving matrix eigenvalue problems. Roughly speaking, these methods iteratively update their eigenvector estimates through repeated mUltiplication by the matrix to be diagonalized. In the
case of PCA, explicitly forming the sample covariance and multiplying by it to perform
such power iterations would be disastrous. However since the sample covariance is in fact
a sum of outer products of individual vectors, we can multiply by it efficiently without ever
computing it. In fact, the EM algorithm is exactly equivalent to performing power iterations
for finding C using this trick. Iterative methods for partial least squares (e.g. the NIPALS
algorithm) are doing the same trick for regression. Taking the singular value decomposition
(SVD) of the data matrix directly is a related way to find the principal subspace. If Lanczos or Arnoldi methods are used to compute this SVD, the resulting iterations are similar
to those of the EM algorithm. Space prohibits detailed discussion of these sophisticated
methods, but two excellent general references are [5, 6]. The second class of methods are
the competitive learning methods for finding the principal subspace such as Sanger's and
Oja's rules. These methods enjoy the same storage and time complexities as the EM algorithm; however their update steps reduce but do not minimize the cost and so they typically
need more iterations and require a learning rate parameter to be set by hand.
Acknowledgements
I would like to thank John Hopfield and my fellow graduate students for constant and excellent
feedback on these ideas. In particular I am grateful to Erik Winfree for significant contributions to the
missing data portion of this work, to Dawei Dong who provided image data to try as a real problem,
as well as to Carlos Brody, Sanjoy Mahajan, and Maneesh Sahani. The work of Zoubin Ghahrarnani
and Geoff Hinton was an important motivation for this study. Chris Bishop and Mike Tipping are
pursuing independent but yet unpublished work on a virtually identical model. The comments of
three anonymous reviewers and many visitors to my poster improved this manuscript greatly.
References
[I] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data via
the EM algorithm. Journal of the Royal Statistical Society series B, 39: 1-38, 1977.
[2] B. S. Everitt. An Introducction to Latent Variable Models. Chapman and Hill, London, 1984.
[3] Zoubin Ghahramani and Geoffrey Hinton. The EM algorithm for mixtures of factor analyzers.
Technical Report CRG-TR-96-1 , Dept. of Computer Science, University of Toronto, Feb. 1997.
[4] Zoubin Ghahramani and Michael I. Jordan. Supervised learning from incomplete data via an
EM approach. In Jack D. Cowan, Gerald Tesauro, and Joshua Alspector, editors, Advances in
Neural Information Processing Systems , volume 6, pages 120-127. Morgan Kaufmann, 1994.
[5] Gene H. Golub and Charles F. Van Loan. Matrix Computations. The Johns Hopkins University
Press, Baltimore, MD, USA, second edition, 1989.
[6] R. B. Lehoucq, D. C. Sorensen, and C. Yang. Arpack users' guide: Solution of large
scale eigenvalue problems with implicitly restarted Arnoldi methods. Technical Report
from http://www.caam.rice.edu/software/ARPACK/, Computational and Applied Mathematics, Rice University, October 1997.
[7] L. Sirovich. Turbulence and the dynamics of coherent structures. Quarterly Applied Mathematics, 45(3):561-590, 1987.
[8] Michael Tipping and Christopher Bishop. Mixtures of probabilistic principal component analyzers. Technical Report NCRG/97/003 , Neural Computing Research Group, Aston University,
June 1997.
[9] Michael Tipping and Christopher Bishop. Probabilistic principal component analysis. Technical
Report NCRG/97/010, Neural Computing Research Group, Aston University, September 1997.
[10] J. H. Wilkinson. The AlgebraiC Eigenvalue Problem. Claredon Press, Oxford, England, 1965.
| 1398 |@word determinant:1 version:1 inversion:3 compression:2 norm:1 covariance:37 decomposition:1 fonn:1 asks:1 pick:1 tr:1 solid:1 accommodate:1 shot:3 reduction:1 initial:1 series:1 xiy:1 denoting:1 diagonalized:1 current:3 yet:1 must:4 written:2 john:2 additive:1 informative:1 shape:1 update:2 generative:2 fewer:1 guess:6 toronto:1 lx:3 along:2 constructed:1 direct:2 become:1 consists:1 fitting:1 introduce:2 expected:3 alspector:1 roughly:2 cct:2 spherical:3 actual:1 becomes:2 project:3 estimating:2 notation:1 provided:1 panel:5 xx:1 what:1 fantasy:1 minimizes:3 eigenvector:6 prohibits:1 developed:1 finding:4 extremum:1 perfonn:2 pseudo:1 remember:1 every:2 fellow:1 exactly:5 k2:1 unit:1 ly:1 normally:1 enjoy:1 eigendirection:2 arnoldi:2 positive:1 before:1 local:3 limit:4 despite:1 oxford:1 solely:1 interpolation:1 plus:1 limited:5 collapse:1 lop:2 range:1 factorization:1 graduate:1 unique:1 procedure:1 maneesh:1 poster:1 projection:7 word:2 regular:1 numbered:1 zoubin:3 get:1 cannot:1 onto:1 cal:1 turbulence:1 storage:2 twodimensional:1 restriction:2 equivalent:2 www:1 reviewer:1 missing:11 yt:1 straightforward:1 graphically:1 identifying:1 rule:1 spanned:1 datapoints:8 variation:1 coordinate:3 limiting:2 imagine:1 controlling:1 user:1 us:3 origin:1 trick:2 observed:6 mike:1 capture:2 thousand:4 compressing:1 highest:1 sirovich:1 lxj:1 intuition:2 dempster:1 complexity:12 wilkinson:1 dynamic:1 gerald:1 grateful:1 solving:1 basis:1 easily:3 joint:1 hopfield:1 geoff:1 various:1 xxt:2 shortcoming:2 london:1 kp:1 whose:2 widely:1 snap:3 furthennore:1 cov:1 itself:3 laird:1 online:1 hoc:1 advantage:1 eigenvalue:11 reconstruction:6 dimensionallatent:1 product:4 aligned:3 roweis:4 qr:1 convergence:7 requirement:1 produce:1 converges:1 progress:1 p2:1 come:1 xly:3 indicate:4 direction:7 closely:1 correct:1 pea:4 pinned:1 require:5 behaviour:2 assign:1 fix:1 dill:1 anonymous:1 mathematically:1 crg:1 knp:3 normal:1 infonnation:3 gaussian:18 always:2 avoid:1 varying:1 np2:2 derived:2 release:1 ax:1 june:1 properly:1 rank:3 likelihood:12 tech:2 greatly:1 am:1 whence:1 inference:5 flaw:1 typically:1 hidden:5 interested:1 nonnal:1 orientation:3 denoted:2 equal:2 once:1 having:1 chapman:1 identical:1 represents:1 report:5 few:5 employ:1 oja:1 individual:1 floating:2 cns:1 interest:1 investigate:1 multiply:1 evaluation:1 golub:1 truly:1 mixture:2 pc:3 behind:1 sorensen:1 fu:1 partial:1 necessary:1 xy:1 modest:1 orthogonal:1 incomplete:4 column:6 lanczos:1 maximization:2 cost:2 deviation:2 hundred:2 uniform:1 front:2 stored:1 reported:1 synthetic:2 yxt:1 my:2 density:3 probabilistic:4 dong:1 invertible:3 michael:3 iy:1 hopkins:1 again:1 squared:5 central:1 recorded:3 choose:4 possibly:1 worse:1 leading:10 li:1 student:1 explicitly:8 ad:1 depends:1 performed:2 try:1 doing:1 portion:1 competitive:1 carlos:1 contribution:1 minimize:4 square:3 ni:1 variance:2 who:1 efficiently:3 kaufmann:1 guessed:1 ofthe:1 generalize:1 produced:1 multiplying:1 cc:7 datapoint:4 reach:1 suffers:1 infinitesimal:1 nonetheless:1 energy:1 obvious:2 naturally:1 proof:1 workstation:1 dataset:2 ask:1 recall:1 lim:1 dimensionality:4 cj:1 subtle:1 sophisticated:1 appears:1 manuscript:1 tipping:4 supervised:1 follow:1 improved:1 iai:1 done:2 evaluated:1 though:3 furthermore:1 until:2 hand:3 working:3 favourably:1 replacing:1 christopher:2 cnew:2 defines:2 indicated:1 usa:1 effect:1 requiring:1 true:3 assigned:1 symmetric:1 iteratively:1 mahajan:1 illustrated:1 deal:1 attractive:2 during:1 criterion:1 generalized:2 trying:1 hill:1 complete:3 dawei:1 image:1 jack:1 charles:1 ctc:4 physical:1 volume:1 jl:3 discussed:1 ncrg:2 significant:1 everitt:1 trivially:1 mathematics:2 similarly:1 decompression:1 analyzer:2 dot:1 stable:1 feb:1 multivariate:1 posterior:3 tesauro:1 certain:1 tenns:2 joshua:1 morgan:1 minimum:1 fortunately:1 greater:1 converge:1 maximize:1 dashed:3 full:3 ofy:1 multiple:1 reduces:2 technical:4 england:1 variant:1 regression:1 vision:1 expectation:2 winfree:1 iteration:18 ion:1 want:1 baltimore:1 singular:1 diagonalize:2 unlike:1 sure:1 comment:1 virtually:1 cowan:1 transfonnation:2 seem:1 jordan:1 call:2 near:2 presence:1 yang:1 spca:15 constraining:1 easy:1 enough:1 identically:1 variety:1 fit:1 restrict:1 inner:2 reduce:1 idea:1 rod:6 pca:18 algebraic:1 speaking:1 cause:1 matlab:3 detailed:1 eigenvectors:10 ylx:2 amount:1 reduced:1 generate:3 http:1 exist:1 notice:4 estimated:1 correctly:1 popularity:1 per:1 write:1 shall:1 group:2 key:1 achieving:1 drawn:4 merely:2 sum:1 run:2 uncertainty:1 family:1 pursuing:1 p3:2 scaling:2 brody:1 ct:3 constraint:1 software:1 dominated:1 argument:1 extremely:1 span:1 spring:2 attempting:1 performing:1 infinitesimally:1 according:2 combination:2 across:1 em:33 sam:1 reconstructing:1 modification:1 making:1 caam:1 projecting:1 restricted:4 computationally:1 equation:1 remains:1 know:1 unusual:1 operation:6 permit:1 apply:1 quarterly:1 away:1 save:2 altogether:1 original:1 running:1 trouble:1 completed:1 sanger:1 ghahramani:2 ellipse:1 society:1 question:1 costly:2 usual:1 diagonal:3 md:1 said:1 lehoucq:1 september:1 subspace:17 distance:3 thank:1 accommodates:1 sensible:4 outer:1 chris:1 eigs:1 evaluate:1 assuming:2 erik:1 length:1 relationship:1 ellipsoid:1 ratio:2 difficult:1 october:1 disastrous:1 holding:1 trace:2 proper:4 unknown:4 perform:1 observation:4 datasets:2 discarded:1 flop:1 lyn:1 nonnalized:1 ever:1 hinton:2 lb:1 pair:1 required:5 unpublished:1 california:1 coherent:1 learned:2 hour:1 bar:1 below:5 royal:1 power:4 critical:2 difficulty:1 attach:1 diagonalizing:1 scheme:1 aston:2 axis:2 created:1 nipals:1 naive:1 sahani:1 ict:1 acknowledgement:1 multiplication:4 interesting:1 analogy:1 geoffrey:1 rubin:1 editor:1 row:1 summary:1 repeat:1 free:3 transpose:1 enjoys:2 guide:1 allow:2 institute:1 explaining:1 taking:3 benefit:1 distributed:3 curve:1 dimension:7 feedback:1 van:1 contour:1 stuck:1 collection:1 projected:2 far:2 implicitly:1 arpack:2 gene:1 global:2 assumed:2 afinite:1 latent:2 iterative:2 why:1 lineargaussian:1 investigated:1 excellent:2 pk:1 linearly:1 motivation:1 noise:12 arise:1 edition:1 n2:1 scarcity:2 repeated:1 fashion:1 lc:2 explicit:1 xl:2 lie:2 third:1 ix:4 formula:1 bishop:4 showing:1 symbol:1 concern:1 diagonalization:3 cx:4 simply:1 forming:1 ordered:1 scalar:1 restarted:1 extracted:1 rice:2 viewed:2 goal:1 identity:2 careful:1 loan:1 principal:31 lemma:1 called:2 sanjoy:1 svd:2 la:1 searched:1 rotate:1 visitor:1 dept:1 ex:1 |
437 | 1,399 | Task and Spatial Frequency Effects on Face
Specialization
Matthew N. Dailey
Garrison W. Cottrell
Department of Computer Science and Engineering
U.C. San Diego
La Jolla, CA 92093-0114
{mdailey,gary}@cs.ucsd.edu
Abstract
There is strong evidence that face processing is localized in the brain.
The double dissociation between prosopagnosia, a face recognition
deficit occurring after brain damage, and visual object agnosia, difficulty
recognizing otber kinds of complex objects, indicates tbat face and nonface object recognition may be served by partially independent mechanisms in the brain. Is neural specialization innate or learned? We suggest that this specialization could be tbe result of a competitive learning mechanism that, during development, devotes neural resources to the
tasks they are best at performing. Furtber, we suggest that the specialization arises as an interaction between task requirements and developmental constraints. In this paper, we present a feed-forward computational
model of visual processing, in which two modules compete to classify
input stimuli. When one module receives low spatial frequency information and the other receives high spatial frequency information, and
the task is to identify the faces while simply classifying the objects, the
low frequency network shows a strong specialization for faces. No otber
combination of tasks and inputs shows this strong specialization. We
take these results as support for the idea that an innately-specified face
processing module is unnecessary.
1 Background
Studies of the preserved and impaired abilities in brain damaged patients provide important
clues on how the brain is organized. Cases of prosop agnosia, a face recognition deficit often
sparing recognition of non-face objects, and visual object agnosia, an object recognition
deficit that can occur witbout appreciable impairment of face recognition, provide evidence
that face recognition is served by a "special" mechanism. (For a recent review of this
18
M. N. Dailey and G. W. Cottrell
evidence, see Moscovitch, Winocur, and Behrmann (1997?. In this study, we begin to
provide a computational account of the double dissociation.
Evidence indicates that face recognition is based primarily on holistic, configural information, whereas non-face object recognition relies more heavily on local features and analysis
of the parts of an object (Farah, 1991; Tanaka and Sengco, 1997). For instance, the distance
between the tip of the nose and an eye in a face is an important factor in face recognition,
but such subtle measurements are rarely as critical for distinguishing, say, two buildings.
There is also evidence that con figural information is highly relevant when a human becomes an "expert" at identifying individuals within other visually homogeneous object
classes (Gauthier and Tarr, 1997).
What role might configural information play in the development of a specialization for face
recognition? de Schonen and Mancini (1995) have proposed that several factors, including
different rates of maturation in different areas of cortex, an infant's tendency to track the
faces in its environment, and the gradual increase in visual acuity as an infant develops,
all combine to force an early specialization for face recognition. If this scenario is correct,
the infant begins to form configural face representations very soon after birth, based primarily on the low spatial frequency information present in face stimuli. Indeed, Costen,
Parker, and Craw (1996) showed that although both high-pass and low-pass image filtering decrease face recognition accuracy, high-pass filtering degrades identification accuracy
more quickly than low-pass filtering. Furthermore, Schyns and Oliva (1997) have shown
that when asked to recognize the identity of the "face" in a briefly-presented hybrid image
containing a low-pass filtered image of one individual's face and a high-pass filtered image
of another individual's face, subjects consistently use the lOW-frequency compone.nt of the
image for the task. This work indicates that low spatial frequency information may be more
important for face identification than high spatial frequency information.
Jacobs and Kosslyn (1994) showed how differential availability of large and small receptive
field sizes in a mixture of experts network (Jacobs, Jordan, Nowlan, and Hinton, 1991)
can lead to experts that specialize for "what" and "where" tasks. In previous work, we
proposed that a neural mechanism allocating resources according to their ability to perform
a given task could explain the apparent specialization for face recognition evidenced by
prosopagnosia (Dailey, Cottrell, and Padgett, 1997). We showed that a model based on
the mixture of experts architecture, in which a gating network implements competitive
learning between two simple homogeneous modules, could develop a specialization such
that damage to one module disproportionately impaired face recognition compared to nonface object recognition.
In the current study, we consider how the availability of spatial frequency information affects face recognition specialization given this hypothesis of neural resource allocation by
competitive learning. We find that when high and low frequency information is "split"
between the two modules in our system, and the task is to identify the faces while simply
classifying the objects, the low-frequency module consistently specializes for face recognition. After describing the study, we discuss its results and their implications.
2 Experimental Methods
We presented a modular feed-forward neural network preprocessed images of 12 different faces, 12 different books, 12 different cups, and 12 different soda cans. We gave the
network two types of tasks:
1. Learning to recognize the superordinate classes of all four object types (hereafter
referred to as classification).
2. Learning to distinguish the individual members of one class (hereafter referred to
Task and Spatial Frequency Effects on Face Specialization
19
as identification) while simply classifying objects of the other three types.
For each task, we investigated the effects of high and low spatial frequency information on
identification and classification in a visual processing system with two competing modules.
We observed how splitting the range of spatial frequency information between the two
modules affected the specializations developed by the network.
2.1
Image Data
We acquired face images from the Cottrell and Metcalfe facial expression database (1991 )
and captured multiple images of several books, cups, and soda cans with a CCD camera
and video frame grabber. For the face images, we chose five grayscale images of each of 12
individuals. The images were photographed under controlled lighting and pose conditions;
the subjects portrayed a different facial expression in each image. For each of the non-face
object classes, we captured five different grayscale images of each of 12 books, 12 cups,
and 12 cans. These images were also captured under controlled lighting conditions, with
small variations in position and orientation between photos. The entire image set contained
240 images, each of which we cropped and scaled to a size of 64x64 pixels.
2.2 Image Preprocessing
To convert the raw grayscale images to a biologically plausible representation more suitable
for network learning and generalization, and to experiment with the effect of high and
low spatial frequency information available in a stimulus, we extracted Gabor jet features
from the images at multiple spatial frequency scales then performed a separate principal
components analysis on the data from each filter scale separately to reduce input pattern
dimensionality.
2.2.1
Gabor jet features
The basic two-dimensional Gabor wavelet resembles a sinusoid grating restricted by a twodimensional Gaussian, and may be tuned to a particular orientation and sinusoidal frequency scale. The wavelet can be used to model simple cell receptive fields in cat primary
visual cortex (Jones and Palmer, 1987). Buhmann, Lades, and von der Malsburg (1990)
describe the Gabor "jet," a vector consisting of filter responses at multiple orientations and
scales.
We convolved each of the 240 images in the input data set with two-dimensional Gabor
? elg
? h t onentatlons
.
.
(0 '8'
11' 11' 311' 11' 511' 311' 711')
filters at fi ve scaIes m
4' 8' 2' T' 4' 8 an d sub sampled an
8x8 grid of the responses to each filter. The process resulted in 2560 complex numbers
describing each image.
2.2.2 Principal components analysis
To reduce the dimensionality of the Gabor jet representation while maintaining a segregation of the responses from each filter scale, we performed a separate PCA on each spatial
frequency component of the pattern vector described above. For each of the 5 filter scales
in the jet, we extracted the subvectors corresponding to that scale from each pattern in the
training set, computed the eigenvectors of their covariance matrix, projected the sub vectors
from each of the patterns onto these eigenvectors, and retained the eight most significant
coefficients. Reassembling the pattern set resulted in 240 40-dimensional vectors.
M. N. Dailey and G. W. Cottrell
20
Module 1
Inputs
Figure 1: Modular network architecture. The gating network units mix the outputs of the
hidden layers multiplicatively.
2.3 The Model
The model is a simple modular feed-forward network inspired by the mixture of experts
architecture (Jordan and Jacobs, 1995); however, it contains hidden layers and is trained by
backpropagation of error rather than maximum likelihood estimation or expectation maximization. The connections to the output units come from two separate inputlhidden layer
pairs; these connections are gated multiplicatively by a simple linear network with softmax outputs. Figure 1 illustrates the model's architecture. During training, the network's
weights are adjusted by backpropagation of error. The connections from the softmax units
in the gating network to the connections between the hidden layers and output layer can
be thought of as multiplicative connections with a constant weight of 1. The resulting
learning rules gate the amount of error feedback received by a module according to the
gating network's current estimate of its ability to process the current training pattern. Thus
the model implements a form of competitive learning in which the gating network learns
which module is better able to process a given pattern and rewards the "winner" with more
error feedback.
2.4 Training Procedure
Preprocessing the images resulted in 240 40-dimensional vectors; four examples of each
face and object composed a I92-element training set, and one example of each face and
object composed a 48-element test set. We held out one example of each individual in
the training set for use in determining when to stop network training. We set the learning
rate for all network weights to 0.1 and their momentum to 0.5. Both of the hidden layers
contained 15 units in all experiments. For the identification tasks, we determined that a
mean squared error (MSE) threshold of 0.02 provided adequate classification performance
on the hold out set without overtraining and allowed the gate network to settle to stable
values. For the four-way classification task, we found that an MSE threshold of 0.002 was
necessary to give the gate network time to stabilize and did not result in overtraining. On
all runs reported in the results section, we simply trained the network until it reached the
relevant MSE threshold.
For each of the tasks reported in the results section (four-way classification, book identification, and face identification), we performed two experiments. In the first, as a control,
both modules and the gating network were trained and tested with the fu1l40-dimensional
pattern vector. In the second, the gating network received the full 40-dimensional vector,
Task and Spatial Frequency Effects on Face Specialization
21
but module 1 received a vector in which the elements corresponding to the largest two Gabor filter scales were set to 0, and the elements corresponding to the middle filter scale were
reduced by 0.5. Module 2, on the other hand, received a vector in which the elements corresponding to the smallest two filter scales were set to 0 and the elements corresponding to
the middle filter were reduced by 0.5. Thus module 1 received mostly high-frequency information, whereas module 2 received mostly low-frequency information, with deemphasized
overlap in the middle range.
For each of these six experiments, we trained the network using 20 different initial random
weight sets and recorded the softmax outputs learned by the gating network on each training
pattern.
3 Results
Figure 2 displays the resulting degree of specialization of each module on each stimulus
class. Each chart plots the average weight the gating network assigns to each module for
the training patterns from each stimulus class, averaged over 20 training runs with different
initial random weights. The error bars denote standard error. For each of the three reported
tasks (four-way claSSification, book identification, and face identification), one chart shows
division of labor between the two modules in the control situation, in which both modules
receive the same patterns, and the other chart shows division of labor between the two
modules when one module receives low-frequency information and the other receives highfrequency information.
When required to identify faces on the basis of high- or lOW-frequency information, compared with the four-way-classification and same-pattern controls, the lOW-frequency module wins the competition for face patterns extremely consistently (lower right graph). Book
identification specialization, however, shows considerably less sensitivity to spatial frequency.
We have also performed the equivalent experiments with a cup discrimination and a can
discrimination task. Both of these tasks show a low-frequency sensitivity lower than that
for face identification but higher than that for book identification. Due to space limitations,
these results are not presented here.
The specialized face identification networks also provide good models of prosopagnosia
and visual object agnosia: when the face-specialized module's output is "damaged" by removing connections from its hidden layer to the output layer, the overall network's generalization performance on face identification drops dramatically, while its generalization performance on object recognition drops much more slowly. When the non-face-specialized
(high frequency) module'S outputs are damaged, the opposite effect occurs: the overall network's performance on each of the object recognition tasks drops, whereas its performance
on face identification remains high.
4 Discussion
The results in Figure 2 show a strong preference for low-frequency information in the face
identification task, empirically demonstrating that, given a choice, a competitive mechanism will choose a module receiving low-frequency, large receptive field information for
this task. This result concurs with the psychological evidence for configural face representations based upon low spatial frequency information, and suggests how the developing
brain could be biased toward a specialization for face recognition by the infant's initially
low visual acuity.
On the basis of these results, we predict that human subjects performing face and object
22
M. N Dailey and G. W. Cottrell
Classification (split frequencies)
Classification (control)
i
'0;
~
i
:
1.0
\.0
0.8
0.8
0.6
1m Module 1
a
Module 2
0.4
~
i"11
r
~
CModule 1
(highfreq)
? Module 2
(low freq)
0.6
0.4
~
0.2
0.2
0.0
0.0
Faces Books Cups Cans
Faces Books Cups Cans
SdmulusType
Stimulus Type
Bookid task (control)
Book id task (split frequencies)
\.0
1.0
0.8
0.8
j
j
-;
f
0.6
r:::::J Module 1
. . Module 2
0.4
-;
0.6
r
0.4
:
0.2
0.2
0.0
0.0
Faces Books Cups Cans
Faces Books Cups Cans
Sdmul.. Type
Stimulus Type
Face id task (control)
r
:
(highfreq)
. . Module 2
(low freq)
~
~
i-;
c:I Module 1
Face id task (split frequencies)
1.0
1.0
0.8
0.8
i
0.6
C!I Module 1
. . Module 2
0.4
~
-;
r
Cl Module 1
0.6
(high freq)
IilII Module 2
0.4
(low freq)
~
0.2
0.2
0.0
0.0
Faces Books Cup. Cans
Stimulus Type
Faces Books Cups Cans
Stimulus Type
Figure 2: Average weight assigned to each module broken down by stimulus class. For each
task, in the control experiment, each module receives the same pattern; the split-frequency
charts summarize the specialization resulting when module 1 receives high-frequency Gabor filter information and module 2 receives low-frequency Gabor filter information.
Task and Spatial Frequency Effects on Face Specialization
23
identification tasks will show more degradation of performance in high-pass filtered images
of faces than in high-pass filtered images of other objects. To our knowledge, this has not
been empirically tested, although Costen et al. (1996) have investigated the effect of highpass and low-pass filtering on face images in isolation, and Parker, Lishman, and Hughes
(1996) have investigated the effect of high-pass and low-pass filtering of face and object
images used as 100 ms cues for a same/different task. Their results indicate that relevant
high-pass filtered images cue object processing better than low-pass filtered images, but the
two types of filtering cue face processing equally well. Similarly, Schyns & Oliva's (1997)
results described earlier suggest that the human face identification network preferentially
responds to low spatial frequency inputs.
Our results suggest that simple data-driven competitive learning combined with constraints
and biases known or thought to exist during visual system development can account for
some of the effects observed in normal and brain-damaged humans. The study lends support to the claim that there is no need for an innately-specified face processing module face recognition is only "special" insofar as faces form a remarkably homogeneous category of stimuli for which Within-category discrimination is ecologically beneficial.
References
Buhmann, J., Lades, M., and von der Malsburg, C. (1990). Size and distortion invariant object recognition by hierarchical graph matching. In Proceedings of the IJCNN
International Joint Conference on Neural Networks, volume II, pages 411-416.
Costen, N., Parker, D., and Craw, I. (1996). Effects of high-pass and low-pass spatial
filtering on face identification. Perception & Psychophysics, 38(4):602-612.
Cottrell, G. and Metcalfe, J. (1991). Empath: Face, gender and emotion recognition using
holons. In Lippman, R., Moody, J., and Touretzky, D., editors, Advances in Neural
Information ProceSSing Systems 3, pages 564-571.
Dailey, M., Cottrell, G., and Padgett, C. (1997). A mixture of experts model exhibiting
pro sop agnosia. In Proceedings of the Nineteenth Annual Conference of the Cognitive
Science Society, pp. 155-160. Stanford, CA, Mahwah: Lawrence Erlbaum.
de Schonen, S. and Mancini, J. (1995). About functional brain specialization: The development of face recognition. TR 95.1, MRC Cognitive Development Unit, London,
UK.
Farah, M. (1991). Patterns of co-occurrence among the associative agnosias: Implications
for visual object representation. Cognitive Neuropsychology, 8:1-19.
Gauthier, I. and Tarr, M. (1997). Becoming a "greeble" expert: Exploring mechanisms for
face recognition. Vision Research. In press.
Jacobs, R. and Kosslyn, S. (1994). Encoding shape and spatial relations - The role of
receptive field size in coordinating complementary representations. Cognitive Science,
18(3):361-386.
Jacobs, R, Jordan, M., Nowlan, S., and Hinton, G. (1991). Adaptive mixtures of local
experts. Neural Computation, 3:79-87.
Jones, J. and Palmer, L. (1987). An evaluation of the two-dimensional Gabor filter model
of simple receptive fields in cat striate cortex. 1 Neurophys., 58(6):1233-1258.
Moscovitch, M., Winocur, G., and Behrmann, M. (1997). What is special about face
recognition? Nineteen experiments on a person with visual object agnosia and dyslexia
but normal face recognition. Journal of Cognitive Neuroscience, 9(5):555-604.
Parker, D., Lishman, J., and Hughes, J. (1996). Role of coarse and fine spatial information
in face and object processing. Journal of Experimental Psychology: Human Perception
and Performance, 22(6):1445-1466.
Schyns, P. and Oliva, A. (1997). Dr. Angry and Mr. Smile: The multiple faces of perceptual
categorizations. Submitted for publication.
Tanaka, J. and Sengco, 1. (1997). Features and their configuration in face recognition.
Memory and Cognition. In press.
| 1399 |@word middle:3 briefly:1 gradual:1 jacob:5 covariance:1 tr:1 initial:2 configuration:1 contains:1 hereafter:2 empath:1 tuned:1 current:3 nt:1 nowlan:2 neurophys:1 cottrell:8 shape:1 plot:1 drop:3 discrimination:3 infant:4 cue:3 filtered:6 coarse:1 preference:1 five:2 differential:1 specialize:1 combine:1 acquired:1 indeed:1 brain:8 inspired:1 subvectors:1 becomes:1 begin:2 provided:1 what:3 kind:1 developed:1 configural:4 holons:1 scaled:1 uk:1 control:7 unit:5 engineering:1 local:2 encoding:1 id:3 becoming:1 might:1 chose:1 resembles:1 suggests:1 co:1 palmer:2 range:2 averaged:1 camera:1 hughes:2 implement:2 witbout:1 backpropagation:2 lippman:1 procedure:1 area:1 gabor:10 thought:2 matching:1 suggest:4 onto:1 prosopagnosia:3 twodimensional:1 equivalent:1 identifying:1 splitting:1 assigns:1 rule:1 x64:1 variation:1 diego:1 play:1 damaged:4 heavily:1 padgett:2 homogeneous:3 distinguishing:1 hypothesis:1 element:6 recognition:29 database:1 observed:2 role:3 module:43 sparing:1 decrease:1 neuropsychology:1 developmental:1 environment:1 broken:1 reward:1 asked:1 trained:4 upon:1 division:2 basis:2 joint:1 cat:2 describe:1 london:1 tbat:1 birth:1 apparent:1 modular:3 stanford:1 plausible:1 nineteenth:1 say:1 distortion:1 ability:3 otber:2 associative:1 interaction:1 relevant:3 figural:1 holistic:1 competition:1 double:2 requirement:1 impaired:2 categorization:1 object:28 develop:1 pose:1 received:6 grating:1 strong:4 c:1 come:1 indicate:1 exhibiting:1 correct:1 filter:13 human:5 settle:1 disproportionately:1 generalization:3 adjusted:1 exploring:1 hold:1 normal:2 visually:1 lawrence:1 cognition:1 predict:1 claim:1 matthew:1 early:1 smallest:1 winocur:2 estimation:1 largest:1 gaussian:1 rather:1 publication:1 acuity:2 consistently:3 indicates:3 likelihood:1 entire:1 initially:1 hidden:5 relation:1 pixel:1 overall:2 classification:9 orientation:3 among:1 development:5 spatial:21 special:3 softmax:3 psychophysics:1 field:5 emotion:1 tarr:2 jones:2 stimulus:11 develops:1 primarily:2 composed:2 recognize:2 ve:1 individual:6 resulted:3 consisting:1 highly:1 evaluation:1 mixture:5 held:1 implication:2 allocating:1 dailey:6 necessary:1 facial:2 psychological:1 instance:1 classify:1 earlier:1 maximization:1 recognizing:1 erlbaum:1 reported:3 considerably:1 combined:1 person:1 international:1 sensitivity:2 receiving:1 tip:1 quickly:1 moody:1 von:2 squared:1 recorded:1 containing:1 choose:1 slowly:1 dr:1 cognitive:5 book:14 expert:8 account:2 sinusoidal:1 de:2 stabilize:1 availability:2 coefficient:1 devotes:1 innately:2 performed:4 multiplicative:1 reached:1 competitive:6 farah:2 chart:4 accuracy:2 dissociation:2 identify:3 greeble:1 identification:19 raw:1 ecologically:1 served:2 lighting:2 mrc:1 submitted:1 overtraining:2 explain:1 touretzky:1 frequency:38 pp:1 reassembling:1 con:1 sampled:1 stop:1 knowledge:1 dimensionality:2 organized:1 subtle:1 feed:3 higher:1 maturation:1 response:3 furthermore:1 until:1 hand:1 receives:7 gauthier:2 innate:1 building:1 effect:11 lades:2 assigned:1 sinusoid:1 freq:4 during:3 nonface:2 m:1 pro:1 image:29 fi:1 specialized:3 functional:1 empirically:2 winner:1 volume:1 measurement:1 significant:1 cup:10 grid:1 similarly:1 stable:1 cortex:3 recent:1 showed:3 jolla:1 driven:1 scenario:1 der:2 captured:3 grabber:1 mr:1 ii:1 multiple:4 mix:1 full:1 jet:5 equally:1 controlled:2 basic:1 oliva:3 patient:1 expectation:1 vision:1 cell:1 receive:1 preserved:1 remarkably:1 background:1 whereas:3 separately:1 cropped:1 fine:1 biased:1 subject:3 nineteen:1 member:1 smile:1 jordan:3 split:5 insofar:1 affect:1 isolation:1 gave:1 psychology:1 architecture:4 competing:1 opposite:1 reduce:2 idea:1 specialization:20 expression:2 pca:1 six:1 adequate:1 impairment:1 dramatically:1 eigenvectors:2 amount:1 category:2 reduced:2 exist:1 coordinating:1 neuroscience:1 track:1 affected:1 four:6 threshold:3 demonstrating:1 preprocessed:1 graph:2 tbe:1 convert:1 compete:1 run:2 soda:2 layer:8 angry:1 distinguish:1 display:1 annual:1 occur:1 ijcnn:1 constraint:2 extremely:1 performing:2 photographed:1 department:1 developing:1 according:2 combination:1 beneficial:1 kosslyn:2 biologically:1 restricted:1 invariant:1 resource:3 segregation:1 remains:1 concurs:1 discus:1 describing:2 mechanism:6 nose:1 photo:1 available:1 eight:1 hierarchical:1 occurrence:1 dyslexia:1 gate:3 convolved:1 deemphasized:1 mancini:2 ccd:1 maintaining:1 malsburg:2 society:1 occurs:1 damage:2 degrades:1 receptive:5 primary:1 responds:1 highfrequency:1 striate:1 win:1 lends:1 distance:1 deficit:3 separate:3 toward:1 retained:1 multiplicatively:2 preferentially:1 mostly:2 sop:1 perform:1 gated:1 situation:1 hinton:2 frame:1 ucsd:1 highpass:1 evidenced:1 pair:1 required:1 specified:2 connection:6 learned:2 tanaka:2 able:1 bar:1 superordinate:1 pattern:15 perception:2 summarize:1 including:1 memory:1 video:1 critical:1 suitable:1 difficulty:1 force:1 hybrid:1 overlap:1 buhmann:2 eye:1 specializes:1 x8:1 review:1 determining:1 limitation:1 filtering:7 allocation:1 localized:1 degree:1 editor:1 classifying:3 soon:1 bias:1 face:78 feedback:2 forward:3 clue:1 san:1 preprocessing:2 projected:1 adaptive:1 unnecessary:1 grayscale:3 craw:2 ca:2 schyns:3 investigated:3 complex:2 mse:3 cl:1 did:1 mahwah:1 allowed:1 complementary:1 referred:2 parker:4 garrison:1 sub:2 position:1 momentum:1 portrayed:1 perceptual:1 behrmann:2 wavelet:2 learns:1 removing:1 down:1 gating:9 agnosia:7 evidence:6 illustrates:1 occurring:1 simply:4 visual:11 labor:2 contained:2 partially:1 gender:1 gary:1 relies:1 extracted:2 identity:1 appreciable:1 determined:1 principal:2 degradation:1 pas:15 tendency:1 la:1 experimental:2 rarely:1 metcalfe:2 support:2 moscovitch:2 arises:1 tested:2 |
438 | 14 | 301
ENCODING GEOMETRIC INVARIANCES IN
HIGHER-ORDER NEURAL NETWORKS
C.L. Giles
Air Force Office of Scientific Research, Bolling AFB, DC 20332
R.D. Griffin
Naval Research Laboratory, Washington, DC
20375-5000
T. Maxwell
Sachs-Freeman Associates, Landover, MD 20785
ABSTRACT
We describe a method of constructing higher-order neural
networks that respond invariantly under geometric transformations on
the input space. By requiring each unit to satisfy a set of
constraints on the interconnection weights, a particular structure is
imposed on the network. A network built using such an architecture
maintains its invariant performance independent of the values the
weights assume, of the learning rules used, and of the form of the
nonlinearities in the network. The invariance exhibited by a firstorder network is usually of a trivial sort, e.g., responding only to
the average input in the case of translation invariance, whereas
higher-order networks can perform useful functions and still exhibit
the invariance. We derive the weight constraints for translation,
rotation, scale, and several combinations of these transformations,
and report results of simulation studies.
INTRODUCTION
A persistent difficulty for pattern recognition systems is the
requirement that patterns or objects be recognized independent of
irrelevant parameters or distortions such as orientation (position,
rotation, aspect), scale or size, background or context, doppler
shift, time of occurrence, or signal duration. The remarkable
performance of humans and other animals on this problem in the visual
and auditory realms is often taken for granted, until one tries to
build a machine with similar performance. Thoufh many methods have
been developed for dealing with these problems, we have classified
them into two categories: 1) preprocessing or transformation
(inherent) approaches, and 2) case-specific or "brute force"
(learned) approaches. Common transformation techniques include:
Fourier, Hough, and related transforms; moments; and Fourier
descriptors of the input signal. In these approaches the signal is
usually transformed so that the subsequent processing ignores
arbitrary parameters such as scale, translation, etc. In addition,
these techniques are usually computationally expensive and are
sensitive to noise in the input signal. The "brute force" approach
is exemplified by training a device, such as a perceptron, to
classify a pattern independent of it's position by presenting the
@
American Institute of Physics 1988
302
training pattern at all possible positions. MADALINE machines 2 have
been shown to perform well using such techniques. Often, this type
of invariance is pattern specific, does not easily generalize to
other patterns, and depends on the type of learning algorithm
employed. Furthermore, a great deal of time and energy is spent on
learning the invariance, rather than on learning the signal. We
describe a method that has the advantage of inherent invariance but
uses a higher-order neural network approach that must learn only the
desired signal. Higher-order units have been shown to have unique
computational strengths and are quite amenable to the encoding of a
priori know1edge. 3 - 7
MATHEMATICAL DEVELOPMENT
Our approach is similar to the group invariance approach,8,10
although we make no appeal to group theory to obtain our results. We
begin by selecting a transformation on the input space, then require
the output of the unit to be invariant to the transformation. The
resulting equations yield constraints on the interconnection weights,
and thus imply a particular form or structure for the network
architecture.
For the i-th unit Yi of order M defined on a discrete input
space, let the output be given by
Yi[YiM(X),P(x)] - f( WiO + ~ Wi 1 (X1) P(x1)
+ ~~ Wi 2 (X1,X2) P(x1) P(x2) + ...
+~ ... ~ Wi M(X1,? ?XM) P(x1)? ?P(XM) ),
(1)
where p(x) is the input pattern or signal function (sometimes called
a pixel) evaluated at position vector x, wim(xl, ... Xm) is the weight
of order m connecting the outputs of units at Xl, x2, .. Xm to the ith unit, i.e., it correlates m values, f(u) is some threshold or
sigmoid output function, and the summations extend over the input
space. YiM(X) represents the entire set of weights associated with
the i-th unit. These units are equivalent to the sigma-pi units a
defined by Rumelhart, Hinton, and Williams. 7 Systems built from
these units suffer from a combinatorial explosion of terms, hence are
more complicated to build and train. To reduce the severity of this
problem, one can limit the range of the interconnection weights or
the number of orders, or impose various other constraints. We find
that, in addition to the advantages of inherent invariance, imposing
an invariance constraint on Eq. (1) reduces the number of allowed
aThe sigma-pi neural networks are multi-layer networks with
higher-order terms in any layer. As such, most of the neural
networks described here can be considered as a special case of the
sigma-pi units. However, the sigma-pi units as originally formulated
did not have invariant weight terms, though it is quite simple to
incorporate such invariances in these units.
303
weights, thus simplifying the architecture and shortening the
training time.
We now define what we mean by invariance. The output of a unit
is invariant with respect to the transformation T on the input
pattern if 9
(2)
An example of the class of invariant response defined by Eq. (2)
would be invariant detection of an object in the receptive field of a
panning or zooming camera. An example of a different class would be
invariant detection of an object that is moving within the field of a
fixed camera. One can think of this latter case as consisting of a
fixed field of "noise" plus a moving field that contains only the
object of interest. If the detection system does not respond to the
fixed field, then this latter case is included in Eq. (2).
To illustrate our method we derive the weight constraints for
one-dimensional translation invariance. We will first switch to a
continuous formulation, however, for reasons of simplicity and
generality, and because it is easier to grasp the physical
significance of the results, although any numerical simulation
requires a discrete formulation and has significant implications for
the implementation of our results. Instead of an index i, we now
keep track of our units with the continuous variable u. With these
changes Eq. (2) now becomes
y[u;wM(x),p(X)] = f( wO + JrdXl Wl(U;Xl) P(xl) + ...
+
f?? Jr dXl? .dXM
wM(U;Xl,? ?XM) P(Xl)? .P(XM) ),
(3)
The limits on the integrals are defined by the problem and are
crucial in what follows. Let T be a translation of the input pattern
by -xO, so that
(4)
T[p(x)] - p(x+XO)
where xo is the translation of the input pattern.
Ty[u;wM(x) ,p(x)] - y[u;YM(x),p(x+XO?)
=
Then, from eq (2),
y[u;wM(x),p(x)]
(5)
Since p(x) is arbitrary we must impose term-by-term equality in the
argument of the threshold function; i.e.,
f dXl Wl(U;Xl)
P(xl) =
f dxl Wl(U;Xl)
P(xl+XO),
(Sa)
Jr fdxl dX2 W2 (U;Xl,X2) P(xl) P(x2) =
Jr
f dXl
dX2 W2 (U;Xl,X2) P(xl+XO) P(x2+XO),
etc.
(Sb)
304
Making the substitutions xl. xl-XO, x2 ",x2-XO, etc, we find that
f dXl Wl(U;Xl)
f f dxl
P(xl) -
f dxl WI(U;Xl-XO)
P(XI) ,
(6a)
dX2 W2 (U;XI,X2) P(xI) P(x2) -
f f dXI
dX2 W2 (U;XI-XO,X2-XO) P(xI) P(x2),
(6b)
etc.
Note that the limits of the integrals on the right hand side must be
adjusted to satisfy the change-of-variables. If the limits on the
integrals are infinite or if one imposes some sort of periodic
boundary condition, the limits of the integrals on both sides of the
equation can be set equal. We will assume in the remainder of this
paper that these conditions can be met; normally this means the
limits of the integrals extend to infinity. (In an implementation,
it is usually impractical or even impossible to satisfy these
requirements, but our simulation results indicate that these networks
perform satisfactorily even though the regions of integration are not
identical. This question must be addressed for each class of
transformation; it is an integral part of the implementation design.)
Since the functions p(x) are arbitrary and the regions of integration
are the same, the weight functions must be equal. This imposes a
constraint on the functional form of the weight functions or, in the
discrete implementation, limits the allowed connections and thus the
number of weights. In the case of translation invariance, the
constraint on the functional form of the weight functions requires
that
w1(U;XI) - wl(u;X].-XO),
w2(U;XI,X2) - w2(U;XI-XO,X2-XO),
(7a)
(7b)
etc.
These equations imply that the first order weight is independent of
input position, and depends only on the output position u. The
second order weight is a function only of vector differences,IO i.e.,
w1(u;Xj) - J..(u),
(8a)
w2(U;X].,X2) - w2(u:X]. -Xl)?
(8b)
For a discrete implementation with N input units (pixels) fully
connected to an output unit, this requirement reduces the number of
second-order weights from order N2 to order N, i.e., only weights for
differences of indexes are needed rather than all unique pair
combinations. Of course, this advantage is multiplied as the number
of fully-connected output units increases.
FURTHER EXAMPLES
We have applied these techniques to several other
transformations of interest. For the case of transformation of scale
305
define the scale operator S such that
Sp(x) - aIlp(ax)
(9)
where a is the scale factor, and x is a vector of dimension n. The
factor an is used for normalization purposes, so that a given figure
always contains the same "energy" regardless of its scale.
Application of the same procedure to this transformation leads to the
following constraints on the weights:
wl(u;Xjfa) -= wl(u;~,
w2(u;X1Ia,xv'a) .. w2(u;'X.l.'~)'
w3(u;xlla,x2/a ,x3/a) ... w3(U;X].,X2,X3), etc.
(lOa)
(lOb)
(lOc)
Consider a two-dimensional problem viewed in polar coordinates (r,t).
A set of solutions to these constraints is
J.(u;q,tI) - w1(u;Q),
w2(u;rl,r2;tl,t2) - w2(u;rllr2;tl,t2).
w3 (u;rl,r2,r3;tl,t2,t3) - w3 (u;(rl-r2)/r3;tl,t2,t3).
(lla)
(llb)
(llc)
Note that with increasing order comes increasing freedom in the
selection of the functional form of the weights. Any solution that
satisfies the constraint may be used. This gives the designer
additional freedom to limit the connection complexity, or to encode
special behavior into the net architecture. An example of this is
given later when we discuss combining translation and scale
invariance in the same network.
Now consider a change of scale for a two-dimensional system in
rectangular coordinates, and consider only the second-order weights.
A set of solutions to the weight constraint is:
W2 (U;Xl,Yl;X2,Y2)
-
W2 (U;Xl/Yl;X2/Y2),
- W2 (U;Xl/X2;Yl/Y2),
W2 (U;Xl,Yl;X2,Y2) - w2 (U;(Xl-X2)/(Yl-Y2)), etc.
W2 (U;Xl,Yl;X2,Y2)
(12a)
(l2b)
(12c)
We have done a simulation using the form of Eq. (12b). The
simulation was done using a small input space (8x8) and one output
unit. A simple least-mean-square (back-propagation) algorithm was
used for training the network. When taught to distinguish the
letters T and C at one scale, it distinguished them at changes of
scale of up to 4X with about 15 percent maximum degradation in the
output strength. These results are quite encouraging because no
special effort was required to make the system work, and no
corrections or modifications were made to account for the boundary
condition requirements as discussed near Eq. (6). This and other
simulations are discussed further later.
As a third example of a geometric transformation, consider the
case of rotation about the origin for a two-dimensional space in
polar coordinates. One can readily show that the weight constraints
306
are satisfied if
wl(u;rl,tl) ~ wl(u;rl),
w2(u;rl,r2;tl,t2) - w2(u;rl,r2;tl-t2), etc.
(13a)
(l3b)
These results are reminiscent of the results for translation
invariance. This is not uncommon: seemingly different problems
often have similar constraint requirements if the proper change of
variable is made. This can be used to advantage when implementing
such networks but we will not discuss it further here.
An interesting case arises when one considers combinations of
invariances, e.g., scale and translation. This raises the question
of the effect of the order of the transformations, i.e., is scale
followed by translation equivalent to translation followed by scale?
The obvious answer is no, yet for certain cases the order is
unimportant. Consider first the case of change-of-scale by a,
followed by a translation XC; the constraints on the weights up to
second order are:
Wl(U;Xl) - wl(u; (xl-xo)/a),
w2 (u; Xl ,x2)
0=
w2(u; (xl-xo)/a, (x2-xo)/a) ,
(14a)
(l4b)
and for translation followed by scale the constraints are:
wl(u;Xl) - wl(u; (xl/a)-xo). and
(lSa)
w2(U;Xl,X2) = w2(u;(xl/a)-xo,(x2Ia )-XO) .
(lSb)
Consider only the second-order weights for the two-dimensional case.
Choose rectangular coordinate variables (x,y) so that the translation
is given by (xO,YO). Then
W2 (U;Xl,Yl;X2,Y2) =
w2 (u;(xl/a)-xO,(Yl/a)-YO;(x2/a)-xO'(Y2/a)-yO)'
(l6a)
W2 (U;Xl,Yl;X2,Y2) w2 (U;(Xl- x o)/a, (Yl-yo)/a; (x2- xo)/a, (Y2-Yo)/a).
(16b)
or
If we take as our solution
w2(U;Xl,Yl;X2,Y2) = w2(U;(X1-X2)/(Yl-Y2?,
(17)
then w2 is invariant to scale and translation, and the order is
unimportant. With higher-order weights one can be even more
adventurous.
As a final example consider the case of a change of scale by a
factor a and rotation about the origin by an amount to for a twodimensional system in polar coordinates. (Note that the order of
transformation makes no difference.) The weight constraints up to
second order are:
(18a)
307
(18b)
The first-order constraint requires that wI be independent of the
input variables, but for the second-order term one can obtain a more
useful solution:
(19)
This implies that with second-order weights, one can construct a unit
that is insensitive to changes in scale and rotation of the input
space. How useful it is depends upon the application.
SIMULATION RESULTS
We have constructed several higher-order neural networks that
demonstrated invariant response to transformations of scale and of
translation of the input patterns. The systems were small,
consisting of less than 100 input units, were constructed from
second-and first-order units, and contained only one, two, or three
layers. We used a back-propagation algorithm modified for the
higher-order (sigma-pi) units. The simulation studies are still in
the early stages, so the performance of the networks has not been
thoroughly investigated. It seems safe to say, however, that there
is much to be gained by a thorough study of these systems. For
example, we have demonstrated that a small system of second-order
units trained to distinguish the letters T and C at one scale can
continue to distinguish them over changes in scale of factors of at
least four without retraining and with satisfactory performance.
Similar performance has been obtained for the case of translation
invariance.
Even at this stage, some interesting facets of this approach are
becoming clear: 1) Even with the constraints imposed by the
invariance, it is usually necessary to limit the range of connections
in order to restrict the complexity of the network. This is often
cited as a problem with higher-order networks, but we take the view
that one can learn a great deal more about the nature of a problem by
examining it at this level rather than by simply training a network
that has a general-purpose architecture. 2) The higher-order
networks seem to solve problems in an elegant and simple manner.
However, unless one is careful in the design of the network, it
performs worse than a simpler conventional network when there is
noise in the input field. 3) Learning is often "quicker" than in a
conventional approach, although this is highly dependent on the
specific problem and implementation design. It seems that a tradeoff
can be made: either faster learning but less noise robustness, or
slower learning with more robust performance.
DISCUSSION
We have shown a simple way to encode geometric invariances into
neural networks (instead of training them), though to be useful the
networks must be constructed of higher-order units. The invariant
encoding is achieved by restricting the allowable network
308
architectures and is independent of learning rules and the form of
the sigmoid or threshold functions. The invariance encoding is
normally for an entire layer, although it can be on an individual
unit basis. It is easy to build one or more invariant layers into a
multi-layer net, and different layers can satisfy different
invariance requirements. This is useful for operating on internal
features or representations in an invariant manner. For learning in
such a net, a multi-layered learning rule such as generalized backpropagation 7 must be used. In our simulations we have used a
generalized back-propagation learning rule to train a two-layer
system consisting of a second-order, translation-invariant input
layer and a first-order output layer. Note that we have not shown
that one can not encode invariances into layered first-order
networks, but the analysis in this paper implies that such invariance
would be dependent on the form of the sigmoid function.
When invariances are encoded into higher-order neural networks,
the number of interconnections required is usually reduced by orders
of powers of N where N is the size of the input. For example, a
fully connected, first-order, single-layer net with a single output
unit would have order N interconnections; a similar second-order net,
order N2 . If this second-order net (or layer) is made shift
invariant, the order is reduced to N. The number of multiplies and
adds is still of order N2 .
We have limited our discussion in this paper to geometric
invariances, but there seems to be no reason why temporal or other
invariances could not be encoded in a similar manner.
REFERENCES
1.
D.H. Ballard and C.M. Brown, Computer Vision (Prentice-Hall,
Englewood Cliffs, NJ, 1982).
2.
B. Widrow, IEEE First Int1. Conf. on Neural Networks, 87TH019l7, Vol. 1, p. 143, San Diego, CA, June 1987.
3.
J.A. Feldman, Biological Cybernetics 46, 27 (1982).
4.
C.L. Giles and T. Maxwell, App1. Optics 26, 4972 (1987).
5.
G.E. Hinton, Proc. 7th IntI. Joint Conf. on Artificial
Intelligence, ed. A. Drina, 683 (1981).
6.
Y.C. Lee, G. Doolen, H.H. Chen, G.Z. Sun, T. Maxwell, H.Y. Lee,
C.L. Giles, Physica 22D, 276 (1986).
7.
D.E. Rume1hart, G.E. Hinton, and R.J. Williams, Parallel
Distributed Processing, Vol. 1, Ch. 8, D.E. Rume1hart and J.L.
McClelland, eds., (MIT Press, Cambridge, 1986).
309
8.
T . Maxwell, C.L. Giles, Y.C. Lee, and H.H. Chen, Proc. IEEE
IntI. Conf. on Systems, Man, and Cybernetics, 86CH2364-8, p.
627, Atlanta, GA, October 1986.
9.
W. Pitts and W.S. McCulloch, Bull. Math. Biophys. 9, 127
(1947).
10.
M. Minsky and S, Papert, Perceptrons (MIT Press, Cambridge,
Mass., 1969).
| 14 |@word seems:3 retraining:1 simulation:9 simplifying:1 moment:1 substitution:1 contains:2 loc:1 selecting:1 yet:1 must:7 readily:1 reminiscent:1 numerical:1 subsequent:1 intelligence:1 device:1 ith:1 math:1 simpler:1 mathematical:1 constructed:3 persistent:1 manner:3 behavior:1 multi:3 freeman:1 encouraging:1 increasing:2 becomes:1 begin:1 mass:1 mcculloch:1 what:2 developed:1 transformation:15 impractical:1 nj:1 temporal:1 thorough:1 firstorder:1 ti:1 brute:2 unit:27 normally:2 lsa:1 xv:1 limit:9 io:1 encoding:4 cliff:1 becoming:1 plus:1 limited:1 range:2 unique:2 camera:2 satisfactorily:1 x3:2 backpropagation:1 procedure:1 lla:1 ga:1 selection:1 operator:1 layered:2 prentice:1 context:1 impossible:1 twodimensional:1 equivalent:2 imposed:2 demonstrated:2 conventional:2 williams:2 regardless:1 duration:1 rectangular:2 simplicity:1 rule:4 doolen:1 coordinate:5 diego:1 us:1 origin:2 associate:1 rumelhart:1 recognition:1 expensive:1 quicker:1 region:2 connected:3 sun:1 complexity:2 trained:1 raise:1 upon:1 basis:1 easily:1 joint:1 various:1 train:2 describe:2 artificial:1 quite:3 encoded:2 solve:1 distortion:1 say:1 interconnection:5 think:1 final:1 seemingly:1 advantage:4 net:6 remainder:1 combining:1 requirement:6 object:4 spent:1 derive:2 illustrate:1 widrow:1 sa:1 eq:7 indicate:1 come:1 met:1 implies:2 safe:1 human:1 implementing:1 require:1 biological:1 summation:1 adjusted:1 correction:1 physica:1 considered:1 hall:1 great:2 pitt:1 early:1 dxl:7 purpose:2 polar:3 proc:2 combinatorial:1 wim:1 sensitive:1 wl:13 mit:2 always:1 modified:1 rather:3 office:1 encode:3 ax:1 yo:5 june:1 naval:1 dependent:2 sb:1 entire:2 transformed:1 pixel:2 orientation:1 priori:1 multiplies:1 development:1 animal:1 special:3 integration:2 field:6 equal:2 construct:1 washington:1 identical:1 represents:1 report:1 t2:6 inherent:3 individual:1 minsky:1 consisting:3 freedom:2 detection:3 atlanta:1 interest:2 englewood:1 highly:1 grasp:1 uncommon:1 amenable:1 implication:1 integral:6 explosion:1 necessary:1 unless:1 hough:1 desired:1 classify:1 giles:4 facet:1 bull:1 examining:1 answer:1 periodic:1 thoroughly:1 cited:1 lee:3 physic:1 yl:12 connecting:1 ym:1 w1:3 satisfied:1 choose:1 worse:1 conf:3 american:1 account:1 nonlinearities:1 satisfy:4 depends:3 later:2 try:1 view:1 wm:4 sort:2 maintains:1 complicated:1 parallel:1 air:1 square:1 descriptor:1 yield:1 t3:2 generalize:1 cybernetics:2 classified:1 ed:2 ty:1 energy:2 obvious:1 associated:1 dxi:1 auditory:1 realm:1 back:3 maxwell:4 higher:13 originally:1 afb:1 response:2 formulation:2 evaluated:1 though:3 done:2 generality:1 furthermore:1 stage:2 until:1 hand:1 propagation:3 scientific:1 effect:1 requiring:1 y2:12 brown:1 hence:1 equality:1 laboratory:1 satisfactory:1 dx2:4 deal:2 lob:1 generalized:2 allowable:1 presenting:1 performs:1 percent:1 common:1 rotation:5 sigmoid:3 functional:3 physical:1 rl:7 insensitive:1 extend:2 discussed:2 significant:1 cambridge:2 imposing:1 feldman:1 moving:2 operating:1 etc:8 add:1 irrelevant:1 certain:1 continue:1 yi:2 additional:1 impose:2 employed:1 recognized:1 signal:7 reduces:2 faster:1 vision:1 sometimes:1 normalization:1 llb:1 achieved:1 whereas:1 background:1 addition:2 addressed:1 lsb:1 crucial:1 w2:31 exhibited:1 elegant:1 seem:1 near:1 easy:1 switch:1 xj:1 w3:4 architecture:6 restrict:1 reduce:1 tradeoff:1 shift:2 granted:1 panning:1 effort:1 wo:1 suffer:1 useful:5 clear:1 unimportant:2 transforms:1 shortening:1 amount:1 category:1 mcclelland:1 reduced:2 designer:1 track:1 discrete:4 vol:2 taught:1 group:2 four:1 threshold:3 letter:2 respond:2 griffin:1 layer:12 followed:4 distinguish:3 strength:2 optic:1 constraint:19 athe:1 infinity:1 x2:33 aspect:1 fourier:2 argument:1 combination:3 jr:3 wi:5 making:1 modification:1 invariant:14 xo:25 inti:2 taken:1 computationally:1 equation:3 discus:2 r3:2 needed:1 multiplied:1 occurrence:1 yim:2 distinguished:1 robustness:1 slower:1 responding:1 include:1 xc:1 build:3 question:2 receptive:1 md:1 exhibit:1 zooming:1 considers:1 trivial:1 reason:2 index:2 madaline:1 october:1 sigma:5 implementation:6 design:3 proper:1 perform:3 hinton:3 severity:1 dc:2 arbitrary:3 bolling:1 pair:1 required:2 doppler:1 connection:3 learned:1 usually:6 pattern:11 exemplified:1 xm:6 built:2 power:1 difficulty:1 force:3 imply:2 x8:1 geometric:5 fully:3 interesting:2 remarkable:1 rume1hart:2 imposes:2 pi:5 translation:19 course:1 loa:1 side:2 perceptron:1 institute:1 distributed:1 boundary:2 dimension:1 llc:1 ignores:1 made:4 preprocessing:1 san:1 correlate:1 keep:1 dealing:1 xi:8 continuous:2 why:1 learn:2 nature:1 robust:1 ballard:1 ca:1 investigated:1 adventurous:1 constructing:1 did:1 significance:1 sachs:1 sp:1 noise:4 n2:3 allowed:2 x1:7 tl:7 papert:1 position:6 xl:39 third:1 specific:3 appeal:1 r2:5 restricting:1 gained:1 biophys:1 chen:2 easier:1 simply:1 visual:1 contained:1 ch:1 satisfies:1 viewed:1 formulated:1 dxm:1 invariantly:1 careful:1 man:1 change:9 included:1 infinite:1 degradation:1 called:1 invariance:27 wio:1 perceptrons:1 internal:1 latter:2 arises:1 incorporate:1 |
439 | 140 | 20
ASSOCIATIVE LEARNING
VIA INHIBITORY SEARCH
David H. Ackley
Bell Communications Research
Cognitive Science Research Group
ABSTRACT
ALVIS is a reinforcement-based connectionist architecture that
learns associative maps in continuous multidimensional environments. The discovered locations of positive and negative reinforcements are recorded in "do be" and "don't be" subnetworks,
respectively. The outputs of the subnetworks relevant to the current goal are combined and compared with the current location to
produce an error vector. This vector is backpropagated through
a motor-perceptual mapping network. to produce an action vector that leads the system towards do-be locations and away from
don 't-be locations. AL VIS is demonstrated with a simulated robot
posed a target-seeking task.
INTRODUCTION
The "backpropagation algorithm" or generalized delta rule (Rumelhart, Hinton, &
Williams, 1986) is sometimes criticized on the grounds that it is a "supervised"
learning algorithm, which requires a "teacher" to provide correct outputs, and
apparently leaves open the question of how the teacher learned the right answers.
However, work. by Rumelhart (personal communication, 1987) and Miyata (1988)
has shown how the environment that a system is embedded in can serve as the
"teacher." If, as in this paper, a backpropagation network is posed the task of
mapping from a vector 9 of robot arm joint angles to the resulting vector X of
arm coordinates in space (the "forward kinematics problem"), then input-output
training data can be obtained by supplying sets of joint angles to the arm and
observing the resulting configurations.
Although this "environment as teacher" strategy shows how a "teacher" can come to
possess useful information without an infinite regress learning it, it is not a complete
solution. There are problems for which the "laws of physics" of an environment
do not suffice to determine the solution. Suppose, for example, that a robot is
posed the problem of learning to reach for different positions in space depending
on which of a set of signals is currently presented, and that the only feedback
available from the environment is success or failure information about the current
arm configuration.
Associative Learning via Inhibitory Search
d
Figure 1. A "trunk" robot.
What is needed in such a case is a mechanism to search through the space of possible arm configurations, recording the successful configurations associated with the
various inputs. ALVIS - Associative Learning Via Inhibitory Search - provides
one such mechanism. The next section applies backpropagation to the 9 --+ X mapping and shows how the resulting network can sometimes be used to solve X --+ 9
problems. The third section, "Self-supervision and inhibitory search," integrates
that network into the overall ALVIS algorithm. The final section contains some
discussion and conclusions. An expanded version of this paper may be found in
Ackley (1988).
FORWARD AND INVERSE KINEMATICS
The in verse kinematics problem in controlling an arm is the problem of determining
what joint angles are needed to produce a specific position and orientation of a hand.
In the general case it is a difficult problem. An itch on your back suggests the kinds
of questions that arise. Which hand should you use? Should you go up from around
your waist, or down from over your shoulders? Can you be sure you know what
will work without actually trying it?
From a computational standpoint, forward kinematics - deciding where your limbs
will end up given a set of joint angles - is an easier problem.
Figure 1 depicts the planar "robot" that was used in this work. I call it the "trunk"
robot. (The work discussed in Ackley (1988) also used a two-handed "pincer"
robot .) Of course, the trunk is a far cry from a real robot, and the only significant
constraint is that the possible joint angles are limited, but this suffices to pose
non-trivial kinematics problems. The trunk has five joints, and each joint angle is
limited to a range of 4 0 to 176 0 with respect to the previous limb.
I simulated a backpropagation network with five real-valued input units (the joint
angles), sixty hidden units in a single layer, and twelve linear output units (the
Cartesian joint positions). Joint angles were expressed in radians, so the range of
input unit values was from about 0.07 to about 3.07. The configuration vector X
was represented by twelve output units corresponding to six pairs of (x, y) coordi-
21
22
Ackley
Figure 2. A "logarithmic strobe" display of the trunk's asymptotic convergence on
a specified position and orientation. The arm position is displayed after iterations
1,2,4,8, ... ,256.
nates, one pair for each joint a through f. With the trunk robot (though not with
the "pincer") f:e and fy have constant values, since they end up being part of the
"anchor." The state of an output unit equals the sum of its inputs, and the error
propagated out of an output unit equals the error propagated into it.
Errors were defined by the difference between the predicted configuration and the
actual configuration, and after extensive training on the trunk robot forward kinematics problem, the network achieved high accuracy over most of the joint ranges.
In typical backpropagation applications, once the desired mapping has been learned,
the backward "error channels" in the network are no longer used. However, suppose
some other error computation, different from that used to train the weights, was
then incorporated. Those errors can be propagated from the outputs of the trained
network all the way back to the inputs. The goal is no longer to change weights
in the network - since they already represent a useful mapping - but to use the
trained network to translate output-space errors, however defined, into errors at
the inputs to the network.
Figure 2 illustrates one use of this process, showing how a trained forward kinematics network can be used to perform a cheap kind of inverse kinematics. The
figure shows superimposed outputs of the trained network under the influence of
a task-specific error computationj in this case the trunk is trying to reduce the
distances between the front and back of a "target arrow" and the front and back
of its first arm section. The target arrow is defined by a head (h:e, hy) and a tail
(t:e, ty). The errors for output units a:e and ay are defined by e(a:e} = h:e - a:e and
e(ay ) hy-ay , the errors for output units b:e and by are defined by e(b:e) t:e-b:e
and e(by ) = ty - by, and the errors at all other outputs are set to zero.
=
=
The algorithm used to generate this behavior has the following steps:
1. Compute errors for one or more output units based on the current positions
of the joints and the desired positioning and orienting information. If "close
Associative Learning via Inhibitory Search
o
Figure 3. The trunk kinking itself.
enough" to the target, exit, otherwise, store these errors on the selected
output units, and set the other error terms to zero.
2. Backpropagate the errors all the way through the network to the input
units. This produces an error term e(9i) for each joint angle 9i. Produce
new joint angles: 9~ = 9i + ke(9i) where k is a scaling constant. Clip the
joint angles against their minimum and maximum values.
3. Forward propagate through the network based on the new joint angles, to
produce new current positions for the joints. Go to step 1.
Whereas the training phase has forward propagation of activations (states) followed
by back propagation of errors, this usage reverses the order. Backpropagation of
errors is followed by changes in inputs followed by forward propagation of activations. This is a general gradient descent technique usable when a backpropagation
network can learn to map from a control space 9 to an error or evaluation space X.
Figure 3 illustrates how gradient descent's familiar limitation can manifest itself:
The target is reachable but the robot fails to reach it. The initial configuration
was such that while approaching the target, the trunk kinked itself too short to
reach. If the robot had "thought" to open 93 instead of closing it, it could have
succeeded. In that sense, the problem arises because the error computation only
specified errors for the tip of the trunk, and not for the rest of the arm. If, instead,
there were indications where all of the joints were to be placed, failures due to local
minima could be greatly reduced.
SELF-SUPERVISION AND INHIBITORY SEARCH
The feedback control network of the previous section locally minimizes joint position
errors - however they are generated - by translating them into joint angle space
and moving downhill. AL VIS uses the feedback control network for arm control;
this section shows how ALVIS learns to generate appropriate joint position space
errors given only a reinforcement signal. There are two key points. The first is
this: Once an action producing a positive reinforcement has somehow been found,
23
24
Ackley
the problem reduces to associative mapping between the input and the discovered
correct output. In ALVIS, "do-be units" are used to record such successes. The
second point is this: When negative reinforcement occurs, the current configuration
can be associated with the input in a behavior-reversed fashion - as a place to
avoid in the future. In ALVIS, "don't-be units" are used to record such failures.
The overall idea, then, is to perform inhibitory search by remembering failures as
they occur and avoiding them in the future, and to perform associative learning
by remembering successful configurations as the search process uncovers them and
recreating them in the future. In effect, ALVIS constructs input-dependent "attractors" at arm configurations associated with success and "repellors" at configurations
associated with failure. Figure 4 summarizes the algorithm. A few points to note
are these:
? The do-be and don't-be units use the spherical non-linear function explored by
Burr & Hanson (1987). The response of a spherical unit is maximal and equal to
one when the input vector and the weight vector are identical. The response of
the unit decreases monotonically with the Euclidean distance between the two
vectors, and the radius r governs the rate of decay.
? The don't-be units of each subnetwork (i.e., relevant to one goal) are in a competitive network (see, e.g., Feldman 1982). The don't-be unit with the largest
activation value (which is a function of both the distance from the current position and the radius) is the only don't-be unit that has effects on the rest of the
system. In the simulations reported here, I used m 4 don't-be units per goal.
=
? In addition to the parameters associated with spherical units, each do-be and
don't-be unit has a strength parameter a that specifies how much influence the
unit has over the behavior of the arm. Do-be strength (at) grows logarithmically with positive reinforcement and shrinks linearly with negative; don't-be
strength (a;i) grows logarithmically with negative reinforcement and shrinks linearly with positive.
Figure 5 illustrates a situation from early in a run of the system. From left to right,
the three displays show the state of the relevant do-be unit, the relevant don't-be
subnetwork, and the current configuration of the arm. Since this particular goal
has never been achieved before, the do-be map provides no useful information its weight vector contains small random values (as it happens, the origin is below
and right of the display) and its strength is zero. The display of the don't-be map
shows the positions of all four relevant don't-be units, with the currently selected
don't-be (unit number 3) drawn somewhat darker. The don't-be units are spread
around configuration space, creating "hills" that push away the arm if it comes
too close. As the arm moves about without reaching the target, different don't-be
units win the competition and take control. Negative reinforcement accrues, and
the winning don't-be consequently moves toward the various current configurations
and gets stronger, until the arm is pushed elsewhere.
Figure 6 illustrates the behavior of the system after more extensive learning. The
Associative Learning via Inhibitory Search
Figure 4. AL VIS
O. (Initialize) Given: a space X of h dimensions, a backpropagation network
trained on 8 -+ X, a set G of goals and a mapping from G to regions of
space. Create an AL VIS network with n = IGI goal units 91, ... ,9n, n dobe/don't-be subnetworks consisting of one do-be unit
and m don't-be
units dii, ... , dfu., and h current position units ZI,'" Zh. Create modifiable connections Wzi from z's to d's, Wiz from d's to z's, and a modifiable strength s for each d. Set all do-be strengths st and all don't-be
strengths sti to zero. Set all weights Wzi and Wiz to small random values.
Set 8 to a random legal vector and produce a current configuration X.
1. (New stimulus) Choose t at random from 1 ... n.
dt
2. (Do's/Don'ts) Compute activations for do-be's and don't-be's using the
spherical function:
let
dt:
d=
1/ (1 + ~JL./:=1 (Zi -
wz
il 2 ) .
In subnetwork t,
be the unit with the largest activation.
dt
3. (Errors) Let wt denote the weights from
and w~ denote the weights
from dt:, and similarly for strengths st and st:. Compute errors for each
component of X: e(zi) = si(wt - Zi) + Si:.(Zi - w~).
4. (Move) Backpropagate to produce e(8). Generate angle changes: A8i =
min(q, max( -q, k,.e(8 i ))), with parameters q and k,.. Generate new angles respecting the maximum vt and minimum vi possible joint angles:
8~ = min( vt, max(vi ,8i + A8i )). Forward propagate to produce a new
configuration X'.
5. (Positive reinforcement) Determine whether X' satisfies goal t. If it does
not, go to step 6. Otherwise,
+, - I d wzi
+, -- zi'I
5 . 1 ror 1 - 1, ??? , hIt
, e wiz - zi an
5.2 Let st' = min (5, si + p+ / (1 + si)), for positive reinf p+ > O.
5.3 Let ri = k,/ max(.I, st'), with parameter k,.
5.4 For i = 1, ... , m, let Sti' = max(O, sti-p+), and rti = k,f max(.I, sti'l.
5.5 Go to step 1.
6. (Negative reinforcement) Perform the following:
= w~ + 77(zi - w~) and
= w;i + +
6.1 For i = 1, ... , h, let
77(Z; - w~) with parameter 77, where is a uniform random variable
between ?O.O1.
6.2 Let
= min (5, s;: + p- / (1 + s;:)), for negative reinf p- > O.
6.3 Let r t = k,/ max ( .1, S;:').
6.4 Let st' = max(O, si - p-), and ri = k,/ max(O.I, Si'l.
6.5 Go to step 2.
t;\
? -
wi:
S;:'
e
w;/
e
25
26
Ackley
~~~Be~:~0~.~00~O~OO~0___________ I_D_on_'_t_~~:_3~1~.~29~4~70~3________ I_B~e____________________ 1
Figure 5. A display of the internal state of the trunk robot in the process of
learning to associate a set of twelve arbitrary stimuli with specified positions in
space. The current signal (though the system has not discovered this yet) means
"touch 5".
Do Be: 4.998667
Don't Be: 3 0.007976
~~~~~~----------I~~~~~~---------
Be
Figure 6. A display of the internal state of the same trunk robot later in the
learning process. The previous goal was "touch *" and the current goal is "touch
3."
\ do-be map now contains an accurate image of a successful configuration for the
"touch 3" goal, and its strength is high. The strength of the selected don't-be
unit is low. The current configuration map in Figure 6 shows each iteration of the
algorithm between the time it achieved its previous goal and the time it achieved
the current goal.
Finally, Figure 7 displays the average time-per-goal as a function of the number of
goals achieved. For 75 repetitions, the trunk network was initialized and run until
500 goals had been achieved, and the resulting time-per-goal data was averaged to
produce the graph.
The average time-per-goal declines rapidly as goals are presented, then seems to
rise slightly, and then stabilizes around an average value of about 300. To have
some kind of standard of comparison, albeit unsophisticated, if the joint angles are
simply changed by uniform random values between q and -q (see Figure 4) on each
iteration, the average time-per-goal is observed to be about 490.
Associative Learning via Inhibitory Search
:nOD
24DO
2100
liDO
15DD
1200
900
600
300
Figure 7. A graph of the average time taken per goal as a function of the number
of goals achieved. The horizontal line shows the average performance of random
joint changes.
DISCUSSION
ALVIS is a preliminary, exploratory system. Of course, the ALVIS environment
is but a pale shadow of the real world, but even granting the limited scope of the
problem formulation, several aspects of AL VIS are incompletely satisfying and in
need of improvement.
To my eye, the biggest drawback of the current implementation is the local goal
representation - which essentially requires that the set of goals be enumerable at
network definition time. Related problems include the inability to share information
between one goal and another and the inability to pursue more than one goal
simultaneously. To determine behavior, the constraints of goals must be integrated
with the possibilities of the current situation. In AL VIS this is done as a strictly
two-step process: the goal selects a subnetwork, and the current situation selects
units within the subnetwork. Work such as Jordan (1986) and Miyata (1988) shows
how goal information and context information can be integrated by supplying both
as inputs to a single network.
ALVIS is a pure feedback control model, and can suffer from the traditional problem
of that approach: when the errors are small, the resulting joint angle changes are
small, and the arm converges only slowly. If the gain at the joints is increased
to speed con vergence, overshoot and oscillation become more likely. However, in
ALVIS oscillations gradually die out, as don't-be units shift positions under the
negative reinforcement, and sometimes such temporary oscillations actually help
with the search, causing the tip of the arm to explore a variety of different points.
27
28
Ackley
The aspect of AL VIS behavior that I find most irritating reveals something about
the approach in general. In some cases - usually on more "peripheral" targets
- AL VIS learns to hit the very edge of the target region. While approaching
such targets, ALVIS experiences negative reinforcement, and the don't-be units,
consequently, gain a little strength. The resulting interference occasionally causes
a very long search for a goal that had previously been rapidly achieved. AL VIS
has a representation only for its own body; a better system would also be able to
represent other objects in the world, and useful relations on the expanded set. The
"mostly motor" emphasis evident in the present system needs to be balanced by
more sophistication on the perceptual side.
Though limited in scope, ALVIS demonstrates three ideas I think worth highlightmg:
? The reuse of the error channel of a backpropagation network, after training, for
translating arbitrary output-space gradients into input-space gradients.
? The recording of previous actual outputs to be used as future desired outputs.
? The use of "repellors" (don 't-be units) as well as attractors in defining errors, and
the resulting process of search-by-inhibition generated by negative reinforcement.
Characterizing the behavior of a machine in terms of attractor dynamics is a familiar
notion, but "repellor dynamics" seems to be largely unknown territory. Indeed, in
ALVIS there is an ephemeral quality to the don't-be units: When all answers have
been discovered, all strength accrues to the do-be's, good performances become
routine, and AL VIS behavior is essentially attractor-based. In watching such a
"grown-up" AL VIS, it is easy to forget how it was in the beginning, when the world
was big and answers were scarce, and ALVIS was doing well just to discover a new
mistake.
References
Ackley, D.H. (1988). Associative learning via inhibitory search. Teclmical memorandum TM
ARH-012509. Morristown. NJ: Bell Communications Research.
Burr. D.J .? & Hanson. S.J. (1987). Knowledge representation in connectionist networks. Technical
memorandum TM-ARH-008733. Morristown. NJ: Bell Communications Research.
Feldman. J.A. (1982). Dynamic connections in neural networks. Biological Cybernetics, 36, 193202.
Jordan. M.1. (1986). Serial order: A parallel. distributed processing approach. Technical report
ICS-86M. La Jolla, CA: University of California. Institute for Cognitive Science.
Miyata, Y. (1988). The learning and planning of actions. Unpublished doctoral dissertation in
psychology, University of California San Diego.
Rumelhart, D.E .? Hinton. G.E .? & Williams. R.J. (1986).
propagating errors. Nature, 3Z3. 533-536.
Learning representations by back-
Rumelhart. D.E. (personal communication. 1987). Also cited as personal communication in Miyata (1988).
| 140 |@word version:1 stronger:1 seems:2 a8i:2 open:2 simulation:1 propagate:2 uncovers:1 t_:1 initial:1 configuration:19 contains:3 current:18 activation:5 si:6 yet:1 must:1 cheap:1 motor:2 leaf:1 selected:3 beginning:1 short:1 granting:1 record:2 supplying:2 dissertation:1 provides:2 location:4 five:2 become:2 burr:2 indeed:1 behavior:8 planning:1 spherical:4 actual:2 little:1 discover:1 suffice:1 what:3 kind:3 minimizes:1 pursue:1 nj:2 multidimensional:1 morristown:2 demonstrates:1 hit:2 control:6 unit:40 dfu:1 producing:1 positive:6 before:1 local:2 mistake:1 emphasis:1 doctoral:1 suggests:1 limited:4 range:3 averaged:1 unsophisticated:1 backpropagation:9 bell:3 thought:1 get:1 close:2 context:1 influence:2 map:6 demonstrated:1 williams:2 go:5 ke:1 pure:1 rule:1 exploratory:1 memorandum:2 coordinate:1 notion:1 target:10 suppose:2 controlling:1 diego:1 us:1 origin:1 associate:1 logarithmically:2 rumelhart:4 satisfying:1 observed:1 ackley:8 region:2 decrease:1 balanced:1 environment:6 respecting:1 dynamic:3 personal:3 trained:5 overshoot:1 ror:1 serve:1 exit:1 joint:27 various:2 represented:1 grown:1 train:1 posed:3 solve:1 valued:1 otherwise:2 think:1 itself:3 final:1 associative:10 indication:1 maximal:1 pale:1 wzi:3 relevant:5 causing:1 rapidly:2 translate:1 competition:1 convergence:1 produce:10 converges:1 object:1 help:1 depending:1 itch:1 oo:1 propagating:1 pose:1 predicted:1 nod:1 come:2 revers:1 shadow:1 radius:2 drawback:1 correct:2 translating:2 dii:1 suffices:1 preliminary:1 biological:1 strictly:1 around:3 ground:1 ic:1 deciding:1 mapping:7 scope:2 stabilizes:1 early:1 integrates:1 currently:2 largest:2 repetition:1 create:2 reaching:1 avoid:1 improvement:1 superimposed:1 greatly:1 sense:1 dependent:1 integrated:2 hidden:1 relation:1 selects:2 overall:2 orientation:2 initialize:1 equal:3 once:2 construct:1 never:1 identical:1 future:4 connectionist:2 stimulus:2 report:1 few:1 simultaneously:1 familiar:2 phase:1 consisting:1 attractor:4 possibility:1 evaluation:1 sixty:1 accurate:1 succeeded:1 edge:1 experience:1 euclidean:1 initialized:1 desired:3 criticized:1 handed:1 increased:1 uniform:2 successful:3 front:2 too:2 reported:1 answer:3 teacher:5 my:1 combined:1 st:6 cited:1 twelve:3 physic:1 tip:2 recorded:1 choose:1 slowly:1 watching:1 cognitive:2 creating:1 usable:1 repellors:2 igi:1 vi:13 later:1 apparently:1 observing:1 doing:1 competitive:1 parallel:1 il:1 accuracy:1 largely:1 waist:1 territory:1 worth:1 cybernetics:1 reach:3 wiz:3 definition:1 failure:5 verse:1 ty:2 against:1 regress:1 associated:5 con:1 radian:1 propagated:3 gain:2 manifest:1 knowledge:1 routine:1 actually:2 back:6 dt:4 supervised:1 planar:1 response:2 formulation:1 done:1 though:3 shrink:2 just:1 until:2 hand:2 horizontal:1 touch:4 propagation:3 somehow:1 quality:1 grows:2 orienting:1 usage:1 effect:2 self:2 die:1 generalized:1 trying:2 hill:1 ay:3 complete:1 evident:1 image:1 jl:1 discussed:1 tail:1 significant:1 feldman:2 similarly:1 closing:1 had:3 reachable:1 moving:1 robot:14 supervision:2 longer:2 inhibition:1 something:1 own:1 jolla:1 store:1 occasionally:1 success:3 vt:2 minimum:3 remembering:2 somewhat:1 determine:3 monotonically:1 signal:3 reduces:1 positioning:1 technical:2 rti:1 long:1 serial:1 essentially:2 iteration:3 sometimes:3 represent:2 achieved:8 whereas:1 addition:1 standpoint:1 rest:2 posse:1 sure:1 recording:2 jordan:2 call:1 enough:1 easy:1 variety:1 zi:8 psychology:1 architecture:1 approaching:2 reduce:1 idea:2 decline:1 tm:2 shift:1 enumerable:1 whether:1 six:1 reuse:1 suffer:1 cause:1 action:3 useful:4 governs:1 backpropagated:1 locally:1 clip:1 reduced:1 generate:4 specifies:1 inhibitory:10 delta:1 per:6 modifiable:2 group:1 key:1 four:1 drawn:1 backward:1 graph:2 sum:1 run:2 angle:18 inverse:2 you:4 sti:4 place:1 oscillation:3 summarizes:1 scaling:1 pushed:1 layer:1 followed:3 display:7 strength:12 occur:1 constraint:2 your:4 ri:2 hy:2 aspect:2 speed:1 min:4 expanded:2 coordi:1 peripheral:1 slightly:1 wi:1 happens:1 gradually:1 interference:1 taken:1 legal:1 previously:1 trunk:14 kinematics:8 mechanism:2 needed:2 know:1 end:2 subnetworks:3 available:1 limb:2 away:2 appropriate:1 include:1 reinf:2 seeking:1 move:3 question:2 already:1 occurs:1 strategy:1 traditional:1 subnetwork:5 gradient:4 win:1 distance:3 reversed:1 simulated:2 incompletely:1 fy:1 trivial:1 toward:1 o1:1 z3:1 repellor:1 difficult:1 mostly:1 negative:10 rise:1 implementation:1 unknown:1 perform:4 descent:2 t:1 displayed:1 situation:3 hinton:2 communication:6 shoulder:1 incorporated:1 head:1 discovered:4 defining:1 arbitrary:2 david:1 pair:2 unpublished:1 specified:3 extensive:2 connection:2 hanson:2 california:2 learned:2 temporary:1 able:1 below:1 usually:1 strobe:1 wz:1 max:8 scarce:1 arm:18 eye:1 zh:1 determining:1 asymptotic:1 law:1 embedded:1 recreating:1 limitation:1 dd:1 share:1 cry:1 course:2 elsewhere:1 changed:1 placed:1 side:1 institute:1 characterizing:1 distributed:1 feedback:4 dimension:1 world:3 forward:9 reinforcement:13 san:1 far:1 reveals:1 anchor:1 don:28 search:15 continuous:1 vergence:1 channel:2 learn:1 nature:1 ca:1 miyata:4 spread:1 linearly:2 arrow:2 big:1 arise:1 body:1 biggest:1 depicts:1 fashion:1 darker:1 fails:1 position:14 downhill:1 winning:1 perceptual:2 accrues:2 third:1 learns:3 down:1 specific:2 showing:1 nates:1 explored:1 decay:1 albeit:1 illustrates:4 push:1 cartesian:1 easier:1 backpropagate:2 forget:1 logarithmic:1 sophistication:1 simply:1 likely:1 explore:1 expressed:1 applies:1 satisfies:1 goal:30 consequently:2 towards:1 change:5 infinite:1 typical:1 arh:2 wt:2 la:1 internal:2 arises:1 inability:2 avoiding:1 |
440 | 1,400 | Synchronized Auditory and Cognitive 40 Hz
Attentional Streams, and the Impact of
Rhythmic Expectation on Auditory Scene Analysis
Bill Baird
Dept Mathematics, U.C.Berkeley, Berkeley, Ca. 94720.
baird@math.berkeley.edu
Abstract
We have developed a neural network architecture that implements a theory of attention, learning, and trans-cortical communication based on
adaptive synchronization of 5-15 Hz and 30-80 Hz oscillations between
cortical areas. Here we present a specific higher order cortical model of
attentional networks, rhythmic expectancy, and the interaction of hi~her?
order and primar?, cortical levels of processing. It accounts for the' mismatch negativity' of the auditory ERP and the results of psychological
experiments of Jones showing that auditory stream segregation depends
on the rhythmic structure of inputs. The timing mechanisms of the model
allow us to explain how relative timing information such as the relative
order of events between streams is lost when streams are formed. The
model suggests how the theories of auditory perception and attention of
Jones and Bregman may be reconciled.
1 Introduction
Amplitude patterns of synchronized "gamma band" (30 to 80 Hz) oscillation have been observed in ilie ensemble activity (local field potentials) of vertebrate olfactory, visual, auditory, motor, and somatosensory cortex, and in the retlna, thalamus, hippocampus, reticular
formation, and EMG. Such activity has not only been found in primates, cats, rabbits and
rats, but also insects, slugs, fish, amphibians, reptiles, and birds. This suggests that gamma
oscillation may be as fundamental to neural processing at the network level as action potentials are at the cellular level.
We have shown how oscillatory associative memories may be coupled to recognize and
generate sequential behavior, and how a set of novel mechanisms utilizing these complex
dynamics can be configured to solve attentional and perceptual processing problems. For
pointers to full treatment wi th mathematics and complete references see [BaIrG et al., 1994].
An important element of intra-cortical communication in the brain, and between modules
in this architecture, is the ability of a module to detect and respond to the proper input signal from a particular module, when inputs from other modules which are irrelevant to tlie
present computation are contributing crosstalk noise. We have demonstrated that selective
control of synchronization, which we hypothesize to be a model of "attention", can be used
to solve this coding problem and control program flow in an architecture with dynamic attractors [Baird et al., 1994].
Using dynamical systems theory, the architecture is constructed from recurrently interconnected oscillatory associative memory modules that model higher order sensory and motor
areas of cortex. The modules learn connection weights between themselves which cause the
system to evolve under a 5-20 Hz clocked sensory-motor processing cycle by a sequence
4
B. Baird
of transitions of synchronized 30-80 Hz oscillatory attractors within the modules. The architecture employ's selective"attentional" control of the synchronization of the 30-80 Hz
gamma band OSCIllations between modules to direct the flow of computation to recognize
and generate sequences. The 30-80 Hz attractor amplitude patterns code the information
content of a cortIcal area, whereas phase and frequency are used to "softwire" the network,
since only the synchronized areas communicate by exchanging amplitude information. The
system works like a broadcast network where the unavoidable crosstalk to all areas from previous learned connections is overcome by frequency coding to allow the moment to moment
operation of attentional communication only between selected task-relevant areas.
The behavior of the time traces in different modules of the architecture models the temporary appearance and switching of the synchronization of 5-20 and 30-80 Hz oscillations
between cortical areas that is observed during sensorimotor tasks in monkeys and numans.
The architecture models the 5-20 Hz evoked potentials seen in the EEG as the control signals which determine the sensory-motor processing cycle. The 5-20 Hz clocks which drive
these control signals in the archItecture model thalamic pacemakers which are thought to
control the excitabili ty of neocortical tissue through similar nonspecific biasing currents that
cause the cogni tive and sensory evoked potentials of the EEG. The 5-20 Hz cycles "quantize
time" and form the basis of derived somato-motor rhythms with periods up to seconds that
entrain to each other in motor coordination and to external rhythms in speech perception
[Jones et al., 1981].
1.1
Attentional Streams of Synchronized 40 Hz Activity
There is extensive evidence for the claim of the model that the 30-80 Hz gamma band activity in the brain accomplishes attentional processing, since 40 Hz appears in cortex when
and where attention is required. For example, it is found in somatosensory, motor and premotor cortex of monkeys when they must pick a rasin out of a small box, but not when a
habitual lever press delivers the reward. In human attention experiments, 30-80 Hz activity
goes up in the contralateral auditory areas when subjects are mstructed to pay attention to
one ear and not the other. Gamma activity declines in the dominant hemisphere along with
errors in a learnable target and distractors task, but not when the distractors and target vary
at random on each trial. Anesthesiologists use the absence of 40 Hz activity as a reliable
indicator of unconsciousness. Recent work has shown that cats with convergent and divergent strabismus who fail on tasks where perceptual binding is required also do not exhibit
cortical synchrony. This is evidence that gamma synchronization IS perceptually functional
and not epiphenomenal.
The architecture illustrates the notion that synchronization of gamma band activity not
only"binds" the features of inputs in primary sensory cortex into "objects", but further binds
the activity of an attended object to oscillatory activity in associational and higher-order
sensory and motor cortical areas to create an evolving attentional network of intercommunicating cortical areas that directs behavior. The binding of sequences of attractor transitions
between modules of the architecture by synchronization of their activity models the physiological mechanism for the formation of perceptual and cognitive "streams" investigated by
Bregman [Bregman, 1990], Jones [Jones et aI., 1981], and others. In audition, according to
Bregman's work, successive events of a sound source are bound together into a distinct sequence or "stream" and segy:egated from other sequences so that one pays attention to onl?,
one sound source at a time (the cocktail party problem). Higher order cortical or "cognitive'
streams are in evidence when subjects are unable to recall the relative order of the telling of
events between two stories told in alternating segments.
MEG tomographic observations show large scale rostral to caudal motor-sensory sweeps of
coherent thalamo-cortical40Hz activity accross the entire brain, the phase of which is reset
by sensory input in waking, but not in dream states [Llinas and Ribary, 1993]. This suggests
an inner higher order "attentional stream" is constantly cycling between motor (rostral) and
sensory (caudal) areas in the absence of input. It may be interrupted by input "pop out"
from primary areas or it may reach down as a "searchlight" to synchromze with particular
ensembles of primary activity to be attended.
2 Jones Theory of Dynamic Attention
Jones [Jones et al., 1981] has developed a psychological theory of attention,perception, and
motor timing based on the hypotheSIS that these processes are organized by neural rhythms
in the range of 10 to .5 Hz - the range within which subjects perceive pen odic events as a
rhythm. These rhythms provide a multi scale representation of time and selectively synchronize with the prominant periodiCities of an input to provide a -t emporal expectation mechanism for attention to target particular points in time.
40 Hz Attentional Streams, Rhythmic Expectation, and Auditory Scene Analysis
5
For example, some work suggests that the accented parts of speech create a rhythm to which
listeners entrain. Attention can then be focused on these expected locations as recognition
anchor points for inference of less prominant parts of the speech stream. This is the temporal
analog of the body centered spatial coordinate frame and multiscale covert attention window
system in vision. Here the body centered temporal coordinates of the internal time base
orient by entrainment to the external rhythm, and the window of covert temporal attention
can then select a level of the multiscale temporal coordinates.
In this view, just as two cortical areas must synchronize to communicate, so must two nervous systems. Work using frame by frame film analysis of human verbal interaction, shows
evidence of "interactional synchrony" of gesture and body movement changes and EEG of
both speaker and listener With the onsets of phonemes in speech at the level of a 10 Hz "microrhythm" - the base clock rate of our models. Normal infants synchronize their spontaneous body ftailings at this 10 Hz level to the mothers voice accents, while autistic and
s~hitzophrenic children fail to show interactional synchrony. Autistics are unable to tap in
orne t9 a metronome.
Neural expectation rhythms that support Jones' theory have been found in the auditory EEG.
In experiments where the arrival time of a target stimulus is regular enough to be learned by
an experimental subject, it has been shown that the 10 Hz activity in advance of the stimulus
becomes phase locked to that expected arrival time. This fits our model of rhythmic expectation where the 10Hz rhythm IS a fast base clock that is shifted in phase and frequency
to produce a match in timmg between the stimulus arrival and the output of longer period
cycles derived from this base clock.
2.1
Mismatch Negativity
The "mismatch negativity" (MNN) [Naatanen, 1992] of the auditory evoked potential appears to be an important physiological indicator of the action of a neural expectancy system
like that proposed by Jones. It has been localized to areas within primary auditory cortex by
MEG studies [Naatanen, 1992] and it appears as an increased negativity of the ERP in the
region of the N200 peak whenever a psycbologically discriminable deviation of a repetitive
auditory stimulus occurs. Mismatch IS caused by deviations in onset or offset time, rise time,
frequency, loudness, timbre, phonetic structure, or spatial location of a tone in the sequence.
The mismatch is abolished by blockers of the action ofNMDA channels [Naatanen, 1992]
which are important for the synaptic changes underlying the kind of Hebbian learning which
is used in the model.
MNN is not a direct function of echoic memory because it takes several repetitions for the
expectancy to begin to develop, and it decays 10 2 - 4 seconds. It appears only for repetition periods greater that 50-100 msec and less than 2-4 seconds. Thus the time scale of its
operation is 10 the appropriate range for Jones' expectancy system. Stream formation also
takes several cycles of stimulus repetition to builo up over 2-4 seconds and decays away
within 2-4 seconds in the absence of stimulation. Those auditory stimulus features which
cause streaming are also features which cause mismatch. This supports the hypothesis in
the model that these phenomena are functionally related.
Finally, MNN can occur independent of attention - while a subject is reading or doing a
visual discrimination task. ThIS implies that the auditory system at least must have its own
timing system that can generate timmg and expectancies independent of other behavior. We
can talk or do internal verbal thinking while doing other tasks. A further component of
this negativity appears in prefrontal cortex and is thought by Nataanen to initiate attentional
switchmg toward the deVIant event causing perceptual "pop out" [Naatanen, 1992].
Stream formation is known to affect rhythm perception. The galloping rhythm of high H and
low L tones - HLH-HLH-HLH, for example becomes two separate Isochronous rhythmic
streams of H-H-H-Hand L-L-L-L when the H and L tones are spread far enough apart
[Bregman, 1990]. Evidence for the effect of in'putrhythms on stream formation, however, is
more sparse, and we focus here on the simulatIOn of a particular set of experiments by Jones
[J ones et al., 1981] and Bregman [Bregman, 1990] where this effect has been demonstrated.
2.2 Jones-Bregman Experiment
Jones [Jones et al., 1981] replicated and altered a classic streaming experiment of Bregman
and Rudnicky [Bregman, 1990], and found that their result depended on a specific choice
of the rhythm of presentation. The experiment required human subjects to determine of the
order of presentation of a pair of high target tones AB or BA of slIghtly different frequencies. Also presented before and after the target tones were a series of identical much lower
frequency tones called the capture tones CCC and two identical tones of intermediate fre-
B. Baird
6
quency before and after the target tones called the flanking tones F - CCCFABFCCC. Bregman and Rudnicky found that target order determination performance was best when tfie
capture tones were near to the flanking tones in frequency, and deteriorated as the captor
tones were moved away. Their explanation was that the flanking tones were captured by the
background capture tone stream when close in frequency, leav10g the target tones to stand
out by themselves in the attended stream. When the captor tones were absent or far away in
frequency, the flanking tones were included in the attended stream and obscured the target
tones.
Jones noted that the flanking tones and the capture stream were presented at a stimulus onset rate of one per 240 ms and the targets appeared at 80 ms intervals. In her experiments,
when the captor and flanking tones were given a rhythm in common with the targets, no
effect of the distance of captor and flanking tones appeared. This suggested that rfiythmic
distinction of targets and dlstractors was necessary 10 addition to the frequency diStinction
to allow selective attention to segregate out the target stream. Because performance in the
single rhythm case was worse than that for the control condition without captors, it appeared
that no stream segregation of targets and captors and flanking tones was occurring until the
rhythmic difference was added. From this evidence we malie the assumption in the model
that the distance of a stimulus in time from a rhythmic expectancy acts like the distance between stimuli in pitch, loudness, timbre, or spatial location as/actor for theformation of
separate streams.
3
Architecture and Simulation
To implement Jones's theory in the model and account for her data, subsets of the oscillatory
modules are dedicated to form a rhythmic temporal coordinate frame or time base by dividing down a thalamic 10 Hz base clock rate in steps from 10 to .5 Hz. Each derived clock is
created by an associative memory module that has been specialized to act stereotypically as
a counter or shift register by repeatedly cycling through all its attractors at the rate of one
for each time step of its clock. Its overall cycle time is therefore determined by the number
of attractors. Each cycle is guaranteed to be identical, as required for clocklike function, because of the strong attractors that correct the perturbing effect of noise. Only one step of the
cycle can send output back to primary cortex - the one with the largest weight from receiving the most matco to incoming stimuli. Each clock derived in this manner from a thalamic
base clock will therefore phase reset itself to get the best match to incoming rhythms. The
match can be further refined by frequency and phase entrainment of the base clock itself.
Three such counters are sufficient to model the rhythms in Jones' experiment as shown in
the architecture of figure 1. The three counters divide the 12.5 Hz clock down to 6.25 and
4.16 Hz. The first contains one attractor at the base clock rate which has adapted to entrain
to the 80 msec period of target stimulation (12.5 Hz). The second cycles at 12.5/2 = 6.25
Hz, alternating between two attractors, and the third steps through three attractors, to cycle
at 12.5/3 = 4.16 Hz, which is the slow rhythm of the captor tones.
The modules of the time base send theirinternal30-80 Hz activity to {>rimary auditory cortex
in 100msec bursts at these different rhythmic rates through fast adapting connections (which
would use NMDA channels in the brain) that continually attempt to match incoming stimulus patterns using an incremental Hebbian learning rule. The weights decay .to zero over
2-4 sec to simulate the data on the rise and fall of the mismatch negativity. These weights
effectively compute a low frequency discrete Fourier transform over a sliding window of
several seconds, and the basic periodic structure of rhythmic patterns is quickly matched.
This serves to establish a quantized temporal grid of expectations against which expressive
timing deviations in speech and music can be experienced.
Following Jones [Jones et al., 1981], we hypothesize that this happens automatically as a
constant adaptation to environmental rhythms, as suggested by the mismatch negativity. Retained in these weights of the timebase is a special k10d of short term memory of the activity
which includes temporal information since the timebase will partially regenerate the prevIous activity in primary cortex at the expected recurrence time. This top-down input causes
enchanced sensitivity in target units by increasing their gain. Those patterns which meet
these established rhythmic expectancy signals in time are thereby boosted in amplitude and
pulled into synchrony with the 30-80 Hz attentional searchlight stream to become part of
the attentional network sending input to higher areas. In accordance with Jones' theory,
voluntary top-down attention can probe input at different hierarchical levels of periodicity
by selectively synchronizing a particular cortical column in the time base set to the 40 Hz
frequency of the inner attention stream. Then the searchlight into primary cortex is synchro-
40 Hz Attentional Streams, Rhythmic Expectation, and Auditory Scene Analysis
7
Dynamic Attention Architecture
Higher Order AuditoryCortex
Synchronization
Timebase
10 Hz Clock
w=,
:J'~:'
:
'0"
~
~"
Pitch
.'
:'
~
~ Cycle
..... ~: : :" : :. '---...,.
Attentionallnput Stream
Rhythmic Searchlight
-.-.- . -.-.-.-.-.-.- . -.-.-.-.-.~
High _._._._._._._._._.~
Target Tones
"
-'1
B
~.
?
? -
.
II.
@H
I
_
. -
'
? -
? -
? -
? -
? -
? -
? -
? -
? -
? -
? -
? -
? -
? -~
Fast Weights
:-. 1 '1
_ . ~F I
'_._._._._ . _._._._._._._._._._.~
High Flanking Tones
Low Captor Tones
? -
f
? -
_.-._. C
Input
~
I
I
1':'."",
.:::::::(~~
~::-:-:
I._._._._._._._._ ..._. __
?
I
..............
\
I
__.:-._.
I
- '.
C ._._._._._._ . _.r ' C
._._._._~
' ?- "1
Primary Auditory Cortex
~----------------------------------------------------~Time
Figure 1: Horizontally arrayed units at the top model higher order auditory and motor cortical columns which are sequentially clocked by the (thalamic) base clock on the right to alternate attractor transitions between upper hidden (motor) and lower context (sensory) layers
to act as an Elman net. Three cortical regions are shown - sequence representation memory,
attentional synchronization control, and a rhythmic timebase of three counters. The hidden
and context layers consist of binary "units" composed of two oscillatory attractors. Activity
levels oscillate up and down through the plane of the paper. Dotted hnes show frequency
shifting outputs from the synchromzation (attention) control modules. The lower vertical
set of units IS a sample of primary auditory cortex frequency channels at the values used in
the Jones-Bregman expenment. The dashed lines show the rhythmic pattern of the target,
flanking, and captor tones moving in time from left to right to Impact on auditory cortex.
nizing and reading in activity occuring at the peaks of that particular time base rhythm.
3.1
Cochlear and Primary Cortex Model
At present, we have modeled only the minimal aspects of primary auditory conex sufficient
to qualitatively simulate the Jones-Bregman experiment, but the principles at work allow
expansion to larger scale models with more stimulus features. We simulate four sites in auditory conex corresponding to the four frequencies of stimuli used in the experiment, as
shown in figure 1. There are two close high frequency target tones, one high flanking frequency location, and the low frequency location of the captor stream. These cortical locations are modeled as oscillators with the same equations used for associative memory
modules [Baird et al., 1994], with full linear cross coupling weights. This lateral connectivity is sufficient to promote synchrony among simultaneously activated oscillators, but
insufficient to activate them strongly in the absence of externallOput. This makes full synchrony of activated units the default condition in the model conex, as in Brown's model
[Brown and Cooke, 1996]. so that the background activation is coherent, and can be read
into higher order cortical levels which synchronize with it. The system assumes that all
input is due to the same environmental source in the absence of evidence for segregation
[Bregman, 1990].
Brown and Cooke [Brown and Cooke, 1996] model the cochlear and brain stem nuclear output as a set of overlapping bandpass ("gammatone") filters consistent with auditory nerve
responses and psychophysical "critical bands". A tone can excite several filter outputs at
once. We apprOXImate this effect of the gammatone filters as a lateral fan out of input activations with weights that spread the activation in the same way as the overlapping gammatone
8
B. Baird
filters do.
Experiments show that the intrinsic resonant or "natural" frequencies or "eigenfr~uencies"
of cortical tissue within the 30-80 Hz gamma band vary within individuals on different trials
of a task, and that neurotransmitters can quickly alter these resonant frequencies of neural
clocks. Following the evidence that the oscillation frequency of binding in vision goes up
with the speed of motion of an object, we assume that unattended activity in auditory cortex
synchromzes at a default background fr~uency of 35 Hz, while the higher order attentional
stream is at a higher frequency of 40 Hz. Just as fast motion in vision can cause stimulus
driven capture of attention, we hypothesize that expectancy mismatch in audition causes the
deviant activity to be boosted above the default background frequency to facilitate synchronization with the attentional stream at 40 Hz. This models the mechanism of involuntary
stimulus driven attentional "pop out". Multiple streams of primary cortex activity synchronized at different eigenfrequencies can be selectively attended by unifonnly sweeping the
eigenfrequencies of all primary ensembles through the passband of the 40 Hz higner order
attentional stream to "tune in" each in turn as a radio reciever does.
Following, but modifing the approach of Brown and Cooke [Brown and Cooke, 1996], the
core of our primary cortex stream fonning model is a fast learning rule that reduces the lateral coupling and (in our model) spreads apart the intrinsic cortIcal frequencies of sound
frequency cflannels that do not exhibit the same amplitude of activity at the same time.
This coupling and eigenfrequency difference recovers between onsets. In the absence of lateral synchronizing connectlons or coherent top down driving, synchrony between cortical
streams is rapidly lost because of their distant resonant frequencIes. Activity not satisfying
the Gestalt prinCIple of "common fate" [Bregman, 1990] is thus decorrelated.
The trade off of the effect of temporal and sound frequency proximity on stream segregation follows because close stimulus frequencies excite each other's channel filters. Each
produces a similar output in the other, and their activitites are not decorrelated by coupling
reduction and resonant frequency shifts. On the other hand, to the extent that they are distant enough in sound frequency, each tone onset weakens the weights and shifts the eigenfrequencies of the other channels that are not simultaneously active. This effect is greater,
the faster the presentation rate, because the weight recovery rate is overcome. This recovery
rate can then be adjusted to yield stream segregation at the rates reported by van Noorden
[Bregman, 1990] for given sound frequency separations.
3.2 Sequential Grouping by Coupling and Resonant Frequency Labels
In the absence of rhythmic structure in the input, the temporary weights and resonant frequency "labels" serve as a short tenn "stream memory" to brid~e time (up to 4 seconds) so
that the next nearby input is "captured" or "sequentially bound into the same ensemble of
synchronized activity. This pattern of synchrony in primary cortex has been made into a
temporary attractor by the temporary weight and eigenfrequency changes from the previous stimulation. This explains the single tone capture expenments where a series of ioentical tones captures later nearby tones. For two points in tIme to be sequentially grouped by
this mechanIsm, there is no need for activity to continue between onsets as in Browns model
[Brown and Cooke, 1996], or to be held in multiple spatial locations as Wang [Wang, 1995]
does. Since the gamma band response to a single auditory input onset lasts only 100 - 150
ms, there is no 40 Hz activity available in prim~ cortex (at most stimulus rates) for succesive inputs to synchronize with for sequential bmding by these mechanisms.
Furthermore, the decorrelation rule, when added to the mechanism of timing expectancies,
explains the loss of relative timing (order) between streams, since the lateral connections
that normally broadcast actual and expected onsets accross auditory cortex, are cut between
two streams by the decorrelating weight reduction. Expected and actual onset events in
different streams can no longer be directly (locally) compared. Experimental evidence for
the broadcast of expectancies comes from the fast generalization to other frequencies of a
learned expectancy for the onset time of a tone of a particular frequency (Schreiner lab personal commumcation).
When rhythmic structure is present, the expectancy system becomes engaged, and this becomes an additional feature dimension along which stimuli can be segregated. Distance
from expected timing as well as sound quality is now an added factor causing stream formation by decoupling and eigenfrequency ShIft. Feedback of expected input can also partially"fill in" missing input for a cycle or two so that the expectancy protects the binding of
features of a stimulus and stabilizes a perceptual stream accross seconds of time.
40 Hz AItentional Streams, Rhythmic Expectation, and Auditory Scene Analysis
3.3
9
Simulation or the Jones-Bregman Experiment
Figure 2 shows the architecture used to simulate the Jones-Bregman experiment. The case
shown is where the flanking tones are in the same stream as the targets because the captor
stream is at the lower sound frequency channel. At the particular pomt in time shown here,
the first flanking tone has just fimshed, and the first target tone has arrived. Both channels are
therfore active, and synchronized with the attentional stream into the higher order sequence
recognizer.
Our mechanistic explanation of the Bregman result is that the early standard target tones
arriving at the 80 msec rate first prime ttie dynamic attention system by setting the 80 msec
clock to oscillate at 40 Hz and depressing the oscillation frequency of other auditory cortex
background activi~y. Then the slow captor tones at the 240 msec period establish a background stream at 30 Hz with a rhythmic expectancy that is later violated by the appearance of
the fast target tones. These now fall outside the correlation attractor basin of the background
stream because the mismatch increases their cortical oscillation frequency. They are explicitly brought into the 40 Hz foreground frequency by the mismatch pop out mechanism. This
allows the attentional stream into the Elman sequence recognition units to synchronize and
read in activity due to the target tones for order determination. It is assisted by the timebase
searchlight at the 80 msec period which synchronizes and enhances activity arriving at that
rhythm. In the absence of a rhythmic distmction for the target tones, their sound frequency
difference alone is insufficient to separate them from the background stream, and the targets
cannot be reliably discriminated.
In this simulation, the connections to the first two Elman associative memory units are hand
wired to the A and B primary cortex oscillators to act as a latching, order determining switch.
If sy-nchronized to the memory unit at the attentional stream frequency, the A target tone
OSCillator will drive the first memory unit into the 1 attractor whicfi then inhibits the second
unit from being driven to 1 by the B target tone. The second unit has similar wiring from the
B tone oscillator, so that the particular higher order (intermediate term) memory unit which
is left in the 1 state after a tnal indicates to the rest of the brain which tone came first. The
flanking and high captor tone oscillator is connected equally to both memory units, so that a
random attractor transition occurs before the targets amve, when it is interfering at the 40 Hz
attentional frequency, and poor order determination results. If the flanking tone oscillator
is in a separate stream along with the captor tones at the background eigenfrequency of 35
Hz, it is outside the recieving passband of the memory units and cannot cause a spurious
attractor transition.
This architecture demonstrates mechanisms that integrate the theories of Jones and Bregman
about auditory perception. Stream formation is a preattentive process that works well on
non-rhythmic inputs as Bregman asserts, but an equally primary and preattentive rhythmic
expectancy process is also at work as Jones asserts and the mismatch negativity indicates.
This becomes a factor in stream formation when rhythmic structure is present in stimuli as
demonstrated by Jones.
References
[Baird et al., 1994] Baird, B., Troyer, T., and Eeckman, F. H. (1994). Gramatical inference
by attentional control of synchronization in an oscillating elman network. In Hanson, S.,
Cowan, J., and Giles, C., editors, Advances in Neural InjormationProcessing Systems 6,
pages 67-75. Morgan Kaufman.
[Bregman, 1990] Bregman, A. S. (1990). Auditory Scene Analysis. MIT Press, Cambridge.
[Brown and Cooke, 1996] Brown, G. and Cooke, M. (1996). A neural oscillator model
of auditory stream segregation. In JJCAI Workshop on Computational Auditory Scene
Analysis. to appear.
[Jones et al., 1981] Jones, M., Kidd, G., and Wetzel, R. (1981). Evidence for rhythmic
attention. Journal 0/ Experimental Psychology: Human Perception and Performance,
7: 1059-1073.
[Llinas and Ribary, 1993] Llinas, R. and Ribary, U. (1993). Coherent 40-hz oscillation
characterizes dream state in humans. Proc. Natl. Acad. Sci. USA,90:2078-2081.
[Naatanen, 1992] Naatanen, R. (1992). Attention and Brain Function. Erlbaum, New Jersey.
[Wang, 1~95] Wang, D. (1995). ~n oscillatory correlation theory of te~poral pattern segmentatIOn. In Covey, E., Hawkms, H., McMullen, T., and Port, R., edaors, Neural Representations o/Temporal Patterns. Plenum. to appear.
| 1400 |@word trial:2 hippocampus:1 simulation:4 attended:5 pick:1 thereby:1 reduction:2 moment:2 series:2 contains:1 current:1 activation:3 must:4 interrupted:1 distant:2 arrayed:1 motor:13 hypothesize:3 discrimination:1 infant:1 tenn:1 selected:1 pacemaker:1 nervous:1 tone:50 alone:1 plane:1 short:2 core:1 pointer:1 math:1 quantized:1 location:7 successive:1 along:3 constructed:1 direct:2 burst:1 become:1 olfactory:1 manner:1 rostral:2 expected:7 behavior:4 themselves:2 elman:4 multi:1 brain:7 automatically:1 actual:2 accross:3 window:3 vertebrate:1 becomes:5 begin:1 increasing:1 underlying:1 matched:1 kind:1 kaufman:1 monkey:2 developed:2 rudnicky:2 synchro:1 temporal:9 berkeley:3 act:4 demonstrates:1 control:10 unit:14 normally:1 appear:2 continually:1 before:3 bind:2 timing:8 local:1 depended:1 accordance:1 switching:1 acad:1 meet:1 bird:1 evoked:3 suggests:4 range:3 locked:1 crosstalk:2 lost:2 implement:2 area:15 evolving:1 thought:2 adapting:1 regular:1 get:1 cannot:2 close:3 context:2 unattended:1 bill:1 demonstrated:3 missing:1 send:2 nonspecific:1 attention:23 go:2 rabbit:1 focused:1 recovery:2 activi:1 perceive:1 schreiner:1 rule:3 utilizing:1 nuclear:1 fill:1 classic:1 notion:1 coordinate:4 plenum:1 deteriorated:1 target:30 spontaneous:1 hypothesis:2 element:1 recognition:2 satisfying:1 cut:1 observed:2 module:15 wang:4 capture:7 region:2 cycle:12 connected:1 movement:1 counter:4 trade:1 reward:1 dynamic:5 personal:1 segment:1 serve:1 anesthesiologist:1 basis:1 cat:2 listener:2 talk:1 neurotransmitter:1 jersey:1 distinct:1 fast:7 activate:1 accented:1 formation:8 outside:2 refined:1 premotor:1 film:1 solve:2 larger:1 slug:1 reticular:1 ability:1 transform:1 itself:2 associative:5 sequence:9 net:1 interaction:2 interconnected:1 reset:2 adaptation:1 fr:1 causing:2 relevant:1 rapidly:1 gammatone:3 moved:1 asserts:2 produce:2 wired:1 incremental:1 oscillating:1 object:3 coupling:5 develop:1 weakens:1 strong:1 dividing:1 somatosensory:2 implies:1 synchronized:8 tlie:1 come:1 correct:1 filter:5 centered:2 human:5 explains:2 generalization:1 adjusted:1 assisted:1 proximity:1 normal:1 claim:1 stabilizes:1 driving:1 vary:2 early:1 recognizer:1 proc:1 radio:1 label:2 coordination:1 largest:1 grouped:1 repetition:3 create:2 caudal:2 brought:1 mit:1 boosted:2 derived:4 focus:1 directs:1 indicates:2 detect:1 kidd:1 inference:2 streaming:2 entire:1 hidden:2 spurious:1 her:3 selective:3 overall:1 among:1 insect:1 spatial:4 special:1 field:1 once:1 identical:3 poral:1 synchronizing:2 jones:30 thinking:1 promote:1 alter:1 foreground:1 others:1 stimulus:20 employ:1 composed:1 simultaneously:2 gamma:9 recognize:2 individual:1 phase:6 attractor:17 ab:1 attempt:1 intra:1 activated:2 natl:1 held:1 bregman:23 isochronous:1 necessary:1 unifonnly:1 divide:1 obscured:1 minimal:1 psychological:2 increased:1 column:2 giles:1 exchanging:1 deviation:3 subset:1 contralateral:1 erlbaum:1 reported:1 emg:1 autistic:1 discriminable:1 periodic:1 fundamental:1 peak:2 sensitivity:1 told:1 off:1 receiving:1 together:1 quickly:2 connectivity:1 lever:1 unavoidable:1 ear:1 broadcast:3 prefrontal:1 worse:1 cognitive:3 external:2 audition:2 account:2 potential:5 coding:2 sec:1 includes:1 metronome:1 baird:9 configured:1 explicitly:1 caused:1 register:1 depends:1 stream:54 onset:10 view:1 later:2 lab:1 doing:2 characterizes:1 thalamic:4 synchrony:8 formed:1 phoneme:1 who:1 ensemble:4 yield:1 sy:1 drive:2 tissue:2 explain:1 oscillatory:7 reach:1 whenever:1 synaptic:1 decorrelated:2 against:1 ty:1 sensorimotor:1 frequency:43 recovers:1 gain:1 auditory:32 fre:1 treatment:1 recall:1 distractors:2 organized:1 nmda:1 amplitude:5 segmentation:1 back:1 nerve:1 appears:5 higher:13 amphibian:1 llinas:3 response:2 decorrelating:1 depressing:1 box:1 strongly:1 furthermore:1 just:3 clock:16 until:1 hand:3 correlation:2 synchronizes:1 expressive:1 multiscale:2 overlapping:2 accent:1 quality:1 hlh:3 facilitate:1 effect:7 usa:1 brown:10 alternating:2 read:2 wiring:1 during:1 recurrence:1 speaker:1 rhythm:20 noted:1 rat:1 clocked:2 m:3 arrived:1 complete:1 neocortical:1 occuring:1 covert:2 delivers:1 dedicated:1 motion:2 novel:1 common:2 specialized:1 functional:1 stimulation:3 discriminated:1 perturbing:1 analog:1 functionally:1 cambridge:1 ai:1 mother:1 eeckman:1 grid:1 mathematics:2 emporal:1 moving:1 habitual:1 actor:1 cortex:23 longer:2 base:13 dominant:1 own:1 recent:1 irrelevant:1 hemisphere:1 apart:2 driven:3 phonetic:1 prime:1 binary:1 continue:1 came:1 seen:1 captured:2 greater:2 additional:1 morgan:1 accomplishes:1 determine:2 period:6 dashed:1 signal:4 sliding:1 multiple:2 full:3 sound:9 reduces:1 ii:1 thalamus:1 hebbian:2 stem:1 match:4 determination:3 gesture:1 cross:1 faster:1 equally:2 impact:2 pitch:2 basic:1 vision:3 expectation:8 repetitive:1 whereas:1 ilie:1 background:9 addition:1 interval:1 source:3 rest:1 hz:48 subject:6 cowan:1 flow:2 fate:1 near:1 intermediate:2 enough:3 switch:1 affect:1 fit:1 psychology:1 architecture:15 inner:2 decline:1 absent:1 shift:4 speech:5 cause:8 oscillate:2 action:3 repeatedly:1 cocktail:1 tune:1 band:7 locally:1 succesive:1 generate:3 fish:1 shifted:1 dotted:1 per:1 discrete:1 four:2 gramatical:1 erp:2 blocker:1 orient:1 respond:1 communicate:2 resonant:6 tomographic:1 separation:1 oscillation:9 bound:2 hi:1 pay:2 captor:14 guaranteed:1 convergent:1 layer:2 fan:1 activity:29 adapted:1 occur:1 scene:6 protects:1 nearby:2 fourier:1 simulate:4 aspect:1 speed:1 inhibits:1 according:1 alternate:1 poor:1 slightly:1 wi:1 primate:1 happens:1 flanking:15 segregation:6 equation:1 turn:1 mechanism:10 fail:2 initiate:1 mechanistic:1 serf:1 sending:1 echoic:1 available:1 operation:2 probe:1 hierarchical:1 away:3 appropriate:1 voice:1 top:4 ribary:3 assumes:1 music:1 establish:2 passband:2 sweep:1 psychophysical:1 added:3 occurs:2 primary:17 ccc:1 cycling:2 exhibit:2 loudness:2 enhances:1 distance:4 attentional:24 unable:2 separate:4 lateral:5 sci:1 softwire:1 cochlear:2 extent:1 cellular:1 toward:1 dream:2 latching:1 meg:2 code:1 retained:1 modeled:2 insufficient:2 onl:1 trace:1 rise:2 ba:1 reliably:1 proper:1 upper:1 vertical:1 observation:1 regenerate:1 voluntary:1 segregate:1 communication:3 frame:4 sweeping:1 waking:1 fonning:1 tive:1 searchlight:5 required:4 pair:1 extensive:1 connection:5 hanson:1 tap:1 coherent:4 learned:3 distinction:2 temporary:4 established:1 pop:4 trans:1 suggested:2 dynamical:1 perception:6 mismatch:12 pattern:9 involuntary:1 biasing:1 reading:2 appeared:3 program:1 reliable:1 memory:14 explanation:2 shifting:1 event:6 critical:1 natural:1 decorrelation:1 synchronize:6 indicator:2 altered:1 created:1 negativity:8 coupled:1 evolve:1 contributing:1 relative:4 segregated:1 synchronization:11 loss:1 determining:1 abolished:1 mcmullen:1 localized:1 integrate:1 sufficient:3 consistent:1 basin:1 principle:2 editor:1 story:1 port:1 interfering:1 cooke:8 periodicity:2 last:1 arriving:2 quency:1 verbal:2 allow:4 pulled:1 telling:1 fall:2 rhythmic:25 sparse:1 van:1 overcome:2 default:3 cortical:21 transition:5 stand:1 dimension:1 feedback:1 sensory:10 qualitatively:1 adaptive:1 expectancy:15 replicated:1 made:1 reptile:1 party:1 far:2 gestalt:1 approximate:1 sequentially:3 incoming:3 anchor:1 active:2 excite:2 pen:1 learn:1 channel:7 ca:1 decoupling:1 eeg:4 mnn:3 quantize:1 expansion:1 investigated:1 complex:1 troyer:1 reconciled:1 spread:3 noise:2 arrival:3 deviant:2 child:1 body:4 site:1 slow:2 experienced:1 strabismus:1 msec:7 bandpass:1 entrain:3 perceptual:5 third:1 down:7 specific:2 showing:1 prim:1 timebase:5 recurrently:1 learnable:1 divergent:1 physiological:2 offset:1 timbre:2 evidence:10 t9:1 decay:3 consist:1 intrinsic:2 grouping:1 sequential:3 effectively:1 workshop:1 interactional:2 te:1 perceptually:1 illustrates:1 associational:1 thalamo:1 occurring:1 appearance:2 wetzel:1 visual:2 horizontally:1 partially:2 binding:4 environmental:2 constantly:1 presentation:3 oscillator:8 absence:8 content:1 change:3 included:1 determined:1 entrainment:2 called:2 engaged:1 experimental:3 preattentive:2 selectively:3 select:1 internal:2 support:2 violated:1 dept:1 phenomenon:1 |
441 | 1,401 | Coding of Naturalistic Stimuli by
Auditory Midbrain Neurons
H. Attias* and C.E. Schreiner t
Sloan Center for Theoretical Neurobiology and
W.M. Keck Foundation Center for Integrative Neuroscience
University of California at San Francisco
San Francisco, CA 94143-0444
Abstract
It is known that humans can make finer discriminations between
familiar sounds (e.g. syllables) than between unfamiliar ones (e.g.
different noise segments). Here we show that a corresponding enhancement is present in early auditory processing stages. Based on
previous work which demonstrated that natural sounds had robust
statistical properties that could be quantified, we hypothesize that
the auditory system exploits those properties to construct efficient
neural codes. To test this hypothesis, we measure the information rate carried by auditory spike trains on narrow-band stimuli
whose amplitude modulation has naturalistic characteristics, and
compare it to the information rate on stimuli with non-naturalistic
modulation. We find that naturalistic inputs significantly enhance
the rate of transmitted information, indicating that auditiory neural responses are matched to characteristics of natural auditory
scenes.
1
Natural Scene Statistics and the Neural Code
A primary goal of hearing research is to understand how complex sounds that occur
in natural scenes are processed by the auditory system. However, natural sounds
are difficult to describe quantitatively and the complexity of auditory responses
they evoke makes it hard to gain insight into their processing. Hence, most studies
of auditory physiology are restricted to pure tones and noise stimuli, resulting in
a limited understanding of auditory encoding. In this paper we pursue a novel
approach to the study of natural sound encoding in auditory spike trains. Our
? Corresponding author. E-mail: hagai@phy.ucsf.edu.
t E-mail: chris@phy.ucsf.edu .
H. Attias and C. E. Schreiner
104
~
11111111 I II
I II I I 111111
~
I 111111 III 11111111 III II III UII
Figure 1: Left: amplitude modulation stimulus drawn from a naturalistic stimulus
set, and the evoked spike train of an inferior colliculus neuron. Right: amplitude
modulation from a non-naturalistic set and the evoked spike train of the same
neuron.
method consists of measuring statistical characteristics of natural auditory scenes,
and incorporating them into simple stimuli in a systematic manner, thus creating
'naturalistic' stimuli which enable us to study the encoding of natural sounds in a
controlled fashion. The first stage of this program has been described in (Attias
and Schreiner 1997); the second is reported below.
Fig. 1 shows two segments of long stimuli and the corresponding spike trains of
the same neuron, elicited by pure tones that were amplitude-modulated by these
stimuli. While both stimuli appear to be random and to have the same mean and
both spike trains have the same firing rate, one may observe that high and low
amplitudes are more likely to occur in the stimulus on the left; indeed, these stimuli
are drawn from two stimulus sets with different statistical properties. Our present
study of auditory coding focuses on assessing the efficiency of this neural code:
for a given stimulus set, how well can the animal reconstruct the input sound and
discriminate between similar sound segments, based on the evoked spike train, and
how those abilities are affected by changing the stimulus statistics. We quantify
the discrimination capability of auditory neurons in the inferior colliculus of the cat
using concepts from information theory (Bialek et al. 1991; Rieke et al. 1997).
This leads to the issue of optimal coding (Atick 1992). Theoretically, given an
auditory scene with particular statistical properties, it is possible to design an encoding scheme that would exploit those properties, resulting in a neural code that
is optimal for that scene but is consequently less efficient for other scenes. Here we
investigate the hypothesis that the auditory system uses a code that is adapted to
natural auditory scenes. This question is addressed by comparing the discrimination
capability of auditory neurons between sound segments drawn from a naturalistic
stimulus set, to the one for a non-naturalistic set.
2
Statistics of Natural Sounds
As a first step in investigating the relation between neural responses and auditory
inputs, we studied and quantified temporal statistics of natural auditory scenes {Attias and Schreiner 1997}. It is well-known that different locations on the basal membrane respond selectively to different frequency components of the incoming sound
x{t) (e.g., Pickles 1988), hence the frequency v corresponds to a spatial coordinate,
in analogy with retinal location in vision. We therefore analyzed a large database
of sounds, including speech, music, animal vocalizations, and background sounds,
using various filter banks comprising 0 -10kHz. In each frequency band v, the amplitude a{t) ~ 0 and phase r/>{t) ofthe band-limited signal xv(t) = a{t) cos{vt+r/>{t))
were extracted, and the amplitude probability distribution p(a) and auto-correlation
Coding of Naturalistic Stimuli by Auditory Midbrain Neurons
Piano music
105
Symphonic: music:
0.6
0.5
Cat vQcalizatlona
Bird 80ngs
Ba.okground sound.
0 .5
0 .5
0 .4
0 .2
o. ~
o~~--------~~
-4
Figure 2: Log-amplitude distribution in several sound ensembles. Different curves
for a given ensemble correspond to different frequency bands. The low amplitude
peak in the cat plot reflect abundance of silent segments. The theoretical curve p(a)
(1) is plotted for comparison (dashed line).
function c(r) = (a(t)a(t + r)) were computed, as well as those of the instantaneous
frequency d?(t)/dt.
Those statistics were found to be nearly identical in all bands and across all examined sounds. In particular, the distribution of the log-amplitude a = log a,
normalized to have zero mean and unit variance, could be well-fitted to the form
p(a) = ,8 exp (,8a + Q
-
e.Bii+t:t)
(1)
(with normalization constants Q = -.578 and ,8 = 1.29), which should, however, be
corrected at large amplitude (> 5a). Several examples are displayed in Fig. 1. The
log-amplitude distribution (1) corresponds mathematically to the amplitude distribution of musical instruments and vocalizations, found to be p(a) = e- a (known as
the Laplace distribution in speech signal processing), as well as that of background
sounds, where p(a) <X ae- a2 (which can be shown to be the band amplitude distribution for a Gaussian signal). The power spectra of a(t) (Fourier transform of c(r))
were found to have a modified 1/ f form.
Together with the results for ?(t), those findings show that natural sounds are
distinguished from arbitrary ones by robust characteristics. In the present paper
we explore to what extent the auditory system exploits them in constructing efficient
neural codes. Another important point made by (Attias and Schreiner 1997), as well
as by (Ruderman and Bialek 1994) regarding visual signals, is that natural inputs
are very often not Gaussian (e.g. (1)), unlike the signals used by conventional
system-identification methods often applied to the nervous system. In this paper
we use non-Gaussian stimuli to study auditory coding.
3
3.1
Measuring the Rate of Information Transfer
Experiment
Based on our results for temporal statistics of natural auditory scenes, we can
construct 'naturalistic' stimuli by starting with a simple signal and systematically
incorporate successively more complicated characteristics of natural sounds into it.
H. Attias and C. E. Schreiner
106
We cQ.ose to use narrow-band stimuli consisting of amplitude-modulated carriers
a(t) cos(vt) at sound frequencies v = 2 - 9kHz with no phase modulation. Focusing
on one-point amplitude statistics, we constructed a white naturalistic amplitude
by choosing a(t) from an exponential distribution with a cutoff, p(O ::; a ::; ae) ex:
e- a , p(a > ae) = 0 at each time point t independently, using a cutoff modulation
frequency of fe = 100Hz (i.e., 1a(J ::; fe) 1= const., 1a(J > fe) 1= 0, where a(J) is
the Fourier transform of a{t)). We also used a non-naturalistic stimulus set where
a(t) was chosen from a uniform distribution p(O ::; a ~ be) = 1lbe, p(a > be) = 0,
with be adjusted so that both stimulus sets had the same mean. A short segment
from each set is shown in Fig. 1, and the two distributions are plotted in Figs. 3,4
(right) .
Stimuli of 15 - 20min duration were played to ketamine-anesthetized cats. To
minimize adaptation effects we alternated between the two sets using 10sec long
segments. Single-unit recordings were made from the inferior colliculus (IC), a subthalamic auditory processing stage (e.g., Pickles 1988). Each IC unit responds best
to a narrow range of sound frequencies, the center of which is called its 'best frequency' (BF). Neighboring units have similar BF's, in accord with the topographic
frequency organization of the auditory system. For each unit, stimuli with carrier
frequency v at most 500Hz away from the unit's BF were used. Firing rates in
response to those stimuli were between 60 - 100Hz. The stimulus and the electrode
signal were recorded simultaeneously at a sampling rate of 24kHz. After detecting
and sotring the spikes and extracting the stimulus amplitude, both amplitude and
spike train were down-sampled to 3kHz.
3.2
Analysis
In order to assess the ability to discriminate between different inputs based on the
observed spike train, we computed the mutual information Ir,s between the spike
train response r(t) = Li o(t - ti), where ti are the spike times, and the stimulus
amplitude s(t). I consists of two terms, Ir,s = Hs - Hslr' where Hs is the stimulus
entropy (the log-number of different stimuli) and Hslr is the entropy of the stimulus
conditioned on the response (the log-number of different stimuli that could elicit
a given response, and thus could not be discriminated based on that response,
averaged over all responses). Our approach generally follows the ideas of (Bialek et
al. 1991; Rieke et al. 1997).
To simplify the calculation, we first modified the stimuli s(t) to get s'(t) = f(s(t?,
where the function f(s) was chosen so that s' was Gaussian. Hence for exponential
stimuli f(s) = y'(2)erfi(1-2e- S ) and for uniform stimuli f(s) = y'(2)erfi(2slb e-1),
where erfi is the inverse error function. This Gaussianization has two advantages:
first, the expression for the mutual information Ir,s' (= Ir,s) is now simpler, being
given by the frequency-dependent signal-to-noise ratio SNR(J) (see below), since
Hs' depends only on the power spectrum of s'(t); second and more importantly,
the noise distribution was observed to become closer to Gaussian following this
transformation.
ftc
dfH[s'(J) 1 f(J)], the calculation of
To compute Hs'lr we bound it from above by
which requires the conditional distribution p[s'(J) 1 f(J)] (note that these variables
are complex, hence this is the joint ditribution of the real and imaginary parts).
The latter is approximated by a Gaussian with mean s~(J) and variance Nr(f).
This variance is, in fact, the power spectrum of the noise, Nr(J) = (I nr(J) 12 ),
which we define by nr(t) = s'(t) - s~(t). Computing the mutual information for
those Gaussian distributions is straightforward and provides a lower bound on the
Coding of Naturalistic Stimuli by Auditory Midbrain Neurons
107
1
I,
0.8
I,
",
I
I
I
I
,
,
,
1.6
60
100
f
Figure 3: Left: signal-to-noise ratio SNR(f) vs. modulation frequency I for naturalistic stimuli. Right: normalized noise distribution (solid line), amplitude distribution of stimuli (dashed line) and of Gaussianized stimuli (dashed-dotted line).
true Ir,s,
Ir,s = Ir,s'
~
Ie
J
dllog 2 SNR(f) .
(2)
o
The signal-to-noise ratio is given by SNR(f) = S'(J)j(Nr(f))r, where S'(f) = (I
s'(J) 12} is the spectrum of the Gaussianized stimulus and the averaging Or is
performed over all responses.
The main object here is s~(J), which is an estimate of the stimulus from the elicited
spike train, and would optimally be given by the conditional mean J ds's'p(s' 1 f)
at each I (Kay 1993). For Gaussian p( S' ,f) this estimator, which is generally non-linear, becomes linear in f(f) and is given by h(J)f(J), where h(J)
(s'(J)f*(J)}j(f(J)f*(f?) is the Wiener filter. However, since our distributions were
only approximately Gaussians we used the conditional mean, obtained by the kernel
estimate
=
s~(f)
=
(3)
where k is a Gaussian kernel, R(J) is the spectrum of the spike train, and i indexes
the data points obtained by computing FFT using a sliding window. The scaling
by y'Si,.,fii reflects the assumption that the distributions at all I differ only by
their variance, which enables us to use the data points at all frequencies to estimate
s~ at a given I. Our estimate produced a slightly higher SNR(f) than the Wiener
estimate used by (Bialek et al. 1991;Rieke et al. 1997) and others.
4
Information on Naturalistic Stimuli
The SNR(f) for exponential stimuli is shown in Fig. 3 (left) for one of our units.
Ie neurons have a preferred modulation frequency 1m (e.g., Pickles 1988), which is
about 40Hz for this unit; notice that generally SNR(J) ~ 1, with equality when the
stimulus and response are completely independent. Thus, stimulus components at
frequencies higher than 60Hz effectively cannot be estimated from the spike train.
The stimulus amplitude distribution is shown in Fig. 3 (right, dashed line), together
with the noise distribution (normalized to have unit variance; solid line) which is
nearly Gaussian.
H Attias and C. E. Schreiner
J08
4,-----------------------,
.. .. ..
1
3.5
0.8
3
~2 . 6
. '.
~o.e
en
0.4
2
0.2
1.5
100
0
-5
.
Figure 4: Left: signal-to-noise ratio SNR(f) vs. modulation frequency f for nonnaturalistic stimuli (solid line) compared with naturalistic stimuli (dotted line).
Right: normalized noise distribution (solid line), amplitude distribution of stimuli (dashed line) compared with that of naturalistic stimuli (dotted line), and of
Gaussianized stimuli (dashed-dotted line).
Using (2) we obtain an information rate of Ir,s ~ 114bit/sec. For the spike rate
of 82spike/sec measured in this unit, this translates into 1.4bit/spike. Averaging
across units, we have 1.3 ? 0.2bit/spike for naturalistic stimuli.
Although this information rate was computed using the conditional mean estimator
(3), it is interesting to examine the Wiener filter h{t) which provides the optimal
linear estimator of the stimulus, as discussed in the previous section. This filter
is displayed in Fig. 5 (solid line) and has a temporal width of several tens of
milliseconds.
5
Information on Non-Naturalistic Stimuli
The SNR(f) for uniform stimuli is shown in Fig. 4 (left, solid line) for the same
unit as in Fig. 3, and is significantly lower than the corresponding SNR(f) for exponential stimuli plotted for comparison (dashed line). For the mutual information
rate we obtain Ir,B ~ 77bit/sec, which amounts to 0.94bit/spike. Averaging across
units, we have 0.8 ? 0.2bit/spike for non-naturalistic stimuli.
The stimulus amplitude distribution is shown in Fig. 4 (right, dashed line), together
with the exponential distribution (dotted line) plotted for comparison, as well as
the noise distribution (normalized to have unit variance). The noise in this case is
less Gaussian than for exponential stimuli, suggesting that our calculated bound on
Ir,s may be lower for uniform stimuli.
Fig. 5 shows the stimulus reconstruction filter (dashed line). It has a similar time
course as the filter for exponential stimuli, but the decay is significantly slower and
its temporal width is more than 100msec.
6
Conclusion
We measured the rate at which auditory neurons carry information on simple stimuli
with naturalistic amplitude modulation, and found that it was higher than for
stimuli with non-naturalistic modulation. A result along the same lines for the frog
was obtained by (Rieke et al. 1995) using Gaussian signals whose spectrum was
shaped according to the frog call spectrum. Similarly, work in vision (Laughlin 1981;
Field 1987; Atick and Redlich 1990; Ruderman and Bialek 1994; Dong and Atick
1995) suggests that visual receptive field properties are consistent with optimal
coding predictions based on characteristics of natural images. Future work will
explore coding of stimuli with more complex natural statistical characteristics and
Coding a/Naturalistic Stimuli by Auditory Midbrain Neurons
109
0 .6 , - - - - - - - - - - - - - - - - - - - - - - - - .
0 .4
0 .3
~
0 .2
0 . '"
__ -.r..1
-'-=----::;::-.::: v Vrr-~------==---__l
o 1.-___
-0.'"
-0 .2
L---_-=o=-...,=-------::0::-------::0,--...,=------'
t
Figure 5: Impulse response of Wiener reconstruction filter for naturalistic stimuli
(solid line) and non-naturalistic stimuli (dashed line).
will extend to higher processing stages.
Acknowledgements
We thank W. Bialek, K. Miller, S. Nagarajan, and F. Theunissen for useful discussions and B. Bonham, M. Escabi, M. Kvale, L. Miller, and H. Read for experimental
support. Supported by The Office of Naval Research (NOOOI4-94-1-0547), NIDCD
(ROI-02260), and the Sloan Foundation.
References
J.J. Atick and N. Redlich (1990). Towards a theory of early visual processing.
Neural Comput. 2,308-320.
J .J. Atick (1992). Could information theory provide an ecological theory of sensory
processing. Network 3, 213-251.
H. Attias and C.E. Schreiner (1997). Temporal low-order statistics of natural
sounds. In Advances in Neural Information Processing Systems 9, MIT Press.
W. Bialek, F. Rieke, R. de Ruyter van Steveninck, and D. Warland (1991). Reading
the neural code. Science 252, 1854-1857.
D.W. Dong and J.J. Atick (1995). Temporal decorrelation: a theory of lagged and
non-lagged responses in the lateral geniculate nucleus. Network 6, 159-178.
D.J. Field (1987). Relations between the statistics of natural images and the response properties of cortical cells. J. Opt. Soc. Am. 4, 2379-2394.
S.M. Kay (1993). Fundamentals of Statistical Signal Processing: Estimation Theory.
Prentice-Hall, New Jersey.
S.B. Laughlin (1981). A simple coding procedure enhances a neuron's information
capacity. Z. Naturforsch. 36c, 910-912.
J.O. Pickles (1988). An introduction to the physiology of hearing (2nd Ed.). San
Diego, CA: Academic Press.
F. Rieke, D. Bodnar, and W. Bialek (1995). Naturalistic stimuli increase the rate
and efficiency of information transmission by primary auditory neurons. Proc. R.
Soc . Lond. B, 262, 259-265.
F. Rieke, D. Warland, R. de Ruyter van Steveninck, and W. Bialek (1997). Spikes:
Exploring the Neural Code. MIT Press, Cambridge, MA.
D.L. Ruderman and W. Bialek (1994). Statistics of natural images: scaling in the
woods. Phys. Rev. Lett. 73,814-817.
| 1401 |@word h:4 nd:1 bf:3 integrative:1 solid:7 carry:1 phy:2 imaginary:1 comparing:1 si:1 slb:1 enables:1 hypothesize:1 plot:1 discrimination:3 v:2 nervous:1 tone:2 short:1 lr:1 detecting:1 provides:2 location:2 simpler:1 along:1 constructed:1 become:1 consists:2 manner:1 theoretically:1 indeed:1 examine:1 window:1 becomes:1 matched:1 what:1 pursue:1 finding:1 transformation:1 temporal:6 ti:2 unit:14 appear:1 carrier:2 xv:1 encoding:4 firing:2 modulation:11 approximately:1 bird:1 frog:2 studied:1 quantified:2 evoked:3 examined:1 suggests:1 co:2 limited:2 range:1 averaged:1 steveninck:2 procedure:1 elicit:1 significantly:3 physiology:2 dfh:1 naturalistic:27 get:1 cannot:1 prentice:1 conventional:1 demonstrated:1 center:3 straightforward:1 starting:1 independently:1 duration:1 pure:2 schreiner:8 insight:1 estimator:3 importantly:1 kay:2 rieke:7 coordinate:1 laplace:1 diego:1 us:1 hypothesis:2 approximated:1 database:1 theunissen:1 observed:2 complexity:1 segment:7 efficiency:2 completely:1 joint:1 cat:4 various:1 jersey:1 train:13 describe:1 choosing:1 whose:2 reconstruct:1 ability:2 statistic:10 topographic:1 transform:2 vocalization:2 advantage:1 reconstruction:2 adaptation:1 neighboring:1 enhancement:1 keck:1 assessing:1 electrode:1 transmission:1 object:1 measured:2 soc:2 quantify:1 differ:1 gaussianization:1 filter:7 human:1 enable:1 nagarajan:1 opt:1 hagai:1 mathematically:1 adjusted:1 exploring:1 hall:1 ic:2 roi:1 exp:1 early:2 a2:1 estimation:1 proc:1 geniculate:1 pickle:4 reflects:1 mit:2 gaussian:12 modified:2 lbe:1 office:1 focus:1 naval:1 am:1 dependent:1 ftc:1 relation:2 comprising:1 issue:1 animal:2 spatial:1 mutual:4 field:3 construct:2 shaped:1 ng:1 sampling:1 identical:1 nearly:2 future:1 others:1 stimulus:70 quantitatively:1 simplify:1 familiar:1 phase:2 consisting:1 organization:1 investigate:1 analyzed:1 closer:1 plotted:4 theoretical:2 fitted:1 measuring:2 hearing:2 snr:10 uniform:4 optimally:1 reported:1 peak:1 fundamental:1 ie:2 systematic:1 dong:2 enhance:1 together:3 reflect:1 recorded:1 successively:1 creating:1 li:1 suggesting:1 de:2 retinal:1 coding:10 sec:4 sloan:2 depends:1 performed:1 elicited:2 capability:2 complicated:1 minimize:1 ass:1 ir:10 nidcd:1 wiener:4 variance:6 characteristic:7 musical:1 ensemble:2 correspond:1 ofthe:1 miller:2 identification:1 produced:1 finer:1 phys:1 ed:1 frequency:17 gain:1 auditory:29 sampled:1 noooi4:1 amplitude:25 focusing:1 higher:4 dt:1 response:14 stage:4 atick:6 correlation:1 d:1 ruderman:3 impulse:1 effect:1 concept:1 normalized:5 true:1 hence:4 equality:1 read:1 white:1 width:2 inferior:3 image:3 instantaneous:1 novel:1 discriminated:1 khz:4 discussed:1 extend:1 unfamiliar:1 cambridge:1 similarly:1 had:2 fii:1 ecological:1 vt:2 kvale:1 transmitted:1 signal:13 ii:3 dashed:10 sliding:1 sound:22 academic:1 calculation:2 long:2 controlled:1 prediction:1 ae:3 vision:2 normalization:1 kernel:2 accord:1 cell:1 background:2 addressed:1 unlike:1 hz:5 recording:1 call:1 extracting:1 j08:1 iii:3 fft:1 silent:1 regarding:1 idea:1 attias:8 translates:1 expression:1 speech:2 generally:3 useful:1 amount:1 band:7 ten:1 processed:1 millisecond:1 notice:1 dotted:5 estimated:1 neuroscience:1 ketamine:1 affected:1 basal:1 drawn:3 changing:1 cutoff:2 wood:1 colliculus:3 inverse:1 respond:1 scaling:2 bit:6 bound:3 syllable:1 played:1 adapted:1 occur:2 uii:1 scene:10 fourier:2 min:1 lond:1 according:1 membrane:1 across:3 slightly:1 rev:1 midbrain:4 restricted:1 instrument:1 gaussians:1 observe:1 away:1 bii:1 distinguished:1 slower:1 symphonic:1 const:1 music:3 exploit:3 warland:2 question:1 spike:22 receptive:1 primary:2 responds:1 bialek:10 nr:5 enhances:1 thank:1 lateral:1 capacity:1 chris:1 mail:2 extent:1 code:8 index:1 cq:1 ratio:4 difficult:1 fe:3 ba:1 lagged:2 design:1 neuron:13 displayed:2 neurobiology:1 arbitrary:1 california:1 erfi:3 narrow:3 below:2 reading:1 program:1 including:1 power:3 decorrelation:1 natural:20 scheme:1 carried:1 auto:1 alternated:1 understanding:1 piano:1 acknowledgement:1 interesting:1 analogy:1 foundation:2 nucleus:1 consistent:1 bank:1 systematically:1 course:1 supported:1 understand:1 laughlin:2 anesthetized:1 van:2 curve:2 calculated:1 cortical:1 lett:1 sensory:1 author:1 made:2 san:3 preferred:1 evoke:1 investigating:1 incoming:1 francisco:2 spectrum:7 naturforsch:1 gaussianized:3 transfer:1 robust:2 ca:2 ruyter:2 complex:3 constructing:1 main:1 noise:13 fig:11 redlich:2 en:1 fashion:1 ose:1 msec:1 exponential:7 comput:1 abundance:1 down:1 decay:1 incorporating:1 effectively:1 conditioned:1 entropy:2 likely:1 explore:2 visual:3 corresponds:2 extracted:1 ma:1 conditional:4 goal:1 consequently:1 towards:1 hard:1 corrected:1 averaging:3 called:1 discriminate:2 experimental:1 indicating:1 selectively:1 support:1 latter:1 modulated:2 ucsf:2 incorporate:1 bodnar:1 ex:1 |
442 | 1,402 | The Rectified Gaussian Distribution
N. D. Socci, D. D. Lee and H. S. Seung
Bell Laboratories, Lucent Technologies
Murray Hill, NJ 07974
{ndslddleelseung}~bell-labs.com
Abstract
A simple but powerful modification of the standard Gaussian distribution is studied. The variables of the rectified Gaussian are
constrained to be nonnegative, enabling the use of nonconvex energy functions. Two multimodal examples, the competitive and
cooperative distributions, illustrate the representational power of
the rectified Gaussian. Since the cooperative distribution can represent the translations of a pattern, it demonstrates the potential
of the rectified Gaussian for modeling pattern manifolds.
1
INTRODUCTION
The rectified Gaussian distribution is a modification of the standard Gaussian in
which the variables are constrained to be nonnegative. This simple modification
brings increased representational power, as illustrated by two multimodal examples
of the rectified Gaussian, the competitive and the cooperative distributions. The
modes of the competitive distribution are well-separated by regions of low probability. The modes of the cooperative distribution are closely spaced along a nonlinear
continuous manifold. Neither distribution can be accurately approximated by a
single standard Gaussian. In short, the rectified Gaussian is able to represent both
discrete and continuous variability in a way that a standard Gaussian cannot.
This increased representational power comes at the price of increased complexity.
While finding the mode of a standard Gaussian involves solution of linear equations,
finding the modes of a rectified Gaussian is a quadratic programming problem.
Sampling from a standard Gaussian can be done by generating one dimensional
normal deviates, followed by a linear transformation. Sampling from a rectified
Gaussian requires Monte Carlo methods. Mode-finding and sampling algorithms
are basic tools that are important in probabilistic modeling.
Like the Boltzmann machine[l], the rectified Gaussian is an undirected graphical
model. The rectified Gaussian is a better representation for probabilistic modeling
351
The Rectified Gaussian Distribution
(a)
(c)
Figure 1: Three types of quadratic energy functions. (a) Bowl (b) Trough (c) Saddle
of continuous-valued data. It is unclear whether learning will be more tractable for
the rectified Gaussian than it is for the Boltzmann machine.
A different version of the rectified Gaussian was recently introduced by Hinton and
Ghahramani[2, 3]. Their version is for a single variable, and has a singularity at
the origin designed to produce sparse activity in directed graphical models. Our
version lacks this singularity, and is only interesting in the case of more than one
variable, for it relies on undirected interactions between variables to produce the
multimodal behavior that is of interest here.
The present work is inspired by biological neural network models that use continuous dynamical attractors[4]. In particular, the energy function of the cooperative
distribution was previously studied in models of the visual cortex[5], motor cortex[6],
and head direction system[7].
2
ENERGY FUNCTIONS: BOWL, TROUGH, AND
SADDLE
The standard Gaussian distribution P(x) is defined as
P(x)
E(x)
=
Z - l e -{3E(;r:) ,
1
_xT
Ax - bTx .
2
(1)
(2)
The symmetric matrix A and vector b define the quadratic energy function E(x).
The parameter (3 = lIT is an inverse temperature. Lowering the temperature
concentrates the distribution at the minimum of the energy function. The prefactor
Z normalizes the integral of P(x) to unity.
Depending on the matrix A, the quadratic energy function E(x) can have different
types of curvature. The energy function shown in Figure l(a) is convex. The minimum of the energy corresponds to the peak of the distribution. Such a distribution
is often used in pattern recognition applications, when patterns are well-modeled
as a single prototype corrupted by random noise.
The energy function shown in Figure 1(b) is flattened in one direction. Patterns
generated by such a distribution come with roughly equal1ikelihood from anywhere
along the trough. So the direction of the trough corresponds to the invariances of
the pattern. Principal component analysis can be thought of as a procedure for
learning distributions of this form.
The energy function shown in Figure 1(c) is saddle-shaped. It cannot be used
in a Gaussian distribution, because the energy decreases without limit down the
N. D. Socci, D. D. Lee and H. S. Seung
352
sides of the saddle, leading to a non-normalizable distribution. However, certain
saddle-shaped energy functions can be used in the rectified Gaussian distribution,
which is defined over vectors x whose components are all nonnegative. The class of
energy functions that can be used are those where the matrix A has the property
x T Ax > 0 for all x > 0, a condition known as copositivity. Note that this set of
matrices is larger than the set of positive definite matrices that can be used with
a standard Gaussian. The nonnegativity constraints block the directions in which
the energy diverges to negative infinity. Some concrete examples will be discussed
shortly. The energy functions for these examples will have multiple minima, and
the corresponding distribution will be multimodal, which is not possible with a
standard Gaussian.
3
MODE-FINDING
Before defining some example distributions, we must introduce some tools for analyzing them. The modes of a rectified Gaussian are the minima of the energy
function (2), subject to nonnegativity constraints. At low temperatures, the modes
of the distribution characterize much of its behavior.
Finding the modes of a rectified Gaussian is a problem in quadratic programming.
Algorithms for quadratic programming are particularly simple for the case of nonnegativity constraints. Perhaps the simplest algorithm is the projected gradient
method, a discrete time dynamics consisting of a gradient step followed by a rectification
(3)
The rectification [x]+ = max(x, 0) keeps x within the nonnegative orthant (x ~ 0).
If the step size 7J is chosen correctly, this algorithm can provably be shown to
converge to a stationary point of the energy function[8]. In practice, this stationary
point is generally a local minimum.
Neural networks can also solve quadratic programming problems. We define the
synaptic weight matrix W = I - A, and a continuous time dynamics
x+x = [b+ Wx]+
(4)
For any initial condition in the nonnegative orthant, the dynamics remains in the
nonnegative orthant, and the quadratic function (2) is a Lyapunov function of the
dynamics.
Both of these methods converge to a stationary point of the energy. The gradient
of the energy is given by 9 = Ax - b. According to the Kiihn-Tucker conditions, a
stationary point must satisfy the conditions that for all i, either gi = 0 and Xi > 0,
or gi > 0 and Xi = O. The intuitive explanation is that in the interior of the
constraint region, the gradient must vanish, while at the boundary, the gradient
must point toward the interior. For a stationary point to be a local minimum, the
Kiihn-Tucker conditions must be augmented by the condition that the Hessian of
the nonzero variables be positive definite.
Both methods are guaranteed to find a global minimum only in the case where A is
positive definite, so that the energy function (2) is convex. This is because a convex
energy function has a unique minimum. Convex quadratic programming is solvable
in polynomial time. In contrast, for a nonconvex energy function (indefinite A), it
is not generally possible to find the global minimum in polynomial time, because of
the possible presence of local minima. In many practical situations, however, it is
not too difficult to find a reasonable solution.
The Rectified Gaussian Distribution
353
(a)
(b)
Figure 2: The competitive distribution for two variables. (a) A non-convex energy
function with two constrained minima on the x and y axes. Shown are contours of
constant energy, and arrows that represent the negative gradient of the energy. (b)
The rectified Gaussian distribution has two peaks.
The rectified Gaussian happens to be most interesting in the nonconvex case, precisely because of the possibility of multiple minima. The consequence of multiple
minima is a multimodal distribution, which cannot be well-approximated by a standard Gaussian. We now consider two examples of a multimodal rectified Gaussian.
4
COMPETITIVE DISTRIBUTION
The competitive distribution is defined by
Aij
bi
=
-dij
1;
+2
(5)
(6)
We first consider the simple case N = 2. Then the energy function given by
X2 +y2
E(x,y)=2
+(x+y)2_(x+y)
(7)
has two constrained minima at (1,0) and (0,1) and is shown in figure 2(a). It
does not lead to a normalizable distribution unless the nonnegativity constraints are
imposed. The two constrained minima of this nonconvex energy function correspond
to two peaks in the distribution (fig 2(b)). While such a bimodal distribution
could be approximated by a mixture of two standard Gaussians, a single Gaussian
distribution cannot approximate such a distribution. In particular, the reduced
probability density between the two peaks would not be representable at all with a
single Gaussian.
The competitive distribution gets its name because its energy function is similar
to the ones that govern winner-take-all networks[9]. When N becomes large, the
N global minima of the energy function are singleton vectors (fig 3), with one
component equal to unity, and the rest zero. This is due to a competitive interaction
between the components. The mean of the zero temperature distribution is given
by
(8)
The eigenvalues of the covariance
(XiXj) - (Xi)(Xj)
1
= N dij -
1
N2
(9)
354
N. D. Socci, D. D. Lee and H. S. Seung
-
.:(a)
.: (b)
: (c)
-
.,
r-
?
?n
0
,
2
J
a
I
?
.,
?
?
to
'2
S
???
.,
??
'I
r-
III
n
n
I
? . . , . . . , . u. .
Figure 3: The competitive distribution for N = 10 variables. (a) One mode (zero
temperature state) of the distribution. The strong competition between the variables results in only one variable on. There are N modes of this form, each with a
different winner variable. (b) A sample at finite temperature (13 ~ 110) using Monte
Carlo sampling. There is still a clear winner variable. (c) Sample from a standard
Gaussian with matched mean and covariance. Even if we cut off the negative values
this sample still bears little resemblance to the states shown in (a) and (b), since
there is no clear winner variable.
all equal to 1/N, except for a single zero mode. The zero mode is 1, the vector of
all ones, and the other eigenvectors span the N - 1 dimensional space perpendicular
to 1. Figure 3 shows two samples: one (b) drawn at finite temperature from the
competitive distribution, and the other (c) drawn from a standard Gaussian distribution with the same mean and covariance. Even if the sample from the standard
Gaussian is cut so negative values are set to zero the sample does not look at all
like the original distribution. Most importantly a standard Gaussian will never be
able to capture the strongly competitive character of this distribution.
5
COOPERATIVE DISTRIBUTION
To define the cooperative distribution on N variables, an angle fh = 27ri/N is
associated with each variable Xi, so that the variables can be regarded as sitting on
a ring. The energy function is defined by
1
4
Aij
6ij + N - N COS(Oi - OJ)
(10)
bi
=
1;
(11)
The coupling Aij between Xi and Xj depends only on the separation Oi - 03. between
them on the ring.
The minima, or ground states, of the energy function can be found numerically by
the methods described earlier. An analytic calculation of the ground states in the
large N limit is also possible[5]. As shown in Figure 4(a), each ground state is a
lump of activity centered at some angle on the ring. This delocalized pattern of
activity is different from the singleton modes of the competitive distribution, and
arises from the cooperative interactions between neurons on the ring. Because the
distribution is invariant to rotations of the ring (cyclic permutations of the variables
xd, there are N ground states, each with the lump at a different angle.
The mean and the covariance of the cooperative distribution are given by
(Xi)
(XiXj) - (Xi}(Xj)
=
const
C(Oi - OJ)
(12)
(13)
=
A given sample of x, shown in Figure 4(a), does not look anything like the mean,
which is completely uniform. Samples generated from a Gaussian distribution with
355
The Rectified Gaussian Distribution
, (b)
'(a)
(c)
r
Figure 4: The cooperative distribution for N = 25 variables. (a) Zero temperature
state. A cooperative interaction between the variables leads to a delocalized pattern
of activity that can sit at different locations on the ring. (b) A finite temperature
(/3 = 50) sample. (c) A sample from a standard Gaussian with matched mean and
covariance.
the same mean and covariance look completely different from the ground states of
the cooperative distribution (fig 4(c)).
These deviations from standard Gaussian behavior reflect fundamental differences
in the underlying energy function. Here the energy function has N discrete minima
arranged along a ring. In the limit of large N the barriers between these minima
become quite small. A reasonable approximation is to regard the energy function
as having a continuous line of minima with a ring geometry[5] . In other words, the
energy surface looks like a curved trough, similar to the bottom of a wine bottle.
The mean is the centroid of the ring and is not close to any minimum.
The cooperative distribution is able to model the set of all translations of the lump
pattern of activity. This suggests that the rectified Gaussian may be useful in
invariant object recognition, in cases where a continuous manifold of instantiations
of an object must be modeled. One such case is visual object recognition, where
the images of an object from different viewpoints form a continuous manifold.
6
SAMPLING
Figures 3 and 4 depict samples drawn from the competitive and cooperative distribution. These samples were generated using the Metropolis Monte Carlo algorithm.
Since full descriptions of this algorithm can be found elsewhere, we give only a brief
description of the particular features used here . The basic procedure is to generate
a new configuration of the system and calculate the change in energy (given by
eq. 2). If the energy decreases, one accepts the new configuration unconditionally.
If it increases then the new configuration is accepted with probability e-{3AE.
In our sampling algorithm one variable is updated at a time (analogous to single
spin flips). The acceptance ratio is much higher this way than if we update all the
spins simultaneously. However, for some distributions the energy function may have
approximately marginal directions; directions in which there is little or no barrier.
The cooperative distribution has this property. We can expect critical slowing down
due to this and consequently some sort of collective update (analogous to multi-spin
updates or cluster updates) might make sampling more efficient. However, the type
of update will depend on the specifics of the energy function and is not easy to
determine.
356
7
N D. Socci, D. D. Lee and H. S. Seung
DISCUSSION
The competitive and cooperative distributions are examples of rectified Gaussians
for which no good approximation by a standard Gaussian is possible. However,
both distributions can be approximated by mixtures of standard Gaussians. The
competitive distribution can be approximated by a mixture of N Gaussians, one
for each singleton state. The cooperative distribution can also be approximated by
a mixture of N Gaussians, one for each location of the lump on the ring. A more
economical approximation would reduce the number of Gaussians in the mixture,
but .make each one anisotropic[IO].
Whether the rectified Gaussian is superior to these mixture models is an empirical
question that should be investigated empirically with specific real-world probabilistic modeling tasks. Our intuition is that the rectified Gaussian will turn out to be
a good representation for nonlinear pattern manifolds, and the aim of this paper
has been to make this intuition concrete.
To make the rectified Gaussian useful in practical applications, it is critical to
find tractable learning algorithms. It is not yet clear whether learning will be more
tractable for the rectified Gaussian than it was for the Boltzmann machine. Perhaps
the continuous variables of the rectified Gaussian may be easier to work with than
the binary variables of the Boltzmann machine.
Acknowledgments We would like to thank P. Mitra, L. Saul, B. Shraiman and
H. Sompolinsky for helpful discussions. Work on this project was supported by Bell
Laboratories, Lucent Technologies.
References
[1] D. H. Ackley, G. E. Hinton, and T. J. Sejnowski. A learning algorithm for
Boltzmann machines. Cognitive Science, 9:147-169, 1985.
[2] G. E. Hinton and Z. Ghahramani. Generative models for discovering sparse
distributed representations. Phil. Trans. Roy. Soc., B352:1177-90, 1997.
[3] Z. Ghahramani and G. E. Hinton. Hierarchical non-linear factor analysis and
topographic maps. Adv. Neural Info. Proc. Syst., 11, 1998.
[4] H. S. Seung. How the brain keeps the eyes still. Proc. Natl. Acad. Sci. USA,
93:13339-13344, 1996.
[5] R. Ben-Yishai, R. L. Bar-Or, and H. Sompolinsky. Theory of orientation tuning
in visual cortex. Proc. Nat. Acad. Sci. USA, 92:3844-3848, 1995.
[6] A. P. Georgopoulos, M. Taira, and A. Lukashin. Cognitive neurophysiology of
the motor cortex. Science, 260:47-52, 1993.
[7] K. Zhang. Representation of spatial orientation by the intrinsic dynamics of
the head-direction cell ensemble: a theory. J. Neurosci., 16:2112-2126, 1996.
[8] D. P. Bertsekas. Nonlinear programming. Athena Scientific, Belmont, MA,
1995.
[9] S. Amari and M. A. Arbib. Competition and cooperation in neural nets. In
J. Metzler, editor, Systems Neuroscience, pages 119-165. Academic Press, New
York, 1977.
[10] G. E. Hinton, P. Dayan, and M. Revow. Modeling the manifolds of images of
handwritten digits. IEEE Trans. Neural Networks, 8:65-74, 1997.
| 1402 |@word neurophysiology:1 version:3 polynomial:2 covariance:6 initial:1 cyclic:1 configuration:3 com:1 yet:1 must:6 belmont:1 wx:1 analytic:1 motor:2 designed:1 depict:1 update:5 stationary:5 generative:1 discovering:1 slowing:1 short:1 location:2 zhang:1 along:3 become:1 introduce:1 roughly:1 behavior:3 multi:1 brain:1 inspired:1 little:2 becomes:1 project:1 matched:2 underlying:1 finding:5 transformation:1 nj:1 xd:1 demonstrates:1 bertsekas:1 positive:3 before:1 local:3 mitra:1 limit:3 consequence:1 io:1 acad:2 analyzing:1 approximately:1 might:1 studied:2 suggests:1 co:1 bi:2 perpendicular:1 directed:1 unique:1 practical:2 acknowledgment:1 practice:1 block:1 definite:3 digit:1 procedure:2 empirical:1 bell:3 thought:1 word:1 get:1 cannot:4 interior:2 close:1 imposed:1 map:1 phil:1 convex:5 importantly:1 regarded:1 analogous:2 updated:1 programming:6 origin:1 roy:1 approximated:6 recognition:3 particularly:1 cut:2 metzler:1 cooperative:17 bottom:1 ackley:1 prefactor:1 capture:1 calculate:1 region:2 adv:1 sompolinsky:2 decrease:2 intuition:2 govern:1 complexity:1 seung:5 dynamic:5 depend:1 completely:2 multimodal:6 bowl:2 separated:1 monte:3 sejnowski:1 whose:1 quite:1 larger:1 valued:1 solve:1 amari:1 gi:2 topographic:1 eigenvalue:1 net:1 interaction:4 representational:3 intuitive:1 description:2 competition:2 cluster:1 diverges:1 produce:2 generating:1 ring:10 ben:1 object:4 illustrate:1 depending:1 coupling:1 ij:1 eq:1 strong:1 soc:1 involves:1 come:2 lyapunov:1 direction:7 concentrate:1 closely:1 centered:1 biological:1 singularity:2 ground:5 normal:1 fh:1 wine:1 proc:3 tool:2 gaussian:49 aim:1 ax:4 contrast:1 centroid:1 kiihn:2 helpful:1 dayan:1 provably:1 orientation:2 constrained:5 spatial:1 marginal:1 equal:2 never:1 shaped:2 having:1 sampling:7 lit:1 look:4 simultaneously:1 geometry:1 consisting:1 taira:1 attractor:1 interest:1 acceptance:1 possibility:1 mixture:6 natl:1 yishai:1 integral:1 unless:1 increased:3 modeling:5 earlier:1 deviation:1 uniform:1 dij:2 too:1 characterize:1 corrupted:1 density:1 peak:4 fundamental:1 lee:4 probabilistic:3 off:1 concrete:2 reflect:1 cognitive:2 leading:1 syst:1 potential:1 singleton:3 trough:5 satisfy:1 depends:1 lab:1 competitive:15 sort:1 oi:3 spin:3 ensemble:1 spaced:1 correspond:1 sitting:1 handwritten:1 accurately:1 carlo:3 economical:1 rectified:29 synaptic:1 energy:40 tucker:2 associated:1 higher:1 arranged:1 done:1 strongly:1 anywhere:1 nonlinear:3 lack:1 mode:14 brings:1 perhaps:2 resemblance:1 scientific:1 name:1 usa:2 y2:1 symmetric:1 laboratory:2 nonzero:1 illustrated:1 anything:1 hill:1 shraiman:1 temperature:9 image:2 recently:1 superior:1 rotation:1 empirically:1 winner:4 anisotropic:1 discussed:1 numerically:1 tuning:1 cortex:4 surface:1 curvature:1 certain:1 nonconvex:4 binary:1 minimum:21 converge:2 determine:1 multiple:3 full:1 academic:1 calculation:1 copositivity:1 basic:2 btx:1 ae:1 represent:3 bimodal:1 cell:1 rest:1 subject:1 undirected:2 lump:4 presence:1 iii:1 easy:1 xj:3 arbib:1 reduce:1 prototype:1 whether:3 hessian:1 york:1 generally:2 useful:2 clear:3 eigenvectors:1 simplest:1 reduced:1 generate:1 neuroscience:1 correctly:1 discrete:3 indefinite:1 delocalized:2 drawn:3 neither:1 lowering:1 inverse:1 angle:3 powerful:1 reasonable:2 separation:1 followed:2 guaranteed:1 quadratic:9 nonnegative:6 activity:5 constraint:5 infinity:1 precisely:1 georgopoulos:1 normalizable:2 x2:1 ri:1 span:1 according:1 representable:1 xixj:2 character:1 unity:2 metropolis:1 modification:3 happens:1 invariant:2 rectification:2 equation:1 previously:1 remains:1 turn:1 lukashin:1 flip:1 tractable:3 gaussians:6 hierarchical:1 shortly:1 original:1 graphical:2 const:1 ghahramani:3 murray:1 question:1 unclear:1 gradient:6 thank:1 sci:2 athena:1 manifold:6 toward:1 modeled:2 ratio:1 difficult:1 info:1 negative:4 collective:1 boltzmann:5 neuron:1 enabling:1 finite:3 orthant:3 curved:1 defining:1 hinton:5 variability:1 head:2 situation:1 introduced:1 bottle:1 accepts:1 trans:2 able:3 bar:1 dynamical:1 pattern:10 max:1 oj:2 explanation:1 power:3 critical:2 solvable:1 technology:2 brief:1 eye:1 unconditionally:1 deviate:1 expect:1 bear:1 permutation:1 interesting:2 viewpoint:1 editor:1 translation:2 normalizes:1 elsewhere:1 cooperation:1 supported:1 aij:3 side:1 saul:1 barrier:2 sparse:2 distributed:1 regard:1 boundary:1 world:1 contour:1 projected:1 approximate:1 keep:2 global:3 instantiation:1 xi:7 continuous:9 socci:4 investigated:1 neurosci:1 arrow:1 noise:1 n2:1 augmented:1 fig:3 nonnegativity:4 vanish:1 down:2 lucent:2 xt:1 specific:2 sit:1 intrinsic:1 flattened:1 nat:1 easier:1 saddle:5 visual:3 corresponds:2 relies:1 ma:1 consequently:1 price:1 revow:1 change:1 except:1 principal:1 invariance:1 accepted:1 arises:1 |
443 | 1,403 | Modeling Complex Cells in an A wake
Macaque During Natural Image Viewing
William E. Vinje
vinjeCsocrates.berkeley.edu
Department of Molecular and
Cellular Biology, Neurobiology Division
University of California, Berkeley
Berkeley, CA, 94720
Jack L. Gallant
gallantCsocrates.berkeley.edu
Department of Psychology
University of California, Berkeley
Berkeley, CA, 94720
Abstract
We model the responses of cells in visual area VI during natural
vision. Our model consists of a classical energy mechanism whose
output is divided by nonclassical gain control and texture contrast
mechanisms. We apply this model to review movies, a stimulus
sequence that replicates the stimulation a cell receives during free
viewing of natural images. Data were collected from three cells
using five different review movies, and the model was fit separately
to the data from each movie. For the energy mechanism alone we
find modest but significant correlations (rE = 0.41, 0.43, 0.59,
0.35) between model and data. These correlations are improved
somewhat when we allow for suppressive surround effects (rE+G =
0.42, 0.56, 0.60, 0.37). In one case the inclusion of a delayed
suppressive surround dramatically improves the fit to the data by
modifying the time course of the model's response.
1
INTRODUCTION
Complex cells in the primary visual cortex (area VI in primates) are tuned to
localized visual patterns of a given spatial frequency, orientation, color, and drift
direction (De Valois & De Valois, 1990). These cells have been modeled as linear
spatio-temporal filters whose output is rectified by a static nonlinearity (Adelson
& Bergen, 1985); more recent models have also included a divisive contrast gain
control mechanism (Heeger, 1992; Wilson & Humanski, 1993; Geisler & Albrecht,
1997). We apply a modified form of these models to a stimulus that simulates
natural vision. Our model uses relatively few parameters yet incorporates the cells'
temporal response properties and suppressive influences from beyond the classical
receptive field (C RF).
Modeling Complex Cells during Natural Image Viewing
2
237
METHODS
Data Collection: Data were collected from One awake behaving Macaque monkey,
using single unit recording techniques described elsewhere (Connor et al., 1997).1
First, the cell's receptive field size and location were estimated manually, and tuning curves were objectively characterized using two-dimensional sinusoidal gratings.
Next a static color image of a natural scene was presented to the animal and his
eye position was recorded continuously as he freely scanned the image for 9 seconds
(Gallant et al., 1998). 2 Image patches centered on the position of the cell's C RF
(and 2-4 times the CRF diameter) were then extracted using an automated procedure. The sequence of image patches formed a continuous 9 second review movie
that simulated all of the stimulation that had occurred in and around the C RF
during free viewing. 3 Although the original image was static, the review movies
contain the temporal dynamics of the saccadic eye movements made by the animal
during free viewing. Finally, the review movies were played in and around the C RF
while the animal performed a fixation task.
During free viewing each eye position is unique, so each image patch is likely to
enter the C RF only once. The review movies were therefore replayed several times
and the cell's average response with respect to the movie timestream was computed
from the peri-stimulus time histogram (PSTH). These review movies also form the
model's stimulus input, while its output is relative spike probability versus time
(the model cell's PSTH).
Before applying the model each review movie was preprocessed by converting to
gray scale (since the model does not consider color tuning), setting the average
luminance level to zero (on a frame by frame basis) and prefiltering with the human
contrast sensitivity function to more accurately reflect the information reaching cells
in VI.
Divisive Normalization Model: The model consists of a classical receptive field
energy mechanism, ECRF, whose output is divided by two nonclassical suppressive
mechanisms, a gain control field, G, and a texture contrast field, T.
ECRF(t)
PSTHmodel(t) ex 1 + Q G(t - d) + f3T(t - d)
(1)
We include a delay parameter for suppressive effects, consistent with the hypothesis
that these effects may be mediated by local cortical interactions (Heeger, 1992;
Wilson & Humanski, 1993). Any latency difference between the central energy
mechanism and the suppressive surround will be reflected as a positive delay offset
(15 > 0 in Equation 1).
Classical Receptive Field Energy Mechanism: The energy mechanism, ECRF,
is composed of four phase-dependent subunits, Uti>. Each subunit computes an inner
product in space and a convolution in time between the model cell's space-time
classical receptive field, CRFtI>(x, y, r), and the image, I(x, y, t).
U<P(t) =
JJJCRFtI>(x, y, r) . I(x, y, t - r) dx dydr
(2)
1 Recorcling was performed under a university-approved protocol and conformed to all
relevant NIH and USDA guidelines.
2 Images were taken from a Corel Corporation photo-CD library at 1280xl024 resolution.
3Eye position data were collected at 1 KHz, whereas the monitor clisplay rate was 72.5
Hz (14 ms per frame). Therefore each review movie frame was composed of the average
stimulation occurring during the corresponcling 13.8 ms of free viewing.
W. E. Vinje and 1. L Gallant
238
The model presented here incorporates the simplifying assumption of a space-time
separable receptive field structure, CRF4>(x, y, r) = CRF4>(x, y) CRF(r).
u4>(t)
= L: CRF(r) (L: L: CRF4>(x, y) . I(x, y, t T
X
r))
(3)
Y
Time is discretized into frames and space is discretized into pixels that match the
review movie input. CRF4>(x, y) is modeled as a sinusoidal grating that is spatially
weighted by a Gaussian envelope (i.e. a Gabor function). In this paper CRF(r)
is approximated as a delta function following a constant latency. This minimizes
model parameters and highlights the model's responses to the stimulus present at
each fixation. The latency, orientation and spatial frequency of the grating, and
the size of the C RF envelope, are all determined empirically by maximizing the fit
between model and data. 4
A static non-linearity ensures that the model PSTH does not become negative.
We have e~amined both half-wave rectification, fj4>(t) = max[U4>(t), O], and halfsquaring, U4>(t) = (max[U4>(t) , 0])2; here we present the results from half-wave
rectification. Half-squaring produces small changes in the model PSTH but does
not improve the fit to the data.
The energy mechanism is made phase invariant by averaging over the rectified phasedependent subunits:
(4)
Gain Control Field: Cells in V 1 incorporate a contrast gain control mechanism
that compensates for changes in local luminance. The gain control field, G, models
this effect as the total image power in a region encompassing the C RF and surround.
G(t-<5)
= L:CRF(r)
T
(L:L:VP(kx,ky,r) )
k%
(5)
ky
P(kx, ky, r) = F FT[PG(x, y, r)] F FT*[JlG(x, y, r)]
(6)
JlG(x, y, r) = vG(x, y) I(x, y, (t - <5) - r)
(7)
P(kx, ky, r) is the spatial Fourier power of JlG(x, y, r) and VG is a two dimensional
Gaussian weighting function whose width sets the size of the gain control field.
Heeger's (1992) divisive gain control term sums over many discrete energy mechanisms that tile space in and around the area of the C RF. Equation 5 approximates
Heeger's approach in the limiting case of dense tiling.
Texture Contrast Field: Cells in area VI can be affected by the image surrounding the region of the CRF (Knierim & Van Essen, 1992) . The responses of many
VI cells are highest when the optimal stimulus is presented alone within the CRF,
and lowest when that stimulus is surrounded with a texture of similar orientation
and frequency. The texture contrast field, T, models this effect as the image power
4 As a fit statistic we use the linear correlation coefficient (Pearson's r) between model
and data. Fitting is done with a gradient ascent algorithm. Our choice of correlation as a
statistic eliminates the need to explicitly consider model normalization as a variable, and is
very sensitive to latency mismatches between model and data. However, linear correlation
is more prone to noise contamination than is X2 ?
Modeling Complex Cells during Natural Image Viewing
239
in the spatial region surrounding the C RF that matches the C RF's orientation and
spatial frequency.
T(t-J) =
1
4
1: 1: CRF (r) "?1: Jp4>(kx,ky,r)
90,180,270[ (
4>=0
k",
T
lICRF(X,
(8)
ky
P4>(kx, ky, r) = F FT[p~(x, y, r)] F FT?[p~(x, y, r)]
J.t~(x, y, r) = ~*(x, y) (1 -
)]
y)) I(x, y, (t - J) - r)
(9)
(10)
~* is a Gabor function whose orientation and spatial frequency match those of the
best' fit C RF4> (x, y). The envelope of ~* defines the size of the texture contrast field.
lICRF is a two dimensional Gaussian weighting function whose width matches the
C RF envelope, and which suppresses the image center. Thus the texture contrast
term picks up oriented power from an annular region of the image surrounding the
C RF envelope. T is made phase invariant by averaging over phase.
3
RESULTS
Thus far our model has been evaluated on a small data set collected as part of
a different study (Gallant et ai., 1998). Two cells, 87A and 98C, were examined
with one review movie each, while cell 97 A was examined with three review movies.
Using this data set we compare the model's response in two interesting situations:
cell 97 A, which had high orientation-selectivity, versus cell 87 A, which had poor
orientation-selectivity; and cell 98C, which was directionally-selective, versus cell
97 A, which was not directionally-selective.
CRF Energy Mechanism: We separately fit the energy mechanism parameters
to each of the three different cells. For cell 97A the three review movies were fit
independently to test for consistency of the best fit parameters.
Table 1 shows the correlation between model and data using only the C RF energy
mechanism (a = f3 = 0 in Equation 1). The significance of the correlations was
assessed via a permutation test. The correlation values for cells 97 A and 98C,
though modest, are significant (p < 0.01). For these cells the 95% confidence
intervals on the best fit parameter values are consistent with estimates from the
flashed grating tests. The best fit parameter values for cell 97 A are also consistent
across the three independently fit review movies.
The model best accounts for the data from cell 97 A. This cell was highly selective for
vertical gratings and was not directionally-selective. Figure 1 compares the PSTH
obtained from cell 97 A with movie B to the model PSTH. The model generally
responds to the same features that drive the real cell, though the match is imperfect.
Much of the discrepancy between the model and data arises from our approximation
of CRF(t) as a delta function. The model's response is roughly constant during
97A
87A
97A
97A
98C
Cell
B
A
C
A
A
Movie
Yes
Yes
Yes
Yes
Oriented
No
Yes
No
No
No
Directional
No
0.35
NA
0.41
0.43
0.59
rE
Table 1: Correlations between model and data PSTHs. Oriented cells showed
orientation-selectivity in the flashed grating test while Directional cells showed
directional-selectivity during manual characterization. rE is the correlation between
ECRF and the data. No fit was obtained for cell 87 A.
W. E. Vinje and J L. Gallant
240
1~--~~--~----~----~----T-----~--~----~----~
.-cO.8
.-.
~
~
? 0.6
~
~
'a
~ 0.4
~
.~
.-.
~
~
0.2
1
2
3
4
5
Time (seconds)
6
7
8
Figure 1: CRF energy mechanism versus data (Cell 97A, Movie B) . White indicates that
the model response is greater than the data, while black indicates the data is greater than
the model and gray indicates regions of overlap. A perfect match between model and data
would result in the entire area under the curve being gray. Our approximation of CRF(t)
leads to a relatively constant model PSTH during each fixation. In contrast the real cell
generally gives a phasic response as each saccade brings a new stimulus into the CRF. In
general the same movie features drive both model and cell.
each fixation, which causes the model PSTH to appear stepped. In contrast the
data PSTH shows a strong phasic response at the beginning of each fixation when
a new stimulus patch enters the cell's CRF .
The model is less successful at accounting for the responses of the directionallyselective cell, 98C. This is probably because the model's space-time separable receptive field misses motion energy cues that drive the cell. The model completely
failed to fit the data from cell 87 A . This cell was not orientation-selective, so the
fitting procedure was unable to find an appropriate orientation for the CRF?(x, y)
Gabor function. 5
CRF Energy Mechanism with Suppressive Surround: Table 2 lists the improvements in correlation obtained by adding the gain control term (a > 0, fJ = 0
in Equation 1). For cell 97 A (all three movies) the best correlations are obtained
when the surround effects are delayed by 56 ms relative to the center. The best
correlation for cell 98C is obtained when the surround is not delayed.
In three out of four cases the correlation values are barely improved when the
surround effects are included, suggesting that the cells were not strongly surroundinhibited by these review movies. However, the improvement is quite striking in the
SFor cell 87 A the correlation values in the orientation and spatial frequency parameter
subspace contained three roughly equivalent maxima. Contamination by multiple cells
was unlikely due to this cell's excellent isolation.
9
Modeling Complex Cells during Natural Image Viewing
Cell
Movie
97A
A
0.42
+0.01
rE+G
~r
97A
B
0.56
+0.13
241
97A
C
0.60
+0.01
98C
A
0.37
+0.02
Table 2: Correlation improvements due to surround gain control mechanism. rE+G
gives the correlation value between the best fit model and the data. ~r gives the
improvement over rEo Including G in Equation 1 leads to a dramatic correlation
increase for cell 97 A, movie B, but not for the other review movies.
case of cell 97 A, movie B. Figure 2 compares the data with a model using both Ecr f
and G in Equation 1. Here the delayed surround suppresses the sustained responses
seen in Figure 1 and results in a more phasic model PSTH that closely matches the
data.
We consider G and T fields both independently and in combination. For each
we independently fit for Q, {3, &, and the size of the suppressive fields. However,
the oriented Fourier power correlates with the total Fourier power for our sample of
natural images, so that G and T are highly correlated. Combined fitting of G and T
terms leads to competition and dominance by G (i.e. (3 -r 0). In this paper we only
report the effects of the gain control mechanism; the texture contrast mechanism
results in similar (though slightly degraded) results.
1~--~----~----~----~----~----~---.----~----~
cO.8
.....
......
.....
~
~
ct 0.6
]
.....
0-
~ 0.4
.....>
~
......
(I)
0::: 0.2
o
o
1
2
3
456
Time (seconds)
7
8
Figure 2: C RF energy mechanism with delayed surround gain control versus data (Cell
97A, Movie B). Color scheme as in Figure 1. The inclusion of the delayed G term results
in a more phasic model response which greatly improves the match between model and
data.
9
W. E. Vinje and 1. L. Gallant
242
4
DISCUSSION
This preliminary study suggests that models of the form outlined here show great
promise for describing the responses of area V1 cells during natural vision. For
comparison consider the correlation values obtained from an earlier neural network
model that attempted to reproduce V1 cells' responses to a variety of spatial patterns (Lehky et al. 1992) . They report a median correlation value of 0.65 for
complex stimuli, whereas the average correlation score from Table 2 is 0.49. This is
remarkable considering that our model has only 7 free parameters, a very hmited
data set for fitting, doesn't yet consider color tuning or directional-selectivity and
considers response across time.
Future implementations of the model will use a more sophisticated energy mechanism that allows for nonseparable space time receptive field structure and more
realistic temporal response dynamics. We will also incorporate more detail into
the surround mechanisms, such as asymmetric surround structure and a broadband
texture contrast term.
By abstracting physiological observation into approximate functional forms our
model balances explanatory power against parametric complexity. A cascaded series
of these models may form the foundation for future modeling of cells in extra-striate
areas V2 and V4. Natural image stimuli may provide an appropriate stimulus set
for development and validation of these extrastriate models.
Acknowledgements
We thank Joseph Rogers for assistance in this study, Maneesh Sahani for the extremely useful suggestion of fitting the CRF parameters, Charles Connor for help
with data collection and David Van Essen for support of data collection.
References
Adelson, E. H. & Bergen, J. R. (1985) Spatiotemporal energy models for the perception of motion. Journal of the Optical Society of America, A, 2, 284-299.
Connor, C. C., Preddie, D. C., Gallant, J . L. & Van Essen, D. C. (1997) Spatial
attention effects in macaque area V4 . Journal of Neuroscience, 77, 3201-3214.
De Valois, R. L. & De Valois, K. K. (1990) Spatial Vision. New York: Oxford
University Press.
Gallant, J. L., Connor, C. E., & Van Essen, D. C. (1998) Neural Activity in Areas
V1 , V2 and V4 During Free Viewing of Natural Scenes Compared to Controlled
Viewing. NeuroReport, 9 .
Geisler, W. S., Albrecht, D. G. (1997) Visual cortex neurons in monkeys and cats:
Detection, discrimination, and identification. Visual Neuroscience, 14, 897-919.
Heeger, D. J. (1992) Normalization of cell responses in cat striate cortex. Visual
Neuroscience, 9, 181-198.
Knierim, J . J . & Van Essen, D. C. (1992) Neuronal responses to static texture
patterns in area V1 of the alert macaque monkey. Journal of Neurophysiology, 67,
961-980.
Lehky, S. R., Sejnowski, T . J . & Desimone, R. (1992) Predicting Responses of
Nonlinear Neurons in Monkey Striate Cortex to Complex Patterns. Journal of
Neuroscience, 12, 3568-3581.
Wilson, H. R. & Humanski, R. (1993) Spatial frequency adaptation and contrast
gain control. Vision Research, 33, 1133-1149.
| 1403 |@word neurophysiology:1 approved:1 simplifying:1 accounting:1 pg:1 pick:1 dramatic:1 extrastriate:1 valois:4 series:1 score:1 tuned:1 yet:2 dx:1 realistic:1 discrimination:1 alone:2 half:3 cue:1 beginning:1 characterization:1 location:1 psth:10 five:1 alert:1 become:1 consists:2 fixation:5 sustained:1 fitting:5 roughly:2 nonseparable:1 discretized:2 considering:1 linearity:1 lowest:1 minimizes:1 monkey:4 suppresses:2 corporation:1 temporal:4 berkeley:6 control:13 unit:1 appear:1 before:1 positive:1 local:2 oxford:1 black:1 examined:2 suggests:1 co:2 unique:1 procedure:2 area:10 maneesh:1 gabor:3 confidence:1 influence:1 applying:1 equivalent:1 center:2 maximizing:1 attention:1 independently:4 resolution:1 his:1 limiting:1 us:1 hypothesis:1 approximated:1 asymmetric:1 u4:4 ft:4 enters:1 region:5 ensures:1 movement:1 highest:1 contamination:2 complexity:1 dynamic:2 division:1 basis:1 completely:1 cat:2 america:1 surrounding:3 sejnowski:1 pearson:1 whose:6 quite:1 compensates:1 objectively:1 statistic:2 conformed:1 directionally:3 sequence:2 nonclassical:2 interaction:1 product:1 p4:1 adaptation:1 relevant:1 competition:1 ky:7 produce:1 perfect:1 help:1 ex:1 strong:1 grating:6 direction:1 closely:1 modifying:1 filter:1 centered:1 human:1 viewing:11 rogers:1 preliminary:1 around:3 great:1 sensitive:1 weighted:1 gaussian:3 modified:1 reaching:1 wilson:3 improvement:4 indicates:3 greatly:1 contrast:14 bergen:2 dependent:1 squaring:1 entire:1 unlikely:1 explanatory:1 selective:5 reproduce:1 pixel:1 orientation:11 development:1 animal:3 spatial:11 field:18 once:1 f3:1 manually:1 biology:1 adelson:2 discrepancy:1 future:2 report:2 stimulus:12 few:1 oriented:4 composed:2 delayed:6 phase:4 william:1 detection:1 essen:5 highly:2 replicates:1 desimone:1 modest:2 re:6 modeling:5 earlier:1 f3t:1 delay:2 successful:1 spatiotemporal:1 combined:1 peri:1 geisler:2 sensitivity:1 v4:3 continuously:1 na:1 reflect:1 recorded:1 central:1 tile:1 albrecht:2 account:1 suggesting:1 sinusoidal:2 de:4 coefficient:1 explicitly:1 vi:5 performed:2 wave:2 formed:1 degraded:1 yes:5 vp:1 directional:4 identification:1 accurately:1 rectified:2 drive:3 manual:1 against:1 energy:17 frequency:7 static:5 gain:13 color:5 improves:2 sophisticated:1 reflected:1 response:21 improved:2 replayed:1 done:1 evaluated:1 though:3 strongly:1 correlation:21 receives:1 nonlinear:1 defines:1 brings:1 gray:3 effect:9 contain:1 spatially:1 flashed:2 white:1 assistance:1 during:15 width:2 m:3 crf:17 motion:2 fj:1 image:20 jack:1 charles:1 nih:1 psths:1 stimulation:3 functional:1 empirically:1 corel:1 khz:1 he:1 occurred:1 approximates:1 significant:2 surround:13 connor:4 enter:1 ai:1 tuning:3 consistency:1 outlined:1 inclusion:2 nonlinearity:1 had:3 cortex:4 behaving:1 recent:1 showed:2 selectivity:5 seen:1 greater:2 somewhat:1 freely:1 converting:1 multiple:1 annular:1 match:8 characterized:1 divided:2 molecular:1 controlled:1 vision:5 histogram:1 normalization:3 cell:60 whereas:2 separately:2 interval:1 wake:1 median:1 suppressive:8 envelope:5 eliminates:1 extra:1 ascent:1 probably:1 recording:1 hz:1 simulates:1 incorporates:2 automated:1 variety:1 fit:16 psychology:1 isolation:1 inner:1 imperfect:1 york:1 cause:1 dramatically:1 generally:2 latency:4 useful:1 lehky:2 diameter:1 estimated:1 delta:2 per:1 neuroscience:4 discrete:1 promise:1 affected:1 dominance:1 four:2 monitor:1 preprocessed:1 luminance:2 v1:4 sum:1 striking:1 uti:1 patch:4 ct:1 played:1 activity:1 scanned:1 awake:1 scene:2 x2:1 fourier:3 extremely:1 separable:2 optical:1 relatively:2 department:2 combination:1 poor:1 across:2 slightly:1 joseph:1 primate:1 invariant:2 taken:1 rectification:2 equation:6 describing:1 mechanism:23 phasic:4 photo:1 tiling:1 apply:2 v2:2 appropriate:2 humanski:3 original:1 include:1 classical:5 society:1 spike:1 receptive:8 primary:1 saccadic:1 parametric:1 responds:1 striate:3 gradient:1 subspace:1 unable:1 thank:1 simulated:1 stepped:1 collected:4 cellular:1 considers:1 barely:1 modeled:2 balance:1 negative:1 implementation:1 guideline:1 gallant:8 vertical:1 convolution:1 observation:1 neuron:2 subunit:3 neurobiology:1 situation:1 frame:5 knierim:2 drift:1 david:1 california:2 macaque:4 beyond:1 pattern:4 mismatch:1 perception:1 reo:1 rf:14 max:2 including:1 power:7 overlap:1 natural:12 predicting:1 cascaded:1 scheme:1 improve:1 movie:27 eye:4 library:1 mediated:1 sahani:1 review:16 acknowledgement:1 relative:2 encompassing:1 highlight:1 permutation:1 abstracting:1 interesting:1 suggestion:1 vinje:4 localized:1 versus:5 vg:2 remarkable:1 foundation:1 validation:1 consistent:3 usda:1 surrounded:1 cd:1 prone:1 course:1 elsewhere:1 free:7 allow:1 van:5 curve:2 cortical:1 prefiltering:1 computes:1 doesn:1 collection:3 made:3 far:1 correlate:1 approximate:1 spatio:1 continuous:1 table:5 ca:2 excellent:1 complex:7 protocol:1 significance:1 dense:1 noise:1 neuronal:1 broadband:1 position:4 heeger:5 weighting:2 offset:1 list:1 physiological:1 adding:1 texture:10 occurring:1 kx:5 likely:1 visual:6 failed:1 contained:1 saccade:1 extracted:1 change:2 included:2 determined:1 averaging:2 miss:1 preddie:1 total:2 divisive:3 attempted:1 support:1 arises:1 assessed:1 incorporate:2 neuroreport:1 correlated:1 |
444 | 1,404 | Reinforcement Learning for Continuous
Stochastic Control Problems
Remi Munos
CEMAGREF, LISC, Pare de Tourvoie,
BP 121, 92185 Antony Cedex, FRANCE.
Rerni.Munos@cemagref.fr
Paul Bourgine
Ecole Polyteclmique, CREA,
91128 Palaiseau Cedex, FRANCE.
Bourgine@poly.polytechnique.fr
Abstract
This paper is concerned with the problem of Reinforcement Learning (RL) for continuous state space and time stocha.stic control
problems. We state the Harnilton-Jacobi-Bellman equation satisfied by the value function and use a Finite-Difference method for
designing a convergent approximation scheme. Then we propose a
RL algorithm based on this scheme and prove its convergence to
the optimal solution.
1
Introduction to RL in the continuous, stochastic case
The objective of RL is to find -thanks to a reinforcement signal- an optimal strategy
for solving a dynamical control problem. Here we sudy the continuous time, continuous state-space stochastic case, which covers a wide variety of control problems
including target, viability, optimization problems (see [FS93], [KP95])}or which a
formalism is the following. The evolution of the current state x(t) E 0 (the statespace, with 0 open subset of IRd ), depends on the control u(t) E U (compact subset)
by a stochastic differential equation, called the state dynamics:
(1)
dx = f(x(t), u(t))dt + a(x(t), u(t))dw
where f is the local drift and a .dw (with w a brownian motion of dimension rand
(j a d x r-matrix) the stochastic part (which appears for several reasons such as lake
of precision, noisy influence, random fluctuations) of the diffusion process.
For initial state x and control u(t), (1) leads to an infinity of possible traj~tories
x(t). For some trajectory x(t) (see figure I)., let T be its exit time from 0 (with
the convention that if x(t) always stays in 0, then T = 00). Then, we define the
functional J of initial state x and control u(.) as the expectation for all trajectories
of the discounted cumulative reinforcement :
J(x; u(.)) =
Ex,u( .)
{loT '/r(x(t), u(t))dt +,,{ R(X(T))}
R. Munos and P. Bourgine
1030
where rex, u) is the running reinforcement and R(x) the boundary reinforcement.
'Y is the discount factor (0 :S 'Y < 1). In the following, we assume that J, a are of
class C2 , rand Rare Lipschitzian (with constants Lr and LR) and the boundary
80 is C2 .
?
? all ?
II
?
?
xirJ
?
?
?
?
?
?
?
?
Figure 1: The state space, the discretized ~6 (the square dots) and its frontier 8~6
(the round ones). A trajectory Xk(t) goes through the neighbourhood of state ~.
RL uses the method of Dynamic Program~ing (DP) which generates an optimal
(feed-back) control u*(x) by estimating the value function (VF), defined as the
maximal value of the functional J as a function of initial state x :
(2)
Vex) = sup J(x; u(.).
u( .)
In the RL approach, the state dynamics is unknown from the system ; the only
available information for learning the optimal control is the reinforcement obtained
at the current state. Here we propose a model-based algorithm, i.e. that learns
on-line a model of the dynamics and approximates the value function by successive
iterations.
Section 2 states the Hamilton-Jacobi-Bellman equation and use a Finite-Difference
(FD) method derived from Kushner [Kus90] for generating a convergent approximation scheme. In section 3, we propose a RL algorithm based on this scheme and
prove its convergence to the VF in appendix A.
2
A Finite Difference scheme
Here, we state a second-order nonlinear differential equation (obtained from the DP
principle, see [FS93J) satisfied by the value function, called the Hamilton-JacobiBellman equation.
Let the d x d matrix a = a.a' (with' the transpose of the matrix). We consider
the uniformly pambolic case, Le. we assume that there exists c > 0 such that
V$ E 0, Vu E U, Vy E IR d ,2:t,j=l aij(x, U)YiYj 2: c1lY112. Then V is C2 (see [Kry80J).
Let Vx be the gradient of V and VXiXj its second-order partial derivatives.
Theorem 1 (Hamilton-Jacohi-Bellman) The following HJB equation holds :
Vex) In 'Y + sup [rex, u)
uEU
+ Vx(x).J(x, u) + ! 2:~j=l aij V
XiXj
(x)] = 0 for x E 0
Besides, V satisfies the following boundary condition: Vex) = R(x) for x E 80.
1031
Reinforcement Learningfor Continuous Stochastic Control Problems
Remark 1 The challenge of learning the VF is motivated by the fact that from V,
we can deduce the following optimal feed-back control policy:
u*(x) E arg sup [r(x, u)
uEU
+ Vx(x).f(x, u) + ! L:7,j=l aij VXiXj (x)]
In the following, we assume that 0 is bounded. Let eI, ... , ed be a basis for JRd.
Let the positive and negative parts of a function 4> be : 4>+ = ma.x(4),O) and
4>- = ma.x( -4>,0). For any discretization step 8, let us consider the lattices: 8Zd =
{8. L:~=1 jiei} where j}, ... ,jd are any integers, and ~6 = 8Z d n O. Let 8~6, the
frontier of ~6 denote the set of points {~ E 8Zd \ 0 such that at least one adjacent
point ~ ? 8ei E ~6} (see figure 1).
Let U 6 cUbe a finite control set that approximates U in the sense: 8 ~ 8' =>
U6' C U6 and U6U 6 = U. Besides, we assume that: Vi = l..d,
(3)
By replacing the gradient Vx(~) by the forward and backward first-order finitedifference quotients: ~;, V(~) = l [V(~ ? 8ei) - V(~)l and VXiXj (~) by the secondorder finite-difference quotients:
-b [V(~ + 8ei) + V(,'
~XiXi V(~)
~;iXj V(~)
- 8ei) - 2V(O]
= 2P[V(~ + 8ei ? 8ej) + V(~ - 8ei =F 8ej)
-V(~
+ 8ei) -
V(~
- 8ei) -
V(~
+ 8ej) -
V(~
- 8ej)
+ 2V(~)]
in the HJB equation, we obtain the following : for ~ E :?6,
V6(~)In,+SUPUEUh {r(~,u) + L:~=1 [f:(~,u)'~~iV6(~) - fi-(~,U)'~;iV6(~)
+
aii (~.u) ~ X,X,'"
V(C) + "
. .
2
. .
wJ'l=~
(at; (~,'U) ~ +x,x.1'"
.V(C) _ a~ (~,'U) ~ x,x V(C))] } = 0
2
2
-. .
J
'"
Knowing that (~t In,) is an approximation of (,l:l.t -1) as ~t tends to 0, we deduce:
V6(~)
with
T(~,
SUPuEUh
[,"'(~'U)L(EEbP(~,U,()V6?()+T(~,u)r(~,u)] (4)
(5)
u)
which appears as a DP equation for some finite Markovian Decision Process (see
[Ber87]) whose state space is ~6 and probabilities of transition:
p(~,u,~
p(~,
? 8ei)
u, ~ + 8ei ? 8ej)
p(~,u,~
- 8ei ? 8ej)
p(~,u,()
"'~~r) [28Ift(~, u)1 + aii(~' u) - Lj=l=i laij(~, u)l] ,
"'~~r)a~(~,u)fori=f:j,
(6)
"'~~r)a~(~,u) for i =f: j,
o otherwise.
Thanks to a contraction property due to the discount factor" there exists a unique
solution (the fixed-point) V to equation (4) for ~ E :?6 with the boundary condition
V6(~) = R(~) for ~ E 8:?6. The following theorem (see [Kus90] or [FS93]) insures
that V 6 is a convergent approximation scheme.
R. Munos and P. Bourgine
1032
Theorem 2 (Convergence of the FD scheme) V D converges to V as 8 1 0 :
lim
VD(~)
/)10
~-x
= Vex)
un~formly on 0
Remark 2 Condition (3) insures that the p(~, u, () are positive. If this condition
does not hold, several possibilities to overcome this are described in [Kus90j.
3
The reinforcement learning algorithm
Here we assume that f is bounded from below. As the state dynami,:s (J and a)
is unknown from the system, we approximate it by building a model f and from
samples of trajectories Xk(t) : we consider series of successive states Xk = Xk(tk)
and Yk = Xk(tk + Tk) such that:
a
- "It E [tk, tk + Tk], x(t) E N(~) neighbourhood of ~ whose diameter is inferior to
kN.8 for some positive constant kN,
- the control u is constant for t E [tk, tk
- Tk
+ Tk],
satisfies for some positive kl and k2,
(7)
Then incrementally update the model :
.1 ",n
n ~k=l
n (Yk - Xk -;;; Lk=l
1
an(~,u)
Yk - Xk
Tk
Tk.fn(~, u))
and compute the approximated time T( x, u)
transition p(~, u, () by replacing f and a by
(Yk - Xk Tk
Tk?fn(~, u))'
(8)
~d
f
the approximated probabilities of
and a in (5) and (6).
We obtain the following updating rule of the V D-value of state ~ :
V~+l (~)
=
sUPuEU/)
[,~/:(x,u) L( p(~, u, ()V~(() + T(x, u)r(~, u)]
(9)
which can be used as an off-line (synchronous, Gauss-Seidel, asynchronous) or ontime (for example by updating V~(~) as soon as a trajectory exits from the neighbourood of ~) DP algorithm (see [BBS95]).
Besides, when a trajectory hits the boundary [JO at some exit point Xk(T) then
update the closest state ~ E [JED with:
(10)
Theorem 3 (Convergence of the algorithm) Suppose that the model as well
as the V D-value of every state ~ E :ED and control u E U D are regularly updated
(respectively with (8) and (9)) and that every state ~ E [JED are updated with (10)
at least once. Then "Ie> 0, :3~ such that "18 ~ ~, :3N, "In 2: N,
sUP~EE/) IV~(~) - V(~)I ~
e
with probability 1
1033
Reinforcement Learningfor Continuous Stochastic Control Problems
4
Conclusion
This paper presents a model-based RL algorithm for continuous stochastic control
problems. A model of the dynamics is approximated by the mean and the covariance
of successive states. Then, a RL updating rule based on a convergent FD scheme is
deduced and in the hypothesis of an adequate exploration, the convergence to the
optimal solution is proved as the discretization step 8 tends to 0 and the number
of iteration tends to infinity. This result is to be compared to the model-free RL
algorithm for the deterministic case in [Mun97]. An interesting possible future
work should be to consider model-free algorithms in the stochastic case for which a
Q-Iearning rule (see [Wat89]) could be relevant.
A
Appendix: proof of the convergence
Let M f ' M a, M fr. and Ma .? be the upper bounds of j, a, f x and 0' x and m f the lower
bound of f. Let EO = SUP?EI:h !V0 (';) - V(';)I and E! = SUP?EI:b \V~(';) - VO(.;)\.
Estimation error of the model fn and an and the probabilities Pn
A.I
Suppose that the trajectory Xk(t) occured for some occurence Wk(t) of the brownian
motion: Xk(t) = Xk + f!k f(Xk(t),u)dt + f!" a(xk(t),U)dwk. Then we consider a
trajectory Zk (t) starting from .; at tk and following the same brownian motion:
Zk(t) ='; + fttk. f(Zk(t), u)dt + fttk a(zk(t), U)dWk'
Let Zk = Zk(tk + Tk). Then (Yk - Xk) - (Zk -.;) =
ftk [f(Xk(t), u) - f(Zk(t), u)] dt +
t
Tk
ft
[a(xk(t), u) - a(zk(t), u)J dWk. Thus, from the C1 property of f and a,
:.+
II(Yk -
Xk) - (Zk - ';)11 ~ (Mf'"
+ M aJ.kN.Tk. 8.
(11)
The diffusion processes has the following property ~ee for example the ItO- Taylor
majoration in [KP95j) : Ex [ZkJ = ';+Tk.f(';, U)+O(Tk) which, from (7), is equivalent
to: Ex [z~:g]
= j(';,u) + 0(8). Thus from the law of large numbers and (11):
li~-!~p Ilfn(';, u) -
f(';, u)11
-
li;;:s~p II~ L~=l [Yk;kX
& -
(Mf:r:
?.] I + 0(8)
+ M aJ?kN?8 + 0(8) =
0(8) w.p. 1 (12)
Besides, diffusion processes have the following property (again see [KP95J):
Ex [(Zk -.;) (Zk - .;)'] = a(';, U)Tk + f(';, u).f(';, U)'.T~ + 0(T2) which, from (7),
is equivalent to: Ex [(Zk-?-Tkf(S'U)~(kZk-S-Tkf(S'U?/] = a(';, u) + 0(82). Let rk =
= Yk - Xk - Tkfn(';, u) which satisfy (from (11) and (12? :
Ilrk - ikll = (Mf:r: + M aJ.Tk.kN.8 + Tk.o(8)
(13)
Zk -.; - Tkf(';, u) and ik
From the definition of Ci;;(';,u), we have: Ci;;(';,u) - a(';,u)
Ex [r~':k]
+ 0(8 2 )
li~~~p
=
~L~=l
'\:1.' -
and from the law of large numbers, (12) and (13), we have:
11~(';,u) -
a(';,u)11 =
Ilik -rkllli:!s!p~
li~-!~p II~ L~=l rJ./Y
fl (II~II + II~II)
- r~':k I + 0(8
+0(82 ) = 0(8 2 )
2)
R. Munos and P. Bourgine
1034
"In(';, u) - I(';, u)" ~
1Ill;;(';, u) - a(';, u)1I
kf?8 w.p. 1
~ ka .82 w.p. 1
(14)
Besides, from (5) and (14), we have:
2
2
c ) _ -Tn (cr.",U )1 _
< d.(k[.6
kT'UJ:2
1T(r.",U
(d.m,+d.k,,6
.6)2 ) UJ:2 <
_
(15)
and from a property of exponential function,
I,T(~.u) _
,7' . .1?) I= kT.In ~ .8
(?
2.
(16)
We can deduce from (14) that:
.
1(
)
-(
limsupp';,u,( -Pn';,u,(
)1
~
n-+oo
(2.6.Mt+d.Ma)(2.kt+d.k,,)62
k J:
6mr(2.k,+d.ka)62
S; puw.p.l
Estimation of IV~+l(';) - V 6(.;) 1
A.2
Mter having updated V~(';) with rule (9),
IV~+l(';) - V 6(.;) From (4), (9) and (8),
I.
A
(17)
let A denote the difference
< ,T(?.U) L: [P(';, u, () - p(.;, u, ()] V 6(() + (,T(?.1?) - ,7'(~'1??) L p(.;, u, () V 6 (()
(
(
+,7' (?.u) . L:p(.;, u, () [V6(() - V~(()] + L:p(.;, u, ().T(';, u) [r(';, u) - F(';, u)]
(
(
+ L:( p(.;, u, () [T(';, u) -
T(';, u)] r(';, u) for all u E U6
As V is differentiable we have : Vee) = V(';) + VX ' ( ( - . ; ) + 0(1I( - ';11). Let
us define a linear function V such that: Vex) = V(';) + VX ' (x - ';). Then
we have: [P(';, u, () - p(.;, u, ()] V 6(() = [P(';, u, () - p(.;, u, ()] . [V6(() - V(()] +
[P(';,u,()-p(';,u,()]V((), thus: L:([p(';,u,()-p(';,u,()]V6(() = kp .E6.8 +
[V(() +0(8)] = [V(7J)-VUD] + kp .E6.8 + 0(8)
with: 7J = L:( p(';, u, () (( -.;) and 1j = L:( p(.;, u, () (( -
.;).
Besides, from the convergence of the scheme (theorem 2), we have E6.8
=
L([P(';,U,()-p(.;,u,()]
[V(7J) - V(1j)] + 0(8)
0(8). From the linearity of
V,
IL( [P(';, u, () - p(.;, u, ()] V6 (() I
IV(() -
I II( - ZII?Mv",
V(Z) ~
=
S; 2kp 82 . Thus
= 0(8) and from (15), (16) and the Lipschitz prop-
erty of r,
A=
1'l'(?'U), L:( p(.;, u, () [V6(() - V~ (()] 1+ 0(8).
As ,..,.7'(?.u) < 1 - 7'(?.U) In 1 < 1 _ T(?.u)-k.,.6 2 In 1 < 1 _ (
I
-
2
'Y -
2
we have:
A
with k
= 2d(M[~d.M,,).
= (1 -
k.8)E~
'Y -
+ 0(8)
6
2d(M[+d.M,,)
_
!ix..82) In 1
2
'Y '
(18)
1035
Reinforcement Learning for Continuous Stochastic Control Problems
A.3
A sufficient condition for sUP?EE~ IV~(~) - V6(~)1 :S
C2
Let us suppose that for all ~ E ~6, the following conditions hold for some a > 0
E~ > C2 =} IV~+I(O - V6(~)1 :S E~ - a
(19)
E~ :S c2=}IV~+I(~)_V6(~)I:Sc2
(20)
From the hypothesis that all states ~ E ~6 are regularly updated, there exists an
integer m such that at stage n + m all the ~ E ~6 have been updated at least
once since stage n. Besides, since all ~ E 8C 6 are updated at least once with
rule (10), V~ E 8C6, IV~(~) - V6(~)1 = IR(Xk(T)) - R(~)I :S 2.LR.8 :S C2 for any
8 :S ~3 = 2~lR' Thus, from (19) and (20) we have:
E! >
E! :S
C2 =}
E!+m :S E! - a
C2 =}
E!+m :S
Thus there exists N such that: Vn ~ N, E~ :S
A.4
C2
C2.
Convergence of the algorithm
Let us prove theorem 3. For any c > 0, let us consider Cl > 0 and C2 > 0 such that
= c. Assume E~ > ?2, then from (18), A = E! - k.8'?2+0(8) :S E~ -k.8.~
for 8 :S ~3. Thus (19) holds for a = k.8.~. Suppose now that E~ :S ?2. From (18),
A :S (1 - k.8)?2 + 0(8) :S ?2 for 8 :S ~3 and condition (20) is true.
Cl +C2
Thus for 8 :S min {~1, ~2, ~3}, the sufficient conditions (19) and (20) are satisfied.
So there exists N, for all n ~ N, E~ :S ?2. Besides, from the convergence of the
scheme (theorem 2), there exists ~o st. V8:S ~o, sUP?EE~ 1V6(~) - V(~)I :S ?1?
Thus for 8 :S
min{~o, ~1, ~2, ~3},
"3N, Vn
~
sup IV~(~) - V(~)I :S sup IV~(~) - V6(~)1
?EE6
?EEh
N,
+ sup
?EE6
1V6(~) - V(~)I :S
?1
+ c2 = ?.
References
[BBS95j Andrew G. Barto, Steven J. Bradtke, and Satinder P. Singh. Learning to
act using real-time dynamic programming. Artificial Intelligence, (72):81138, 1995.
[Ber87j Dimitri P. Bertsekas. Dynamic Programming: Deterministic and Stochastic Models. Prentice Hall, 1987.
[FS93j Wendell H. Fleming and H. Mete Soner. Controlled Markov Processes and
Viscosity Solutions. Applications of Mathematics. Springer-Verlag, 1993.
[KP95j Peter E. Kloeden and Eckhard Platen. Numerical Solutions of Stochastic
Differential Equations. Springer-Verlag, 1995.
[Kry80j N.V. Krylov. Controlled Diffusion Processes. Springer-Verlag, New York,
1980.
[Kus90j Harold J. Kushner. Numerical methods for stochastic control problems in
continuous time. SIAM J. Control and Optimization, 28:999-1048, 1990.
[Mun97j Remi Munos. A convergent reinforcement learning algorithm in the continuous case based on a finite difference method. International Joint Conference on Art~ficial Intelligence, 1997.
[Wat89j Christopher J.C.H. Watkins. Learning from delayed reward. PhD thesis,
Cambridge University, 1989.
| 1404 |@word open:1 kus90j:2 covariance:1 contraction:1 bourgine:5 initial:3 series:1 ecole:1 kry80j:2 current:2 discretization:2 ixj:1 ka:2 dx:1 fn:3 numerical:2 update:2 intelligence:2 ficial:1 xk:20 lr:4 finitedifference:1 successive:3 c6:1 zii:1 c2:14 differential:3 zkj:1 ik:1 prove:3 hjb:2 discretized:1 bellman:3 discounted:1 ee6:2 estimating:1 bounded:2 linearity:1 erty:1 every:2 act:1 iearning:1 k2:1 hit:1 control:19 hamilton:3 bertsekas:1 positive:4 local:1 tends:3 fluctuation:1 tory:1 unique:1 vu:1 prentice:1 influence:1 equivalent:2 deterministic:2 go:1 starting:1 rule:5 u6:3 dw:2 updated:6 target:1 suppose:4 programming:2 us:1 designing:1 hypothesis:2 secondorder:1 approximated:3 updating:3 ft:1 steven:1 wj:1 yk:8 reward:1 dynamic:7 singh:1 solving:1 exit:3 basis:1 aii:2 joint:1 kp:3 artificial:1 whose:2 otherwise:1 noisy:1 differentiable:1 vud:1 propose:3 maximal:1 fr:3 relevant:1 mete:1 convergence:9 jed:2 generating:1 converges:1 tk:24 oo:1 andrew:1 quotient:2 convention:1 yiyj:1 stochastic:13 exploration:1 vx:6 frontier:2 hold:4 hall:1 estimation:2 fs93j:2 always:1 pn:2 ej:6 cr:1 barto:1 sudy:1 derived:1 sense:1 lj:1 france:2 arg:1 ill:1 art:1 cube:1 once:3 having:1 future:1 t2:1 delayed:1 fd:3 possibility:1 kt:3 partial:1 iv:10 taylor:1 eckhard:1 formalism:1 markovian:1 cover:1 lattice:1 subset:2 rare:1 rex:2 kn:5 antony:1 thanks:2 deduced:1 st:1 siam:1 international:1 ie:1 stay:1 off:1 jo:1 again:1 thesis:1 satisfied:3 derivative:1 dimitri:1 li:4 de:1 wk:1 satisfy:1 mv:1 vi:1 depends:1 lot:1 sup:11 square:1 ir:2 il:1 wendell:1 trajectory:8 ed:2 definition:1 mun97:1 proof:1 jacobi:2 proved:1 lim:1 jacobibellman:1 occured:1 back:2 appears:2 feed:2 dt:5 rand:2 stage:2 christopher:1 replacing:2 ei:14 nonlinear:1 incrementally:1 aj:3 building:1 true:1 evolution:1 round:1 adjacent:1 inferior:1 harold:1 polytechnique:1 vo:1 tn:1 motion:3 bradtke:1 fttk:2 laij:1 fi:1 functional:2 mt:1 rl:10 approximates:2 cambridge:1 mathematics:1 dot:1 v0:1 deduce:3 brownian:3 closest:1 verlag:3 mr:1 eo:1 vxixj:3 sc2:1 signal:1 ii:9 rj:1 jrd:1 seidel:1 ing:1 controlled:2 expectation:1 iteration:2 c1:1 cedex:2 regularly:2 integer:2 ee:4 viability:1 concerned:1 variety:1 fori:1 knowing:1 tkf:3 synchronous:1 motivated:1 ird:1 peter:1 puw:1 york:1 remark:2 adequate:1 v8:1 viscosity:1 discount:2 diameter:1 vy:1 cemagref:2 zd:2 kloeden:1 supueuh:2 diffusion:4 backward:1 vn:2 lake:1 decision:1 vex:5 appendix:2 vf:3 bound:2 fl:1 convergent:5 infinity:2 bp:1 ftk:1 generates:1 min:2 majoration:1 xixj:1 soner:1 stic:1 equation:10 available:1 neighbourhood:2 jd:1 jiei:1 running:1 kushner:2 lipschitzian:1 uj:2 objective:1 strategy:1 gradient:2 dp:4 vd:1 stocha:1 reason:1 besides:8 xixi:1 vee:1 negative:1 policy:1 unknown:2 u6u:1 upper:1 markov:1 finite:7 learningfor:2 drift:1 kl:1 fleming:1 krylov:1 dynamical:1 below:1 challenge:1 fs93:2 program:1 including:1 scheme:10 pare:1 lk:1 occurence:1 kf:1 law:2 interesting:1 sufficient:2 principle:1 ift:1 transpose:1 asynchronous:1 soon:1 free:2 aij:3 wide:1 munos:6 kp95j:3 boundary:5 dimension:1 overcome:1 transition:2 cumulative:1 kzk:1 forward:1 reinforcement:12 palaiseau:1 approximate:1 compact:1 satinder:1 mter:1 continuous:11 un:1 zk:14 traj:1 poly:1 cl:2 paul:1 dynami:1 precision:1 ueu:2 exponential:1 kus90:2 watkins:1 ito:1 learns:1 ix:1 theorem:7 rk:1 exists:6 ci:2 phd:1 kx:1 crea:1 mf:3 platen:1 remi:2 formly:1 insures:2 v6:16 springer:3 satisfies:2 supueu:1 ma:4 prop:1 lipschitz:1 uniformly:1 dwk:3 called:2 eeh:1 gauss:1 e6:3 iv6:2 statespace:1 ex:6 |
445 | 1,405 | Hybrid NNIHMM-Based Speech Recognition
with a Discriminant Neural Feature Extraction
Daniel Willett, Gerhard RigoU
Department of Comfuter Science
Faculty of Electrica Engineering
Gerhard-Mercator-University Duisburg, Germany
{willett,rigoll}@tb9-ti.uni-duisburg.de
Abstract
In this paper, we present a novel hybrid architecture for continuous speech
recognition systems. It consists of a continuous HMM system extended
by an arbitrary neural network that is used as a preprocessor that takes
several frames of the feature vector as input to produce more discriminative feature vectors with respect to the underlying HMM system. This
hybrid system is an extension of a state-of-the-art continuous HMM system, and in fact, it is the first hybrid system that really is capable of outperforming these standard systems with respect to the recognition accuracy.
Experimental results show an relative error reduction of about 10% that
we achieved on a remarkably good recognition system based on continuous HMMs for the Resource Management 1OOO-word continuous speech
recognition task.
1 INTRODUCTION
Standard state-of-the-art speech recognition systems utilize Hidden Markov Models
(HMMs) to model the acoustic behavior of basic speech units like phones or words. Most
commonly the probabilistic distribution functions are modeled as mIXtures of Gaussian distributions. These mixture distributions can be regarded as output nodes of a Radial-BasisFunction (RBF) network that is embedded in the HMM system [1]. Contrary to neural training procedures the parameters of the HMM system, including the RBF network, are usually estimated to maximize the training observations' likelihood. In order to combine the
time-warping abilities of HMMs and the more discriminative power of neural networks,
several hybrid approaches arose during the past five years, that combine HMM systems and
neural networks. The best known approach is the one proposed by Bourlard [2]. It replaces
the HMMs' RBF -net with a Multi-Layer-Perceptron (MLP) which is trained to output each
HMM state's posterior probability. At last year's NIPS our group presented a novel hybrid
speech recognition approach that combines a discrete HMM speech recognition system and
a neural quantizer (3). By maximizing the mutual information between the VQ-Iabels and
the assigned phoneme-classes, this apl'roach outperforms standard discrete recognition systems. We showed that this approach IS capable of building up very accurate systems with
an extremely fast likelihood computation, that only consists of a quantization and a table
lookup. This resulted in a hybrid system with recognition performance equivalent to the best
D. Willett and G. Rigoll
764
feature
extraction
x (t-P)
x(t)
x(t+F)
Neural Network
HMM-System
x'(t)
(linear transfonnation,
MLP or recurrent MLP)
(RBF-network)
p(x(t)IW[)
p(x(t)l~
Figure I: Architecture of the hybrid NN/HMM system
continuous systems, but with a much faster decoding. Nevertheless, it has turned out that
this hybrid approach is not really capable of substantially outperforming very good continuous systems with respect to the recognition accuracy. This observation is similar to experiences with Bourlard's MLP approach. For the decoding procedure, this architecture offers
a very efficient pruning technique (phone deactivation pruning [4)) that is much more efficient than pruning on likelihoods, but until today this approach did not outperform standard
continuous HMM systems in recognition performance.
2
HYBRID CONTINUOUS HMMlMLP APPROACH
Therefore, we followed a different approach, namely the extension of a state-of-the-art continuous system that achieves extremely good recognition rates with a neural net that is
trained with MMI-methods related to those in [5]. The major difference in this approach
is the fact that the acoustic processor is not replaced by a neural network, but that the Gaussian probability density component is retained and combined with a neural component in an
appropriate manner. A similar approach was presented in [6] to improve a speech recognitIOn system for the TIMIT database. We propose to regard the additional neural component
as being part of the feature extraction, and to reuse it in recognition systems of higher complexity where discriminative training is extremely expensive.
2.1
ARCHITECTURE
The basic architecture of this hybrid system is illustrated in Figure 1. The neural net functions as a feature transformation that takes several additional past and future feature vectors into account to produce an improved more discriminant feature vector that is fed into
the HMM system. This architecture allows (at least) three ways of interpretation; 1. as a
hybrid system that combines neural nets and continuous HMMs, 2. as an LDA-like transformation that incorporates the HMM parameters into the calculation of the transformation
matrix and 3. as feature extraction method, that allows the extraction offeatures according
to the underlying HMM system. The considered types of neural networks are linear transformations, MLPs and recurrent MLPs. A detailed 3escription of the possible topologies is
given in Section 3.
With this architecture, additional past and future feature vectors can be taken into account
in the probability estimation process without increasing the dimensionality of the Gaussian
mixture components. Instead of increasing the HMM system's number of parameters the
neural net is trained to produce more discriminant feature vectors with respect to the trained
HMM system. Of course, adding some kind of neural net increases the number of parameters too, but the increase is much more moderate than it would be when increasing each
Gaussian's dimensionality.
765
Speech Recognition with a Discriminant Neural Feature Extraction
2.2
TRAINING OBJECTIVE
The original purpose of this approach was the intention to transfer the hybrid approach
presented in [3], based on MMI neural network, to (semi-) continuous systems. ThIS way,
we hoped to be able to achieve the same remarkable improvements that we obtained on
discrete systems now on continuous systems, which are the much better and more flexible
baseline systems. The most natural way to do this would be the re-estimation of the codebook of Gaussian mean vectors of a semi-continuous system using the neural MMI training
algorithm presented in [31. Unfortunately though, this won't work, as this codebook of a
semi-continuous system does not determine a separation of the feature space, but is used as
means of Gaussian densities. The MMI-principle can be retained, however, by leaving the
original HMM system unmodified and instead extending it with a neural component, trained
according to a frame-based MMI approach, related to the one in [3] . The MMI criterion is
usually formulated in the following way:
= argmaxi).(X, W) = argmax(H>.(X) - H>.(XIW)) = argmax P>.(~l~)
>.
>.
>.
P>.
(1)
This means that following the MMI criterion the system's free parameters ,\ have to be estimated to maximize the quotient of the observation's likelihood p>.(XIW) for the known
transcription Wand its overall likelihood P>. (X). With X = (x(l), x(2), .. .x(T)) denot(w(l), w(2) , .. .w(T)) denoting the HMM statesing the training observations and W
assigned to the observation vectors in a Viterbi-alignment - the frame-based MMI criterion
becomes
)..MMI
=
T
)..MMI
~ arg~ax
L i).(x(i), w(i))
i=1
ITT
IT p>.(x(i)lw(i))
p>.(x(i))
:::::::arg~ax .
=arg~ax .
p>.(x(i)lw(i))
T
%=1
1:1
s.
2: P>.(X(Z)lwk)p(Wk)
(2)
k=1
where S is the total number ofHMM states, (W1 , '" ws) denotes the HMM states and p( Wk)
denotes each states' prior-probability that is estimated on the alignment of the training data
or by an analysis of the lan~age model.
Eq. 2 can be used to re-estImate the Gaussians of a continuous HMM system directly. In
[7] we reported the slight improvements in recognition accuracy that we achieved with this
parameter estimation. However, it turned out, that only the incorporation of additional features in the probability calculation pipeline can provide more discriminative emission probabilities and a major advance in recognition accuracy. Thus, we experienced it to be more
convenient to train an additional neural net in order to maximize Eq. 2. Besides, this approach offers the possibility of improving a recognition system by applying a trained feature
extraction network taken from a different system. Section 5 will report our positive experiences with this procedure.
At first, for matter of simplicity, we will consider a linear network that takes P past feature vectors and F future feature vectors as additional input. With the linear net denoted as
a (P + F + 1) x N matrix NET, each component x' (t)[c] of the network output x'(t)
computes to
P+F N
x'(t)[c] =
L L x(t -
P
+ i)[jJ . N ET[i * N + j][c]
Vc E {L.N}
(3)
i=O j=1
so that the derivative with respect to a component of NET easily computes to
8x'(t)[cJ
_ 6 -x(t - P
8NET[i*N+j][c] - C,c
+
i)['J
(4)
J
In a continuous HMM system with diagonal covariance matrices the pdf of each HMM state
w is modeled by a mixture of Gaussian components like
N
C
p>.(xlw) =
Ld
j=1
1 ~ (m;[I)-x[l])
1
1U
wj
y'(2rr)nIUjl
-"2 L,
e
1=1
2
ajll)
(5)
766
D. Willett and G. Rigoll
A pdfs derivative with respect to a component x'[e] of the net's output becomes
N
1 ~ (mj[ll-",/[I))2
C
8p,\.(x'lw) _ ~ d . (:e[e] - mj[e))
1
e- 2 ~
8x'[e) - ~ WJ
oAe)
y'(21l')nIO'jl
aj[!}
(6)
With x(t) in Eq. 2 now replaced by the net output x'(t) the partial derivative ofEq. 2 with
respect to a probabilistic distribution function p( x' (i) IWk) computes to
8h(x'(i), w(i)) _ <5W(i) ,Wk
8p,\.(x'(i)lwk) - p,\.(x(i)lwk)
s
(7)
2: p,\.(x(i)lwI)p(wd
1=1
Thus, using the chain rule the derivative of the net's parameters with respect to the framebased MMI criterion can be computed as displayed in Eq. 8
8h(X, W) = t(t(8h(:e(i)IW(i))) 8p,\.(x'(i)l wk) ox' (i)[c] )
8N ET[lHe)
i=l k=l op,\.(x'(i)lwk)
ox' (i)[e] oN ET[lHe)
(8)
and a gradient descent procedure can be used to determine the optimal parameter estimates.
2.3
ADVANTAGES OF THE PROPOSED APPROACH
When using a linear network, the proposed approach strongly resembles the well known
Linear Discriminant Analysis (LDA) [8] in architecture and training objective. The main
difference is the way the transformation is set up. In the proposed approach the transformation is computed by taking directly the HMM parameters into account whereas the LDA
only tries to separate the features according to some class assignment. With the incorporation of a trained continuous HMM system the net's parameters are estimated to produce
feature vectors that not only have a good separability in general, but also have a distribution that can be modeled with mixtures ofGaussians very well. Our experiments given at the
end ofthis paper prove this advantage. Furthermore, contrary to LDA, that produces feature
vectors that don't have much in common with the original vectors, the proposed approach
only slightly modifies the input vectors. Thus, a well trained continuous system can be extended by the MMI-net approach, in order to improve its recognition performance without
the need for completely rebuilding it. In addition to that, the approach offers a fairly easy
extension to nonlinear networks (MLP) and recurrent networks (recurrent MLP). This will
be outlined in the following Section. And, maybe as the major advantage, the approach allows keeping up the division of the input features into streams of features that are strongly
uncorrelated and which are modeled with separate pdfs. The case of multiple streams is discussed in detail in Section 4. Besides) the MMI approach offers the possibility of a unified
training of the HMM system and the reature extraction network or an iterative procedure of
training each part alternately.
3
NETWORK TOPOLOGIES
Section 2 explained how to train a linear transformation with respect to the frame-based
MMI criterion. However, to exploit all the advantages of the proposed hybrid approach the
network should be able to perform a nonlinear mapping, in order to produce features whose
distribution is (closer to) a mixture of Gaussians although the original distribution is not.
3.1
MLP
When using a fully connected MLP as displayed in Figure 2 with one hidden layer of H
nodes, that perform the nonlinear function /, the activation of one of the output nodes
x'(t)[e] becomes
H
x'(tHe) =
t;
P+F N
L2[hJ[e)?
f( BIAS. + ~ ~ x(t - P + i)li)? L1[i * N + jJ[h))
(9)
767
Speech Recognition with a Discriminant Neural Feature Extraction
original features
(multiple frames)
: :~--:EA6
:::
I: I
I
I
!: :'::
:::
o
I
I
x(t-l)
x(t)
x(t+1)
1
~
~
~
~
x'(t)
I
'
"''"~' ~... \NlIl,~l/\f)(JI\L.-i.M/I/
\,,~~
Ll
'~~:\ . .
"'~~~'~~~~~11111~
~
p(x(t)lw ~
I
I
network
~~~~
:::
:::
:::
I
RBF-
I
: : ~ _______________
1- '--_-_-.:_-_-_-_-_-_-_-_-_-_-_-_-_
p(x(t)lwi
p(x(t)lwj
?~~
transformed features
- __
context-dependent
Hidden Markov Model
Figure 2: Hybrid system with a nonlinear feature transfonnation
which is easily differentiable with respect to the nonlinear network's parameters. In our
experiments we chose f to be defined as the hyperbolic tangents f(x) := tanh(x) = (2(1+
e -X) -1 - 1) so that the partial derivative with respect to i.e. a weight L 1[i . N + J] [h] of
the first layer computes to
ax' (t)[e]
aL1[i . N + 3][h]
x(t - p
+ i)[J] . L2[h][c]
.COSh(BIAS. +
%~
(10)
x(t - P + i)lil? L1[i * N
+ j][h1r'
and the gradient can be set up according to Eq. 8.
3.2
RECURRENT MLP
With the incorporation of several additional past feature vectors as explained in Section 2,
more discriminant feature vectors can be generated. However, this method is not capable of
modeling longer term relations, as it can be achieved by extending the network with some
recurrent connections. For the sake of simplicity, in our experiments we simply extended
the MLP as indicated with the dashed lines in Figure 2 by propagating the output x (t) back
to the input of the network (with a delay of one discrete time step). This type of recurrent
neural net is often referred to as a 'Jordan' -network. Certainly, the extension of the network
with additional hidden nodes in order to model the recurrence more independently would
be possible as well.
4
MULTI STREAM SYSTEMS
In HMM-based recognition systems the extracted features are often divided into streams
that are modeled independently. This is useful the less correlated the divided features are.
In this case the overall likelihood of an observation computes to
M
p>.(xlw)
= IT p$>.(xlw)w,
(11)
where each of the stream pdfs p$>.(xlw) only uses a subset of the features in x. The stream
weights W$ are usually set to unity.
D. Willett and G. Rigoll
768
Table 1: Word error rates achieved in the experiments
A multi stream system can be improved by a neural extraction for each stream and an independent training of these neural networks. However, it has to be considered that the
subdivided features usually are not totally independent and by considering multiple input
frames as illustrated in Figure 1 this dependence often increases. It is a common practice,
for instance, to model the features' first and second order delta coefficients in independent streams. So, for sure the streams lose independence when considering multiple frames,
as these coefficients are calculated using the additional frames. Nevertheless, we found it
to give best results to maintain this subdivision into streams, but to consider the stronger
correlation by training each stream's net dependent on the other nets' outputs. A training
criterion follows straight from Eq. 11 inserted in Eq. 2.
\
_
/\MMI -
argmax
).
rrT p).(x(i)lw(i)) _
( (.))
i=l
P). X t
- argmax
.).
rrT 11M (P3).(X(i)IW(i)))W'
i=13=1
( (.))
Ps). X
z
The derivative of this equation with respect to the pdf P.;). (xlw) ofa specific stream
pends on the other streams' pdfs. With the Ws set to unity it is
8h(x'(;): w(i)) =
8ps).(x (Z)IWk)
(rr ps).(X(i)I~(i))) (
~.
3r 3
P3).(X(Z))
6w (i):Wk
ps).(X(Z)IWk)
(12)
s de-
_ s
p(Wk)
)
'I\'
((')1) ( )
W P.;). x Z WI P WI
1=1
(13)
Neglecting the correlation among the streams the training of each stream's net can be done
independently. However, the more the incorporation of additional features increases the
streams' correlation, the more important it gets to train the nets in a unified training procedure according to Eq. 13.
5
EXPERIMENTS AND RESULTS
We applied the proposed approach to improve a context-independent (monophones) and a
context-dependent (triphones) continuous speech reco~tion system for the 1000-wordResource Management (RM) task. The systems used lmear HMMs of three emitting states
each. The tying of Gaussian mixture components was perfonned with an adaptive procedure according to [9]. The HMM states of the word-internal triphone system were clustered
in a tree-based phonetic clustering procedure. Decoding was perfonned with a Viterbidecoder and the standard wordpair-grammar of perplexity 60. Training of the MLP was
perfonned with the RPROP algorithm. For training the weights of the recurrent connections we chose real-time recurrent learning. The average error rates were computed using
the test-sets Feb89, Oct89, Feb91 and Sep92.
The table above shows the recognition results with single stream systems in its first section.
These systems simply use a 12-value Cepstrum feature vector without the incorporation of
delta coefficients. The systems with an input transfonnation use one additional past and one
additional future feature vector as input. The proposed approach achieves the same perfonnance as the LDA, but it is not capable of outperfonning It.
The second section of the table lists the recognition results with four stream systems that
use the first and second order delta coefficients in additional streams plus log energy and
this values' delta coefficients in a forth stream. The MLP system trained according to Eq.
Speech Recognition with a Discriminant Neural Feature Extraction
769
II slightly outperforms the other approaches. The incorporation of recurrent network connections does not improve the system's performance.
The third section of the table lists the recognition results with four stream systems with a
context-dependent acoustic modeling (triphones). The applied LDA and the MMI-networks
were taken from the monophone four stream system. On the one hand, this was done to
avoid the computational complexity that the MMI training objective causes on contextdependent systems. On the other hand, this demonstrates that the feature vectors produced
by the trained networks have a good discrimination for continuous systems in general.
Again, the MLP system outperforms the other? approaches and achieves a very remarkable
word error rate. It should be pointed out here, that the structure of the continuous system as
reported in (9) is already highly optimized and it is almost impossible to further reduce the
error rate by means of any acoustic modeling method. This is reflected in the fact that even
a standard LDA cannot improve this system. Only the new neural approach leads to a 10%
reduction in error rate which is a large improvement considering the fact that the error rate
of the baseline system is among the best ever reported for the RM database,
6
CONCLUSION
The paper has presented a novel approach to discriminant feature extraction. A MLP network has successfully been used to compute a feature transformation that outputs extremely
suitable features for continuous HMM systems. The experimental results have proven that
the proposed approach is an appropriate method for including several feature frames in the
probability estimation process without increasing the dimensionality of the Gaussian mixture components in the HMM system. Furthermore did the results on the triphone speech
recognition system prove that the approach provides discriminant features, not only for the
system that the mapping is computed on, but for HMM systems with a continuous modeling
in general: The application of recurrent networks did not improve the recognition accuracy.
The longer range relations seem to be very weak and they seem to be covered well by using
the neighboring feature vectors and first and second order delta coefficients. The proposed
unified training procedure for multiple nets in multi-stream systems allows keeping up the
subdivision of features of weak correlations, and gave us best profits in recognition accuracy.
References
[1) H. Ney, "Speech Recognition in a Neural Network Framework: Discriminative Training of Gaussian Models and Mixture Densities as Radial Basis Functions", Proc. IEEEICASSp, 1991, pp. 573-576.
[2) H Bouriard, N. Morgan, "Connectionist Speech Recognition - A Hybrid Approach",
Kluwer Academic Press, 1994.
[3) G. Rigoll, C. Neukirchen, "A new approach to hybrid HMMIANN speech recognition
using mutual information neural networks", Advances in Neural Information Processing
Systems (NIPS-96), Denver, Dec. 1996, pp. 772-778.
[4] M. M. Hochberg, G. D. Cook, S. J. Renals, A. J. Robinson, A. S. Schechtman, "The
1994 ABBOT Hybrid Connectionist-HMM Large-Vocabulary Recognition System",
Proc. ARPA Spoken Language Systems Technology Workshop, 1995.
[5] G. Rigoll, "Maximum Mutual Information Neural Networks for Hybrid ConnectionistHMM Speech Recognition", IEEE-Trans. Speech Audio Processing, Vol. 2, No.1, Jan.
1994,pp.175-184.
[6] Y. Bengio et aI., "Global Optimization of a Neural Network - Hidden Markov Model
Hybrid" IEEE-Transcations on NN, Vol. 3, No. 2, 1992, pp. 252-259.
[7] D. Willett, C. Neukirchen, R. Rottland, "Dictionary-Based Discriminative HMM Parameter Estimation for Continuous Speech Recognition Systems", Proc. IEEE-ICASSp,
1997,pp.1515-1518.
[8] X. Aubert, R. Haeb-Umbach, H. Ney, "Continuous mixture densities and linear discriminant analysis for improved context-dependent acoustic models", Proc. IEEE-ICASSp,
1993, pp. II 648-651.
[9) D. Willett, G. Rigoll, "A New Approach to Generalized Mixture Tying for Continuous
HMM-Based Speech Recognition",Proc. EUROSPEECH, Rhodes, 1997.
| 1405 |@word faculty:1 oae:1 stronger:1 lwk:4 covariance:1 profit:1 xlw:5 ld:1 feb91:1 reduction:2 daniel:1 denoting:1 past:6 outperforms:3 wd:1 activation:1 discrimination:1 rrt:2 cook:1 provides:1 quantizer:1 codebook:2 node:4 five:1 consists:2 prove:2 combine:4 umbach:1 manner:1 behavior:1 multi:4 considering:3 increasing:4 becomes:3 totally:1 underlying:2 tying:2 kind:1 substantially:1 unified:3 spoken:1 transformation:8 ti:1 ofa:1 rm:2 demonstrates:1 unit:1 positive:1 engineering:1 mercator:1 chose:2 plus:1 resembles:1 hmms:6 range:1 practice:1 procedure:9 jan:1 hyperbolic:1 convenient:1 word:5 radial:2 intention:1 get:1 cannot:1 context:5 applying:1 impossible:1 rottland:1 equivalent:1 maximizing:1 modifies:1 independently:3 simplicity:2 rule:1 oct89:1 regarded:1 today:1 gerhard:2 us:1 recognition:36 expensive:1 database:2 inserted:1 wj:2 connected:1 complexity:2 trained:10 division:1 completely:1 basis:1 easily:2 icassp:2 train:3 fast:1 argmaxi:1 feb89:1 rebuilding:1 whose:1 haeb:1 grammar:1 ability:1 advantage:4 rr:2 differentiable:1 net:22 propose:1 renals:1 neighboring:1 turned:2 achieve:1 forth:1 p:4 extending:2 produce:6 recurrent:11 propagating:1 op:1 eq:9 reco:1 quotient:1 vc:1 subdivided:1 clustered:1 really:2 extension:4 considered:2 viterbi:1 mapping:2 major:3 achieves:3 al1:1 dictionary:1 purpose:1 estimation:5 proc:5 rhodes:1 lose:1 iw:3 tanh:1 successfully:1 gaussian:10 arose:1 avoid:1 hj:1 ax:4 emission:1 improvement:3 pdfs:4 likelihood:6 baseline:2 dependent:5 nn:2 hidden:5 w:2 relation:2 transformed:1 germany:1 overall:2 arg:3 flexible:1 among:2 denoted:1 art:3 fairly:1 mutual:3 extraction:12 future:4 report:1 connectionist:2 resulted:1 hmmiann:1 replaced:2 argmax:4 aubert:1 maintain:1 mlp:14 possibility:2 highly:1 certainly:1 alignment:2 mixture:11 chain:1 accurate:1 capable:5 partial:2 closer:1 experience:2 neglecting:1 perfonnance:1 tree:1 re:2 denot:1 monophone:1 arpa:1 instance:1 neukirchen:2 modeling:4 abbot:1 unmodified:1 assignment:1 subset:1 delay:1 too:1 eurospeech:1 reported:3 combined:1 density:4 probabilistic:2 decoding:3 offeatures:1 w1:1 again:1 management:2 derivative:6 li:1 account:3 de:2 lookup:1 wk:6 coefficient:6 matter:1 escription:1 stream:24 tion:1 try:1 deactivation:1 timit:1 ofeq:1 mlps:2 accuracy:6 phoneme:1 weak:2 mmi:17 produced:1 straight:1 processor:1 rprop:1 energy:1 pp:6 dimensionality:3 cj:1 back:1 higher:1 reflected:1 improved:3 ooo:1 cepstrum:1 done:2 though:1 ox:2 strongly:2 furthermore:2 until:1 correlation:4 hand:2 nonlinear:5 aj:1 lda:7 indicated:1 building:1 assigned:2 illustrated:2 ll:2 during:1 recurrence:1 won:1 criterion:6 generalized:1 pdf:2 l1:2 novel:3 common:2 ji:1 denver:1 jl:1 discussed:1 interpretation:1 slight:1 kluwer:1 willett:7 ai:1 ofgaussians:1 outlined:1 pointed:1 language:1 longer:2 triphones:2 posterior:1 showed:1 moderate:1 phone:2 perplexity:1 phonetic:1 outperforming:2 morgan:1 additional:13 triphone:2 determine:2 maximize:3 dashed:1 semi:3 ii:2 multiple:5 faster:1 academic:1 calculation:2 offer:4 divided:2 basic:2 achieved:4 dec:1 wordpair:1 whereas:1 remarkably:1 addition:1 leaving:1 sure:1 contrary:2 transfonnation:3 incorporates:1 seem:2 jordan:1 bengio:1 easy:1 independence:1 gave:1 architecture:8 topology:2 reduce:1 rigoll:7 reuse:1 monophones:1 speech:20 cause:1 jj:2 useful:1 detailed:1 covered:1 maybe:1 cosh:1 outperform:1 estimated:4 delta:5 discrete:4 vol:2 group:1 four:3 duisburg:2 nevertheless:2 lan:1 utilize:1 year:2 wand:1 almost:1 separation:1 p3:2 hochberg:1 layer:3 followed:1 replaces:1 incorporation:6 sake:1 extremely:4 department:1 according:7 slightly:2 separability:1 unity:2 wi:2 sep92:1 explained:2 taken:3 pipeline:1 resource:1 vq:1 equation:1 fed:1 end:1 gaussians:2 appropriate:2 reature:1 ney:2 original:5 denotes:2 clustering:1 exploit:1 lhe:2 warping:1 objective:3 already:1 xiw:2 dependence:1 diagonal:1 gradient:2 separate:2 hmm:33 discriminant:11 besides:2 modeled:5 retained:2 unfortunately:1 lil:1 perform:2 observation:6 markov:3 roach:1 descent:1 displayed:2 extended:3 ever:1 frame:9 arbitrary:1 namely:1 connection:3 optimized:1 acoustic:5 nip:2 alternately:1 robinson:1 able:2 trans:1 lwj:1 usually:4 including:2 power:1 perfonned:3 suitable:1 natural:1 hybrid:20 transcations:1 bourlard:2 improve:6 technology:1 prior:1 l2:2 tangent:1 relative:1 apl:1 embedded:1 fully:1 ofhmm:1 proven:1 remarkable:2 age:1 principle:1 uncorrelated:1 course:1 last:1 free:1 keeping:2 bias:2 perceptron:1 taking:1 regard:1 lmear:1 calculated:1 vocabulary:1 computes:5 commonly:1 adaptive:1 emitting:1 pruning:3 uni:1 transcription:1 global:1 discriminative:6 don:1 continuous:27 iterative:1 table:5 mj:2 transfer:1 itt:1 improving:1 did:3 main:1 referred:1 experienced:1 lw:5 third:1 preprocessor:1 specific:1 list:2 ofthis:1 workshop:1 quantization:1 adding:1 hoped:1 simply:2 extracted:1 lwi:2 formulated:1 rbf:5 pends:1 total:1 experimental:2 subdivision:2 schechtman:1 internal:1 audio:1 contextdependent:1 correlated:1 |
446 | 1,406 | Bayesian Robustification for Audio Visual
Fusion
Javier Movellan *
movellanOcogsci.ucsd.edu
Department of Cognitive Science
University of California, San Diego
La Jolla, CA 92092-0515
Paul Mineiro
pmineiroOcogsci.ucsd.edu
Department of Cognitive Science
University of California, San Diego
La Jolla, CA 92092-0515
Abstract
We discuss the problem of catastrophic fusion in multimodal recognition systems. This problem arises in systems that need to fuse
different channels in non-stationary environments. Practice shows
that when recognition modules within each modality are tested in
contexts inconsistent with their assumptions, their influence on the
fused product tends to increase, with catastrophic results. We explore a principled solution to this problem based upon Bayesian
ideas of competitive models and inference robustification: each
sensory channel is provided with simple white-noise context models, and the perceptual hypothesis and context are jointly estimated. Consequently, context deviations are interpreted as changes
in white noise contamination strength, automatically adjusting the
influence of the module. The approach is tested on a fixed lexicon
automatic audiovisual speech recognition problem with very good
results.
1
Introduction
In this paper we address the problem of catastrophic fusion in automatic multimodal
recognition systems. We explore a principled solution based on the Bayesian ideas of
competitive models and inference robustification (Clark & Yuille, 1990; Box, 1980;
O'Hagan, 1994). For concreteness, consider an audiovisual car telephony task which
we will simulate in later sections. The task is to recognize spoken phone numbers
based on input from a camera and a microphone. We want the recognition system to
work on environments with non-stationary statistical properties: at times the video
signal (V) may be relatively clean and the audio signal (A) may be contaminated by
sources like the radio, the engine, and friction with the road. At other times the A
signal may be more reliable than the V signal, e.g., the radio is off, but the talker's
mouth is partially occluded. Ideally we want the audio-visual system to combine
the A and V sources optimally given the conditions at hand, e.g., give more weight
to whichever channel is more reliable at that time. At a minimum we expect that
for a wide variety of contexts, the performance after fusion should not be worse than
the independent unimodal systems (Bernstein & Benoit, 1996). When component
modules can Significantly outperform the overall system after fusion, catastrophic
fusion is said to have occurred.
? To whom correspondence should be addressed.
Bayesian Robustification for Audio Visual Fusion
743
Fixed vocabulary audiovisual speech recognition (AVSR) systems typically consist of two independent modules, one dedicated to A signals and one to V signals
(Bregler, Hild, Manke & Waibel, 1993; Wolff, Prasad, Stork & Hennecke, 1994;
Adjondani & Benoit, 1996; Movellan & Chadderdon, 1996). From a Bayesian perspective this modularity reflects an assumption of conditional independence of A
and V signals (i.e., the likelihood function factorizes)
P(XaXvIWi.Aa.Av) <X P(XaIWi.Aa)P(xvlwi.Av),
(1)
where Xa and Xv are the audio and video data, Wi is a perceptual interpretation
of the data (e.g., the word "one") and {.Aa,.A v } are the audio and video models
according to which these probabilities are calculated, e.g., a hidden Markov model ,
a neural network, or an exemplar model. Training is also typically modularized: the
A module is trained to maximize the likelihood of a sample of A signals while the V
module is trained on the corresponding sample of V signals. At test time new data
are presented to the system and each module typically outputs the log probability
of its input given each perceptual alternative. Assuming conditional independence,
Bayes' rule calls for an affine combination of modules
'Ii;
argmax {logp(wilxaxv.Aa.Av)}
Wi
Wi
where 'Ii; is the interpretation chosen by the system, and p(Wi) is the prior probability
of each alternative. This fusion rule is optimal in the sense that it minimizes the
expected error: no other fusion rule produces smaller error rates, provided the
models {.Aa,.A v } and the assumption of conditional independence are correct.
Unfortunately a naive application of Bayes' rule to AVSR produces catastrophic
fusion. The A and V modules make assumptions about the signals they receive,
either explicitly, e.g., a well defined statistical model, or implicitly, e.g., a blackbox trained with a particular data sample. In our notation these assumptions are
reflected by the fact that the log-likelihoods are conditional on models: {.A a , .A v }.
The fact that modules make assumptions implies that they will operate correctly
only within a restricted context, i.e, the collection of situations that meet the assumptions. In practice one typically finds that Bayes' rule assigns more weight to
modules operating outside their valid context, the opposite of what is desired.
2
Competitive Models and Bayesian Robustification
Clark and Yuille (1990) and Yuille and Bulthoff (1996) analyzed information integration in sensory systems from a Bayesian perspective. Modularity is justified in
their view by the need to make assumptions that disambiguate the data available to
the perceptual system (Clark & Yuille, 1990, p. 5). However, this produces modules
which are valid only within certain contexts. The solution proposed by Clark and
Yuille (1990) is the creation of an ensemble of models each of which specializes on
a restricted context and automatically checks whether the context is correct. The
hope is that by working with such an ensemble of models, robustness under a variety
of contexts can be achieved (Clark & Yuille, 1990, p. 13).
Box (1980) investigated the problem of robust statistical inference from a Bayesian
perspective. He proposed extending inference models with additional "nuisance"
parameters a, a process he called Bayesian robustification. The idea is to replace
an implicit assumption about the specific value of a with a prior distribution over
a, representing uncertainty about that parameter.
The approach here combines the ideas of competitive models and robustification.
Each of the channels in the multimodal recognition system is provided with extra
744
1. Movellan and P. Mineiro
parameters that represent non-stationary properties of the environment, what we
call a context model. By doing so we effectively work with an infinite ensemble of
models each of which compete on-line to explain the data. As we will see later even
unsophisticated context models provide superior performance when the environment
is non-stationary.
We redefine the estimation problem as simultaneously choosing the most probable
A and V context parameters and the most probable perceptual interpretation
w = argmax
{max P(WwaO'vIXaXvAaAv)}
Wi
(3)
Uo,U v
where 0' a and 0' v are the context parameters for the audio and visual channels
and Wi are the different perceptual interpretations. One way to think of this joint
decision approach is that we let all context models compete and we let only the
most probable context models have an influence on the fused percept. Hereafter we
refer to this approach as competitive fusion.
Assuming conditional independence of the audio and video data and uninformative
priors for (0' a, 0'v), we have
w= ar~ax {logp(Wi) +
[~~IOgP(XalwwaAa)]
+
[~::XIOgP(xvIWiO'vAv)]}.
(4)
Thus conditional independence allows a modular implementation of competitive
fusion, Le., the A and V channels do not need to talk to each other until the time
to make a joint decision, as follows.
1. For each Wi obtain conditional estimates of the context parameters for the
audio and video signals:
o-~IWi ~ argmax { 10gp(xalwWaAa) } ,
(5)
(To.
and
o-~IWi ~ argmax{ 10gp(xvlwWvAv) }.
(6)
(Tv
2. Find the best Wi using the conditional context estimates.
w= argmax
{IOgp(Wi) + logp(xalwio-alwi Aa) + logp(xv IWio-vlwi Av)}
Wi
3
(7)
Application to AVSR
Competitive fusion can be easily applied to Hidden Markov Models (HMM), an
architecture closely related to stochastic neural networks and arguably the most
successful for AVSR. Typical hidden Markov models used in AVSR are defined by
? Markovian state dynamics: p(qt+ll2.t )
time t and 2.t = (ql," . qt),
= p(qt+llqt), where qt is the state at
? Conditionally independent sensor models linking observations to states
! (Xt Iqt), typically a mixture of multivariate Gaussian densities
!(xtlqt) =
L p(mi Iqt)(27l')
-N/2
1~I-l/2 exp(d(xt, qt, Pi, ~)),
(8)
Bayesian Robustification for Audio VIsual Fusion
745
where N is the dimensionality of the data, mi is the mixture label, p(milqt)
is the mixture distribution for state qt, Pi is the centroid for mixture mi,
E is a covariance matrix, and d is the Mahalanobis norm
(9)
The approach explored here consists on modeling contextual changes as variations
on the variance parameters. This corresponds to modeling non-stationary properties of the environments as variations in white noise power within each channel.
Competitive fusion calls for on-line maximization of the variance parameters at the
same time we optimize with respect to the response alternative.
11; =
arg:ax{ logp(wi)
[
+
~~logp(xalwiEaAa)]
+ [ ~~logp(xvlwiEv'\v)] }.
(10)
The maximization with respect to the variances can be easily integrated into standard HMM packages by simply applying the EM learning algorithm (Dampster,
Laird & Rubin, 1977) on the variance parameters at test time. Thus the only difference between the standard approach and competitive fusion is that we retrain
the variance parameters of each HMM at test time. In practice this training takes
only one or two iterations of the EM algorithm and can be done on-line. We tested
this approach on the following AVSR problem.
Training database We used Tulips1 (Movellan, 1995) a database consisting of
934 images of 9 male and 3 female undergraduate students from the Cognitive
Science Department at the University of California, San Diego. For each of these,
two samples were taken for each of the digits "one" through "four". Thus, the total
database consists of 96 digit utterances. The specifics of this database are explained
in (Movellan, 1995). The database is available at http://cogsci.ucsd.edu.
Visual processing We have tried a wide variety of visual processing approaches
on this database, including decomposition with local Gaussian templates (Movellan,
1995), PCA-based templates (Gray, Movellan & Sejnowski, 1997), and Gabor energy
templates (Movellan & Prayaga, 1996). To date, best performance was achieved
with the local Gaussian approach. Each frame of the video track is soft-thresholded
and symmetrized along the vertical axis, and a temporal difference frame is obtained
by subtracting the previous symmetrized frame from the current symmetrized frame.
We calculate the inner-products between the symmetrized images and a set of basis
images. Our basis images were lOx15 shifted Gaussian kernels with a standard
deviation of 3 pixels. The loadings of the symmetrized image and the differential
image are combined to form the final observation frame. Each of these composite
frames has 300 dimensions (2xlOx15). The process is explained in more detail in
Movellan (1995).
Auditory processing LPC /cepstral analysis is used for the auditory front-end.
First, the auditory signal is passed through a first-order emphasizer to spectrally
flatten it. Then the signal is separated into non-overlapping frames at 30 frames
per second. This is done so that there are an equal number of visual and auditory
feature vectors for each utterance, which are then synchronized with each other. On
each frame we perform the standard LPC / cepstral analysis. Each 30 msec auditory
frame is characterized by 26 features: 12 cepstral coefficients, 12 delta-cepstrals,
1 log-power, and 1 delta-log-power. Each of the 26 features is encoded with 8-bit
accuracy.
J Movellan and P. Mineiro
746
Figure 1: Examples of the different occlusion levels, from left to right: 0%, 10%,
20%, 40%, 60%, 80%. Percentages are in terms of area.
Recognition Engine In previous work (Chadderdon & Movellan, 1995) a wide
variety of HMM architectures were tested on this database including architectures
that did not assume conditional independence. Optimal performance was found
with independent A and V modules using variance matrices of the form uI, where
u is a scalar and I the identity matrix. The best A models had 5 states and 7
mixtures per state and the best V models had 3 states and 3 mixtures per state.
We also determined the optimal weight of A and V modules. Optimal performance
is obtained by weighting the output of V times 0.18.
Factorial Contamination Experiment In this experiment we used the previously optimized architecture and compared its performance under 64 different
conditions using the standard and the competitive fusion approaches. We used a
2 x 8 x 8 factorial design, the first factor being the fusion rule, and the second and
third factors the context in the audio and video channels. To our knowledge this is
the first time an AVSR system is tested with a factorial experimental design with
both A and V contaminated at various levels. The independent variables were:
1. Fusion rule: Classical, and competitive fusion.
2. Audio Context: Inexistent, clean, or contaminated at one of the following
signal to noise ratios: 12 Db, 6 Db, 0 Db, -6 Db, -12 Db and -100 Db. The
contamination was done with audio digitally sampled from the interior of
a car while running on a busy highway with the doors open and the radio
on a talk-show station.
3. Video Context: Inexistent, clean or occluded by a grey level patch. The
percentages of visual area occupied by the patch were 10%,20%,40%,60%,
80% and 100% (see Figure 1).
The dependent variable was performance on the digit recognition task evaluated in
terms of generalization to new speakers. In all cases training was done with clean
signals and testing was done with one of the 64 contexts under study. Since the
training sample is small, generalization performance was estimated using a jackknife
procedure (Efron, 1982). Models were trained with 11 subjects, leaving a different
subject out for generalization testing. The entire procedure was repeated 12 times,
each time leaving a different subject out for testing. Statistics of generalization
performance are thus based on 96 generalization trials (4 digits x 12 subjects x
2 observations per subject). Standard statistical tests were used to compare the
classical and competitive context rules.
The results of this experiment are displayed in Table 1. Note how the experiment
replicates the phenomenon of catastrophic fusion. With the classic approach, when
one of the channels is contaminated, performance after fusion can be significantly
Bayesian Robustificationfor Audio Visual Fusion
Video
747
one
None
-6
-1
95.83
95.83
90.62
80.20
67.70
42 . 70
Clean
84 .37
97.92
97.92
94.80
90.62
89.58
81.25
10%
73.95
93.75
93.75
94 .79
87 .50
80 .20
71.87
19 .80
82 20
1 64.58
.
I
20%
62.50
96.87
96 .87
94.79
89.58
80.20
1 6 2 .501
41.66
40%
37.50
93 .75
89.58
87.50
83.30
70.83
43 .75
30.20
60%
34.37
93.75
91.66
82.29
65.62
42.70
26.04
80%
27.00
95.83
90.62
79.16
64.58
146.871
25 .00
100%
25.00
93.75
92.71
84.37
78.12
63.54
Performance wIth C/a .. 1e Fu" on
Audio
44 .79
26.04
Video
one
20 .83
None
ean
95 .83
94.79
89.58
79. 16
65 .62
40 .62
Clean
86 .45
98.95
96.87
95.83
93.75
87.50
79 . 16
10%
73.95
93.75
93 .75
93.75
89.58
79 .16
70 .83
70 83
1 52 ..58 /
20%
54.16
89.58
84.37
84.37
75.00
29.16
~
43 .00
40%
81.25
78.12
67.20
52 .08
38.54
34.37
60%
32.29
77 .08
72.91
62 .50
47.91
37.50
29 . 16
80%
29 . 16
70.83
68.75
54.16
44 .79
1 33 .831
28.12
100%
25.00
61.46
58.33
51.04
42.70
38.54
29 . 16
61.45
Table 1: Average generalization performance with standard and competitive fusion.
Boxed cells indicate a statistically significant difference a = 0.05 between the two
fusion approaches.
worse than performance with the clean channel alone. For example, when the audio
is clean, the performance of the audio-only system is 95.83%. When combined with
bad video (100% occlusion), this performance drops down to 61.46%, a statistically
significant difference, F(l,ll) = 132.0, p < 10- 6 . Using competitive fusion, the
performance of the joint system is 93.75%, which is not significantly different from
the performance of the A system only, F(l,ll) = 2.4, p= 0.15. The table shows
in boxes the regions for which the classic and competitive fusion approaches were
significantly different (a = 0.05). Contrary to the classic approach, the competitive
approach behaves robustly in all tested conditions.
4
Discussion
Catastrophic fusion may occur when the environment is non-stationary forcing modules to operate outside their assumed context. The reason for this problem is that in
the absence of a context model, deviations from the expected context are interpreted
as information about the different perceptual interpretations instead of information
about contextual changes. We explored a principled solution to this problem inspired by the Bayesian ideas of robustification (Box, 1980) and competitive models
(Clark & Yuille, 1990). Each module was provided with simple white-noise context
models and the most probable context and perceptual hypothesis were jointly estimated. Consequently, context deviations are interpreted as changes in the white
noise contamination strength, automatically adjusting the influence of the module.
The approach worked very well on a fixed lexicon AVSR problem.
References
Adjondani, A. & Benoit, C. (1996). On the Integration of Auditory and Visual
Parameters in an HMM-based ASR. In D. G. Stork & M. E. Hennecke (Eds.),
Speechreading by Humans and Machines: Models, Systems, and Applications,
pages 461-471. New York: NATO/Springer-Verlag.
Bernstein, L. & Benoit, C. (1996). For Speech Perception Three Senses are Bettern
748
1. Movellan and P. Mineiro
than One. In Proc. of the 4th Int. Conf. on Spoken Language Processing,
Philadelphia, PA., USA.
Box, G. E. P. (1980). Sampling and Bayes inference in scientific modeling. J. Roy.
Stat. Soc., A., 143, 383-430.
Bregler, C., Hild, H., Manke, S., & Waibel, A. (1993). Improving Connected Letter
Recognition by Lipreading. In Proc. Int. Conf. on Acoust., Speech, and Signal
Processing, volume 1, pages 557-560, Minneapolis. IEEE.
Biilt hoff, H. H. & Yuille, A. L. (1996). A Bayesian framework for the integration
of visual modules. In T. Inui & J. L. McClelland (Eds.), Attention and performance XVI: Information integmtion in perception and communication, pages
49-70. Cambridge, MA: MIT Press.
Chadderdon, G. & Movellan, J. (1995). Testing for Channel Independence in Bimodal Speech Recognition. In Proceedings of 2nd Joint Symposium on Neuml
Computation, pages 84-90.
Clark, J. J. & Yuille, A. L. (1990). Data Fusion for Sensory Information Processing
Systems. Boston: Kluwer Academic Publishers.
Dampster, A. P., Laird, N. M., & Rubin, D. B. (1977). Maximum likelihood from
incomplete data via the EM algorithm. J. Roy. Stat. Soc., 39, 1-38.
Efron, A. (1982). The jacknife, the bootstmp and other resampling plans. Philadelphia, Pennsylvania: SIAM.
Gray, M. S., Movellan, J. R., & Sejnowski, T. (1997). Dynamic features for visual
speechreading: A systematic comparison. In Mozer, Jordan, & Petsche (Eds.),
Advances in Neuml Information Processing Syste"ms, volume 9. MIT Press.
Movellan, J. R. (1995). Visual speech recognition with stochastic neural networks.
In G. Tesauro, D. Touretzky, & T. Leen (Eds.), Advances in neuml information
processing systems. Cambridge,Massacusetts: MIT Press.
Movellan, J. R. & Chadderdon, G. (1996). Channel Separability in the Audio
Visual Integration of Speech: A Bayesian Approach. In D. G. Stork & M. E.
Hennecke (Eds.), Speechreading by Humans and Machines: Models, Systems,
and Applications, pages 473-487. New York: NATO/Springer-Verlag.
Movellan, J. R. & Prayaga, R. S. (1996). Gabor Mosaics: A description of Local
Orientation Statistics with Applications to Machine Perception. In G. W.
Cottrell (Ed.), proceedings of the Eight Annual Conference of the Cognitive
Science Society, page 817. Mahwah, New Jersey: LEA.
O'Hagan, A. (1994). Kendall's Advanced Theory of Statistics: Volume 2B, Bayesian
Inference. volume 2B. Cambridge University Press.
Wolff, G. J., Prasad, K. V., Stork, D. G., & Hennecke, M. E. (1994). Lipreading by
Neural Networks: Visual Preprocessing, Learning and Sensory Integration. In
J. D. Cowan, G. Tesauro, & J. Alspector (Eds.), Advances in Neuml Information Processing Systems, volume 6, pages 1027-1034. Morgan Kaufmann.
| 1406 |@word trial:1 norm:1 loading:1 nd:1 open:1 grey:1 prasad:2 tried:1 covariance:1 decomposition:1 speechreading:3 hereafter:1 current:1 contextual:2 cottrell:1 drop:1 resampling:1 stationary:6 alone:1 lexicon:2 along:1 differential:1 symposium:1 consists:2 combine:2 redefine:1 expected:2 alspector:1 blackbox:1 audiovisual:3 inspired:1 automatically:3 provided:4 notation:1 what:2 interpreted:3 minimizes:1 spectrally:1 spoken:2 acoust:1 temporal:1 uo:1 arguably:1 local:3 tends:1 xv:2 syste:1 meet:1 statistically:2 minneapolis:1 unsophisticated:1 camera:1 testing:4 iqt:2 practice:3 movellan:17 digit:4 procedure:2 area:2 significantly:4 gabor:2 composite:1 word:1 road:1 flatten:1 interior:1 context:30 influence:4 applying:1 optimize:1 attention:1 assigns:1 rule:8 classic:3 variation:2 diego:3 mosaic:1 hypothesis:2 pa:1 roy:2 recognition:12 hagan:2 database:7 module:18 calculate:1 region:1 connected:1 contamination:4 principled:3 digitally:1 environment:6 mozer:1 ui:1 ideally:1 occluded:2 dynamic:2 trained:4 creation:1 yuille:9 upon:1 basis:2 ll2:1 multimodal:3 joint:4 easily:2 various:1 jersey:1 talk:2 separated:1 sejnowski:2 cogsci:1 outside:2 choosing:1 modular:1 encoded:1 statistic:3 think:1 jointly:2 gp:2 laird:2 final:1 subtracting:1 product:2 date:1 description:1 extending:1 produce:3 tulips1:1 stat:2 exemplar:1 qt:6 soc:2 implies:1 indicate:1 synchronized:1 closely:1 correct:2 stochastic:2 human:2 generalization:6 probable:4 bregler:2 hild:2 exp:1 talker:1 inexistent:2 estimation:1 proc:2 radio:3 label:1 highway:1 reflects:1 hope:1 mit:3 sensor:1 gaussian:4 occupied:1 factorizes:1 ax:2 likelihood:4 check:1 centroid:1 sense:1 inference:6 dependent:1 typically:5 integrated:1 entire:1 hidden:3 pixel:1 overall:1 arg:1 orientation:1 plan:1 integration:5 hoff:1 equal:1 asr:1 sampling:1 contaminated:4 simultaneously:1 recognize:1 argmax:5 consisting:1 occlusion:2 benoit:4 replicates:1 male:1 analyzed:1 mixture:6 sens:1 fu:1 incomplete:1 desired:1 modeling:3 soft:1 markovian:1 ar:1 logp:7 maximization:2 deviation:4 successful:1 front:1 optimally:1 combined:2 density:1 siam:1 systematic:1 off:1 manke:2 fused:2 worse:2 cognitive:4 conf:2 busy:1 student:1 coefficient:1 int:2 explicitly:1 later:2 view:1 kendall:1 doing:1 competitive:17 bayes:4 iwi:2 accuracy:1 variance:6 kaufmann:1 percept:1 ensemble:3 bayesian:15 none:2 explain:1 touretzky:1 ed:7 energy:1 mi:3 sampled:1 auditory:6 adjusting:2 knowledge:1 car:2 dimensionality:1 efron:2 javier:1 reflected:1 response:1 leen:1 done:5 box:5 evaluated:1 xa:1 implicit:1 until:1 hand:1 bulthoff:1 working:1 overlapping:1 gray:2 scientific:1 usa:1 white:5 conditionally:1 mahalanobis:1 ll:2 nuisance:1 speaker:1 m:1 dedicated:1 image:6 superior:1 behaves:1 stork:4 volume:5 linking:1 occurred:1 interpretation:5 he:2 kluwer:1 refer:1 significant:2 cambridge:3 automatic:2 language:1 had:2 operating:1 multivariate:1 female:1 perspective:3 jolla:2 phone:1 forcing:1 inui:1 certain:1 verlag:2 tesauro:2 lipreading:2 morgan:1 minimum:1 additional:1 maximize:1 signal:16 ii:2 unimodal:1 academic:1 characterized:1 iteration:1 represent:1 kernel:1 bimodal:1 achieved:2 cell:1 lea:1 receive:1 justified:1 uninformative:1 want:2 addressed:1 source:2 leaving:2 modality:1 publisher:1 extra:1 operate:2 subject:5 db:6 cowan:1 contrary:1 inconsistent:1 jordan:1 call:3 door:1 bernstein:2 variety:4 independence:7 architecture:4 pennsylvania:1 opposite:1 inner:1 idea:5 iogp:2 whether:1 pca:1 passed:1 speech:7 york:2 factorial:3 mcclelland:1 http:1 outperform:1 percentage:2 shifted:1 estimated:3 delta:2 correctly:1 track:1 per:4 four:1 clean:8 thresholded:1 fuse:1 concreteness:1 compete:2 package:1 letter:1 uncertainty:1 patch:2 decision:2 bit:1 correspondence:1 annual:1 strength:2 occur:1 worked:1 simulate:1 friction:1 relatively:1 jackknife:1 department:3 tv:1 according:1 waibel:2 combination:1 smaller:1 em:3 separability:1 wi:12 explained:2 restricted:2 taken:1 previously:1 discus:1 jacknife:1 whichever:1 end:1 available:2 hennecke:4 eight:1 petsche:1 robustly:1 alternative:3 robustness:1 symmetrized:5 neuml:4 running:1 classical:2 society:1 said:1 hmm:5 whom:1 reason:1 assuming:2 ratio:1 ql:1 unfortunately:1 implementation:1 design:2 perform:1 av:4 observation:3 vertical:1 markov:3 displayed:1 situation:1 communication:1 frame:10 ucsd:3 station:1 optimized:1 california:3 engine:2 xvi:1 address:1 perception:3 lpc:2 reliable:2 max:1 video:11 including:2 mouth:1 power:3 advanced:1 representing:1 axis:1 specializes:1 naive:1 utterance:2 philadelphia:2 prior:3 expect:1 telephony:1 clark:7 affine:1 rubin:2 pi:2 wide:3 template:3 cepstral:3 modularized:1 calculated:1 vocabulary:1 valid:2 dimension:1 sensory:4 collection:1 san:3 preprocessing:1 robustification:9 implicitly:1 nato:2 xtlqt:1 assumed:1 mineiro:4 modularity:2 table:3 disambiguate:1 channel:12 robust:1 ca:2 ean:1 improving:1 boxed:1 investigated:1 did:1 noise:6 paul:1 mahwah:1 repeated:1 retrain:1 msec:1 perceptual:8 weighting:1 third:1 down:1 bad:1 specific:2 xt:2 explored:2 fusion:29 consist:1 undergraduate:1 effectively:1 boston:1 simply:1 explore:2 visual:16 partially:1 scalar:1 springer:2 aa:6 corresponds:1 vav:1 ma:1 conditional:9 identity:1 consequently:2 replace:1 absence:1 change:4 infinite:1 typical:1 determined:1 microphone:1 wolff:2 called:1 total:1 catastrophic:7 experimental:1 la:2 arises:1 audio:18 tested:6 phenomenon:1 |
447 | 1,407 | Reinforcement Learning for Continuous
Stochastic Control Problems
Remi Munos
CEMAGREF, LISC, Pare de Tourvoie,
BP 121, 92185 Antony Cedex, FRANCE.
Rerni.Munos@cemagref.fr
Paul Bourgine
Ecole Polyteclmique, CREA,
91128 Palaiseau Cedex, FRANCE.
Bourgine@poly.polytechnique.fr
Abstract
This paper is concerned with the problem of Reinforcement Learning (RL) for continuous state space and time stocha.stic control
problems. We state the Harnilton-Jacobi-Bellman equation satisfied by the value function and use a Finite-Difference method for
designing a convergent approximation scheme. Then we propose a
RL algorithm based on this scheme and prove its convergence to
the optimal solution.
1
Introduction to RL in the continuous, stochastic case
The objective of RL is to find -thanks to a reinforcement signal- an optimal strategy
for solving a dynamical control problem. Here we sudy the continuous time, continuous state-space stochastic case, which covers a wide variety of control problems
including target, viability, optimization problems (see [FS93], [KP95])}or which a
formalism is the following. The evolution of the current state x(t) E 0 (the statespace, with 0 open subset of IRd ), depends on the control u(t) E U (compact subset)
by a stochastic differential equation, called the state dynamics:
(1)
dx = f(x(t), u(t))dt + a(x(t), u(t))dw
where f is the local drift and a .dw (with w a brownian motion of dimension rand
(j a d x r-matrix) the stochastic part (which appears for several reasons such as lake
of precision, noisy influence, random fluctuations) of the diffusion process.
For initial state x and control u(t), (1) leads to an infinity of possible traj~tories
x(t). For some trajectory x(t) (see figure I)., let T be its exit time from 0 (with
the convention that if x(t) always stays in 0, then T = 00). Then, we define the
functional J of initial state x and control u(.) as the expectation for all trajectories
of the discounted cumulative reinforcement :
J(x; u(.)) =
Ex,u( .)
{loT '/r(x(t), u(t))dt +,,{ R(X(T))}
R. Munos and P. Bourgine
1030
where rex, u) is the running reinforcement and R(x) the boundary reinforcement.
'Y is the discount factor (0 :S 'Y < 1). In the following, we assume that J, a are of
class C2 , rand Rare Lipschitzian (with constants Lr and LR) and the boundary
80 is C2 .
?
? all ?
II
?
?
xirJ
?
?
?
?
?
?
?
?
Figure 1: The state space, the discretized ~6 (the square dots) and its frontier 8~6
(the round ones). A trajectory Xk(t) goes through the neighbourhood of state ~.
RL uses the method of Dynamic Program~ing (DP) which generates an optimal
(feed-back) control u*(x) by estimating the value function (VF), defined as the
maximal value of the functional J as a function of initial state x :
(2)
Vex) = sup J(x; u(.).
u( .)
In the RL approach, the state dynamics is unknown from the system ; the only
available information for learning the optimal control is the reinforcement obtained
at the current state. Here we propose a model-based algorithm, i.e. that learns
on-line a model of the dynamics and approximates the value function by successive
iterations.
Section 2 states the Hamilton-Jacobi-Bellman equation and use a Finite-Difference
(FD) method derived from Kushner [Kus90] for generating a convergent approximation scheme. In section 3, we propose a RL algorithm based on this scheme and
prove its convergence to the VF in appendix A.
2
A Finite Difference scheme
Here, we state a second-order nonlinear differential equation (obtained from the DP
principle, see [FS93J) satisfied by the value function, called the Hamilton-JacobiBellman equation.
Let the d x d matrix a = a.a' (with' the transpose of the matrix). We consider
the uniformly pambolic case, Le. we assume that there exists c > 0 such that
V$ E 0, Vu E U, Vy E IR d ,2:t,j=l aij(x, U)YiYj 2: c1lY112. Then V is C2 (see [Kry80J).
Let Vx be the gradient of V and VXiXj its second-order partial derivatives.
Theorem 1 (Hamilton-Jacohi-Bellman) The following HJB equation holds :
Vex) In 'Y + sup [rex, u)
uEU
+ Vx(x).J(x, u) + ! 2:~j=l aij V
XiXj
(x)] = 0 for x E 0
Besides, V satisfies the following boundary condition: Vex) = R(x) for x E 80.
1031
Reinforcement Learningfor Continuous Stochastic Control Problems
Remark 1 The challenge of learning the VF is motivated by the fact that from V,
we can deduce the following optimal feed-back control policy:
u*(x) E arg sup [r(x, u)
uEU
+ Vx(x).f(x, u) + ! L:7,j=l aij VXiXj (x)]
In the following, we assume that 0 is bounded. Let eI, ... , ed be a basis for JRd.
Let the positive and negative parts of a function 4> be : 4>+ = ma.x(4),O) and
4>- = ma.x( -4>,0). For any discretization step 8, let us consider the lattices: 8Zd =
{8. L:~=1 jiei} where j}, ... ,jd are any integers, and ~6 = 8Z d n O. Let 8~6, the
frontier of ~6 denote the set of points {~ E 8Zd \ 0 such that at least one adjacent
point ~ ? 8ei E ~6} (see figure 1).
Let U 6 cUbe a finite control set that approximates U in the sense: 8 ~ 8' =>
U6' C U6 and U6U 6 = U. Besides, we assume that: Vi = l..d,
(3)
By replacing the gradient Vx(~) by the forward and backward first-order finitedifference quotients: ~;, V(~) = l [V(~ ? 8ei) - V(~)l and VXiXj (~) by the secondorder finite-difference quotients:
-b [V(~ + 8ei) + V(,'
~XiXi V(~)
~;iXj V(~)
- 8ei) - 2V(O]
= 2P[V(~ + 8ei ? 8ej) + V(~ - 8ei =F 8ej)
-V(~
+ 8ei) -
V(~
- 8ei) -
V(~
+ 8ej) -
V(~
- 8ej)
+ 2V(~)]
in the HJB equation, we obtain the following : for ~ E :?6,
V6(~)In,+SUPUEUh {r(~,u) + L:~=1 [f:(~,u)'~~iV6(~) - fi-(~,U)'~;iV6(~)
+
aii (~.u) ~ X,X,'"
V(C) + "
. .
2
. .
wJ'l=~
(at; (~,'U) ~ +x,x.1'"
.V(C) _ a~ (~,'U) ~ x,x V(C))] } = 0
2
2
-. .
J
'"
Knowing that (~t In,) is an approximation of (,l:l.t -1) as ~t tends to 0, we deduce:
V6(~)
with
T(~,
SUPuEUh
[,"'(~'U)L(EEbP(~,U,()V6?()+T(~,u)r(~,u)] (4)
(5)
u)
which appears as a DP equation for some finite Markovian Decision Process (see
[Ber87]) whose state space is ~6 and probabilities of transition:
p(~,u,~
p(~,
? 8ei)
u, ~ + 8ei ? 8ej)
p(~,u,~
- 8ei ? 8ej)
p(~,u,()
"'~~r) [28Ift(~, u)1 + aii(~' u) - Lj=l=i laij(~, u)l] ,
"'~~r)a~(~,u)fori=f:j,
(6)
"'~~r)a~(~,u) for i =f: j,
o otherwise.
Thanks to a contraction property due to the discount factor" there exists a unique
solution (the fixed-point) V to equation (4) for ~ E :?6 with the boundary condition
V6(~) = R(~) for ~ E 8:?6. The following theorem (see [Kus90] or [FS93]) insures
that V 6 is a convergent approximation scheme.
R. Munos and P. Bourgine
1032
Theorem 2 (Convergence of the FD scheme) V D converges to V as 8 1 0 :
lim
VD(~)
/)10
~-x
= Vex)
un~formly on 0
Remark 2 Condition (3) insures that the p(~, u, () are positive. If this condition
does not hold, several possibilities to overcome this are described in [Kus90j.
3
The reinforcement learning algorithm
Here we assume that f is bounded from below. As the state dynami,:s (J and a)
is unknown from the system, we approximate it by building a model f and from
samples of trajectories Xk(t) : we consider series of successive states Xk = Xk(tk)
and Yk = Xk(tk + Tk) such that:
a
- "It E [tk, tk + Tk], x(t) E N(~) neighbourhood of ~ whose diameter is inferior to
kN.8 for some positive constant kN,
- the control u is constant for t E [tk, tk
- Tk
+ Tk],
satisfies for some positive kl and k2,
(7)
Then incrementally update the model :
.1 ",n
n ~k=l
n (Yk - Xk -;;; Lk=l
1
an(~,u)
Yk - Xk
Tk
Tk.fn(~, u))
and compute the approximated time T( x, u)
transition p(~, u, () by replacing f and a by
(Yk - Xk Tk
Tk?fn(~, u))'
(8)
~d
f
the approximated probabilities of
and a in (5) and (6).
We obtain the following updating rule of the V D-value of state ~ :
V~+l (~)
=
sUPuEU/)
[,~/:(x,u) L( p(~, u, ()V~(() + T(x, u)r(~, u)]
(9)
which can be used as an off-line (synchronous, Gauss-Seidel, asynchronous) or ontime (for example by updating V~(~) as soon as a trajectory exits from the neighbourood of ~) DP algorithm (see [BBS95]).
Besides, when a trajectory hits the boundary [JO at some exit point Xk(T) then
update the closest state ~ E [JED with:
(10)
Theorem 3 (Convergence of the algorithm) Suppose that the model as well
as the V D-value of every state ~ E :ED and control u E U D are regularly updated
(respectively with (8) and (9)) and that every state ~ E [JED are updated with (10)
at least once. Then "Ie> 0, :3~ such that "18 ~ ~, :3N, "In 2: N,
sUP~EE/) IV~(~) - V(~)I ~
e
with probability 1
1033
Reinforcement Learningfor Continuous Stochastic Control Problems
4
Conclusion
This paper presents a model-based RL algorithm for continuous stochastic control
problems. A model of the dynamics is approximated by the mean and the covariance
of successive states. Then, a RL updating rule based on a convergent FD scheme is
deduced and in the hypothesis of an adequate exploration, the convergence to the
optimal solution is proved as the discretization step 8 tends to 0 and the number
of iteration tends to infinity. This result is to be compared to the model-free RL
algorithm for the deterministic case in [Mun97]. An interesting possible future
work should be to consider model-free algorithms in the stochastic case for which a
Q-Iearning rule (see [Wat89]) could be relevant.
A
Appendix: proof of the convergence
Let M f ' M a, M fr. and Ma .? be the upper bounds of j, a, f x and 0' x and m f the lower
bound of f. Let EO = SUP?EI:h !V0 (';) - V(';)I and E! = SUP?EI:b \V~(';) - VO(.;)\.
Estimation error of the model fn and an and the probabilities Pn
A.I
Suppose that the trajectory Xk(t) occured for some occurence Wk(t) of the brownian
motion: Xk(t) = Xk + f!k f(Xk(t),u)dt + f!" a(xk(t),U)dwk. Then we consider a
trajectory Zk (t) starting from .; at tk and following the same brownian motion:
Zk(t) ='; + fttk. f(Zk(t), u)dt + fttk a(zk(t), U)dWk'
Let Zk = Zk(tk + Tk). Then (Yk - Xk) - (Zk -.;) =
ftk [f(Xk(t), u) - f(Zk(t), u)] dt +
t
Tk
ft
[a(xk(t), u) - a(zk(t), u)J dWk. Thus, from the C1 property of f and a,
:.+
II(Yk -
Xk) - (Zk - ';)11 ~ (Mf'"
+ M aJ.kN.Tk. 8.
(11)
The diffusion processes has the following property ~ee for example the ItO- Taylor
majoration in [KP95j) : Ex [ZkJ = ';+Tk.f(';, U)+O(Tk) which, from (7), is equivalent
to: Ex [z~:g]
= j(';,u) + 0(8). Thus from the law of large numbers and (11):
li~-!~p Ilfn(';, u) -
f(';, u)11
-
li;;:s~p II~ L~=l [Yk;kX
& -
(Mf:r:
?.] I + 0(8)
+ M aJ?kN?8 + 0(8) =
0(8) w.p. 1 (12)
Besides, diffusion processes have the following property (again see [KP95J):
Ex [(Zk -.;) (Zk - .;)'] = a(';, U)Tk + f(';, u).f(';, U)'.T~ + 0(T2) which, from (7),
is equivalent to: Ex [(Zk-?-Tkf(S'U)~(kZk-S-Tkf(S'U?/] = a(';, u) + 0(82). Let rk =
= Yk - Xk - Tkfn(';, u) which satisfy (from (11) and (12? :
Ilrk - ikll = (Mf:r: + M aJ.Tk.kN.8 + Tk.o(8)
(13)
Zk -.; - Tkf(';, u) and ik
From the definition of Ci;;(';,u), we have: Ci;;(';,u) - a(';,u)
Ex [r~':k]
+ 0(8 2 )
li~~~p
=
~L~=l
'\:1.' -
and from the law of large numbers, (12) and (13), we have:
11~(';,u) -
a(';,u)11 =
Ilik -rkllli:!s!p~
li~-!~p II~ L~=l rJ./Y
fl (II~II + II~II)
- r~':k I + 0(8
+0(82 ) = 0(8 2 )
2)
R. Munos and P. Bourgine
1034
"In(';, u) - I(';, u)" ~
1Ill;;(';, u) - a(';, u)1I
kf?8 w.p. 1
~ ka .82 w.p. 1
(14)
Besides, from (5) and (14), we have:
2
2
c ) _ -Tn (cr.",U )1 _
< d.(k[.6
kT'UJ:2
1T(r.",U
(d.m,+d.k,,6
.6)2 ) UJ:2 <
_
(15)
and from a property of exponential function,
I,T(~.u) _
,7' . .1?) I= kT.In ~ .8
(?
2.
(16)
We can deduce from (14) that:
.
1(
)
-(
limsupp';,u,( -Pn';,u,(
)1
~
n-+oo
(2.6.Mt+d.Ma)(2.kt+d.k,,)62
k J:
6mr(2.k,+d.ka)62
S; puw.p.l
Estimation of IV~+l(';) - V 6(.;) 1
A.2
Mter having updated V~(';) with rule (9),
IV~+l(';) - V 6(.;) From (4), (9) and (8),
I.
A
(17)
let A denote the difference
< ,T(?.U) L: [P(';, u, () - p(.;, u, ()] V 6(() + (,T(?.1?) - ,7'(~'1??) L p(.;, u, () V 6 (()
(
(
+,7' (?.u) . L:p(.;, u, () [V6(() - V~(()] + L:p(.;, u, ().T(';, u) [r(';, u) - F(';, u)]
(
(
+ L:( p(.;, u, () [T(';, u) -
T(';, u)] r(';, u) for all u E U6
As V is differentiable we have : Vee) = V(';) + VX ' ( ( - . ; ) + 0(1I( - ';11). Let
us define a linear function V such that: Vex) = V(';) + VX ' (x - ';). Then
we have: [P(';, u, () - p(.;, u, ()] V 6(() = [P(';, u, () - p(.;, u, ()] . [V6(() - V(()] +
[P(';,u,()-p(';,u,()]V((), thus: L:([p(';,u,()-p(';,u,()]V6(() = kp .E6.8 +
[V(() +0(8)] = [V(7J)-VUD] + kp .E6.8 + 0(8)
with: 7J = L:( p(';, u, () (( -.;) and 1j = L:( p(.;, u, () (( -
.;).
Besides, from the convergence of the scheme (theorem 2), we have E6.8
=
L([P(';,U,()-p(.;,u,()]
[V(7J) - V(1j)] + 0(8)
0(8). From the linearity of
V,
IL( [P(';, u, () - p(.;, u, ()] V6 (() I
IV(() -
I II( - ZII?Mv",
V(Z) ~
=
S; 2kp 82 . Thus
= 0(8) and from (15), (16) and the Lipschitz prop-
erty of r,
A=
1'l'(?'U), L:( p(.;, u, () [V6(() - V~ (()] 1+ 0(8).
As ,..,.7'(?.u) < 1 - 7'(?.U) In 1 < 1 _ T(?.u)-k.,.6 2 In 1 < 1 _ (
I
-
2
'Y -
2
we have:
A
with k
= 2d(M[~d.M,,).
= (1 -
k.8)E~
'Y -
+ 0(8)
6
2d(M[+d.M,,)
_
!ix..82) In 1
2
'Y '
(18)
1035
Reinforcement Learning for Continuous Stochastic Control Problems
A.3
A sufficient condition for sUP?EE~ IV~(~) - V6(~)1 :S
C2
Let us suppose that for all ~ E ~6, the following conditions hold for some a > 0
E~ > C2 =} IV~+I(O - V6(~)1 :S E~ - a
(19)
E~ :S c2=}IV~+I(~)_V6(~)I:Sc2
(20)
From the hypothesis that all states ~ E ~6 are regularly updated, there exists an
integer m such that at stage n + m all the ~ E ~6 have been updated at least
once since stage n. Besides, since all ~ E 8C 6 are updated at least once with
rule (10), V~ E 8C6, IV~(~) - V6(~)1 = IR(Xk(T)) - R(~)I :S 2.LR.8 :S C2 for any
8 :S ~3 = 2~lR' Thus, from (19) and (20) we have:
E! >
E! :S
C2 =}
E!+m :S E! - a
C2 =}
E!+m :S
Thus there exists N such that: Vn ~ N, E~ :S
A.4
C2
C2.
Convergence of the algorithm
Let us prove theorem 3. For any c > 0, let us consider Cl > 0 and C2 > 0 such that
= c. Assume E~ > ?2, then from (18), A = E! - k.8'?2+0(8) :S E~ -k.8.~
for 8 :S ~3. Thus (19) holds for a = k.8.~. Suppose now that E~ :S ?2. From (18),
A :S (1 - k.8)?2 + 0(8) :S ?2 for 8 :S ~3 and condition (20) is true.
Cl +C2
Thus for 8 :S min {~1, ~2, ~3}, the sufficient conditions (19) and (20) are satisfied.
So there exists N, for all n ~ N, E~ :S ?2. Besides, from the convergence of the
scheme (theorem 2), there exists ~o st. V8:S ~o, sUP?EE~ 1V6(~) - V(~)I :S ?1?
Thus for 8 :S
min{~o, ~1, ~2, ~3},
"3N, Vn
~
sup IV~(~) - V(~)I :S sup IV~(~) - V6(~)1
?EE6
?EEh
N,
+ sup
?EE6
1V6(~) - V(~)I :S
?1
+ c2 = ?.
References
[BBS95j Andrew G. Barto, Steven J. Bradtke, and Satinder P. Singh. Learning to
act using real-time dynamic programming. Artificial Intelligence, (72):81138, 1995.
[Ber87j Dimitri P. Bertsekas. Dynamic Programming: Deterministic and Stochastic Models. Prentice Hall, 1987.
[FS93j Wendell H. Fleming and H. Mete Soner. Controlled Markov Processes and
Viscosity Solutions. Applications of Mathematics. Springer-Verlag, 1993.
[KP95j Peter E. Kloeden and Eckhard Platen. Numerical Solutions of Stochastic
Differential Equations. Springer-Verlag, 1995.
[Kry80j N.V. Krylov. Controlled Diffusion Processes. Springer-Verlag, New York,
1980.
[Kus90j Harold J. Kushner. Numerical methods for stochastic control problems in
continuous time. SIAM J. Control and Optimization, 28:999-1048, 1990.
[Mun97j Remi Munos. A convergent reinforcement learning algorithm in the continuous case based on a finite difference method. International Joint Conference on Art~ficial Intelligence, 1997.
[Wat89j Christopher J.C.H. Watkins. Learning from delayed reward. PhD thesis,
Cambridge University, 1989.
Use of a Multi-Layer Percept ron to
Predict Malignancy in Ovarian Tumors
Herman Verrelst,
Yves Moreau and Joos Vandewalle
Dept. of Electrical Engineering
Katholieke Universiteit Leuven
Kard. Mercierlaan 94
B-3000 Leuven, Belgium
Dirk Timmerman
Dept. of Obst. and Gynaec.
University Hospitals Leuven
Herestraat 49
B-3000 Leuven, Belgium
Abstract
We discuss the development of a Multi-Layer Percept ron neural
network classifier for use in preoperative differentiation between
benign and malignant ovarian tumors. As the Mean Squared classification Error is not sufficient to make correct and objective assessments about the performance of the neural classifier, the concepts of sensitivity and specificity are introduced and combined
in Receiver Operating Characteristic curves. Based on objective
observations such as sonomorphologic criteria, color Doppler imaging and results from serum tumor markers, the neural network is
able to make reliable predictions with a discriminating performance
comparable to that of experienced gynecologists.
1
Introd uction
A reliable test for preoperative discrimination between benign and malignant ovarian tumors would be of considerable help to clinicians. It would assist them to select
patients for whom minimally invasive surgery or conservative management suffices
versus those for whom urgent referral to a gynecologic oncologist is needed.
We discuss the development of a neural network classifier/diagnostic tool. The neural network was trained by supervised learning, based on data from 191 thoroughly
examined patients presenting with ovarian tumors of which 140 were benign and 51
malignant. As inputs to the network we chose indicators that in recent studies have
proven their high predictive value [1, 2, 3]. Moreover, we gave preference to those
indicators that can be obtained in an objective way by any gynecologist. Some of
these indicators have already been used in attempts to make one single protocol or
decision algorithm [3, 4].
Use of a MLP to Predict Malignancy in Ovarian Tumors
979
In order to make reliable assessments on the practical performance of the classifier,
it is necessary to work with other concepts than Mean Squared classification Error
(MSE), which is traditionally used as a measure of goodness in the training of a
neural network. We will introduce notions as specificity and sensitivity and combine
them into Receiver Operating Characteristic (ROC) curves. The use of ROC-curves
is motivated by the fact that they are independent of the relative proportion of the
various output classes in the sample population. This enables an objective validation
of the performance of the classifier. We will also show how, in the training of the
neural network, MSE optimization with gradient methods can be refined and/or
replaced with the help of ROC-curves and simulated annealing techniques.
The paper is organized as follows. In Section 2 we give a brief description of the
selected input features. In Section 3 we state some drawbacks to the MSE criterion
and introduce the concepts of sensitivity, specificity and ROC-curves. Section 4
then deals with the technicalities of training the neural network. In Section 5 we
show the results and compare them to human performance.
2
Data acquisition and feature selection
The data were derived from a study group of 191 consecutive patients who were
referred to a single institution (University Hospitals Leuven, Belgium) from August
1994 to August 1996. Table 1 lists the different indicators which were considered,
together with their mean value and standard deviations or together with the relative
presence in cases of benign and malignant tumors.
Table 1
Indicator
Demographic
Age
Postmenopausal
CA 125 (log)
Blood flow present
Abdominal fluid
Bilateral mass
Unilocular cyst
Multiloc/solid cyst
Smooth wall
Irregular wall
Papillations
Serum marker
CD!
Morphologic
Benign
Malignant
49.3 ? 16.0
40%
2.8?1.1
72.9%
12.1%
11.4%
42.1%
16.4%
58.6%
32.1%
7.9%
58.3 ? 14.3
70.6%
5.2 ? 1.9
100%
52.9%
35.3%
5.9%
49.0%
2.0%
76.5%
74.5%
Table 1: Demographic, serum marker, color Doppler imaging and morphologic indicators.
For the continuous valued features the mean and standard deviation for each class are
reported. For binary valued indicators, the last two columns give the presence of the
feature in both classes e.g. only 2% of malignant tumors had smooth walls.
First, all patients were scanned with ultrasonography to obtain detailed gray-scale
images of the tumors. Every tumor was extensively examined for its morphologic
characteristics. Table 1 lists the selected morphologic features: presence of abdominal fluid collection, papillary structures (> 3mm), smooth internal walls, wall
irregularities, whether the cysts were unilocular, multilocular-solid and/or present
on both pelvic sides. All outcomes are binary valued: every observation relates to
the presence (1) or absence (0) of these characteristics.
Secondly, all tumors were entirely surveyed by color Doppler imaging to detect
presence or absence of blood flow within the septa, cyst walls, solid tumor areas or
ovarian tissue. The outcome is also binary valued (1/0).
H. Verrelst, Y. Moreau, 1. Vandewalle and D. TImmennan
980
Thirdly, in 173 out of the total of 191 patients, serum CA 125 levels were measured,
using CA 125 II immunoradiometric assays (Centocor, Malvern, PA). The CA 125
antigen is a glycoprotein that is expressed by most epithelial ovarian cancers. The
numerical value gives the concentration in U Iml. Because almost all values were
situated in a small interval between 0 and 100, and because a small portion took
values up to 30,000, this variable was rescaled by taking its logarithm.
Since age and menopausal status of the patient are considered to be highly relevant,
these are also included. The menopausal score is -1 for premenopausal, + 1 for
postmenopausal. A third class of patients were assigned a 0 value. These patients
had had an hysterectomy, so no menopausal status could be appointed to them.
It is beyond the scope of this paper to give a complete account of the meaning of
the different features that are used or the way in which the data were acquired.
We will limit ourselves to this short description and refer the reader to [2, 3] and
gynecological textbooks for a more detailed explanation.
3
3.1
Receiver Operating Characteristics
Drawbacks to Mean Squared classification Error
Let us assume that we use a one-hidden-Iayer feed-forward NN with m inputs
nh hidden neurons with the tanh(.) as activation function, and one output i1k,
m
nh
Yk(B) =
L
xl,
Wj
tanh(L VijX~
j=l
+ {3j)'
(1)
i=l
parameterized by the vector 0 consisting of the network's weights Wj and Vij and
bias terms {3j. The cost function is often chosen to be the squared difference between
the desired d k and the actual response Yk. averaged over all N samples [12],
1 N
J(O) = N
2:)dk -
Yk(9))2.
(2)
k=l
This type of cost function is continuous and differentiable, so it can be used in
gradient based optimization techniques such as steepest descent (back-propagation),
quasi-Newton or Levenberg-Marquardt methods [8, 9, 11, 12]. However there are
some drawbacks to the use of this type of cost function.
First of all, the MSE is heavily dependent on the relative proportion of
the different output classes in the training set. In our dichotomic case this
can easily be demonstrated by writing the cost function, with superscripts b and m
respectively meaning benign and malignant, as
J(O)
-
Nb
Nb + N m
1 ""Nb (db
)2
Nb wk=l k - Yk
Nm
+ Nb + N m
~
~
,\
(1-,\)
1 ""Nm
N m wk=l
(dm
k - Yk
)2
(3)
If the relative proportion in the sample population is not representative for reality,
the .x parameter should be adjusted accordingly. In practice this real proportion is
often not known accurately or one simply ignores the meaning of .x and uses it as a
design parameter in order to bias the accuracy towards one of the output classes.
A second drawback of the MSE cost function is that it is not very in-
formative towards practical usage of the classification tool. A clinician is
not interested in the averaged deviation from desired numbers, but thinks in terms
of percentages found, missed or misclassified. In the next section we will introduce
the concepts of sensitivity and specificity to express these more practical measures.
Use of a MLP to Predict Malignancy in Ovarian Tumors
3.2
981
Sensitivity, specificity and ROC-curves
If we take the desired response to be 0 for benign and 1 for malignant cases, the
way to make clear cut (dichotomic) decisions is to compare the numerical outcome
of the neural network to a certain threshold value T between 0 and 1. When the
outcome is above the threshold T, the prediction is said to be positive. Otherwise the
prediction is said to be negative. With this convention, we say that the prediction
was
True Positive (TP)
if the prediction was positive when the sample was malignant.
True Negative (TN) if the prediction was negative when the sample was benign.
False Positive (FP)
if the prediction was positive when the sample was benign.
False Negative (FN) if the prediction was negative when the sample was malignant.
To every of the just defined terms T P, TN, F P and F N, a certain subregion of the
total sample space can be associated, as depicted in Figure 1. In the same sense,
we can associate to them a certain number counting the samples in each subregion.
We can then define sensitivity as
the proportion of malignant cases that
Tl:FN'
Total opulation
?~li~nant
....
. " ...
,
\
TP
-
'-
,,
"
Figure 1: The concepts of true and false positive and negative illustrated. The dashed
area indicates the malignant cases in the total sample population. The positive prediction
of an imperfect classification (dotted area) does not fully coincide with this sub area.
F::rN'
are predicted to be malignant and specificity as
the proportion of benign
cases that are predicted to be benign. The false positive rate is I-specificity.
When varying the threshold T, the values of T P, TN, F P, F N and therefore
also sensitivity and specificity, will change. A low threshold will detect almost all
malignant cases at the cost of many false positives. A high threshold will give
less false positives, but will also detect less malignant cases. Receiver Operating
Characteristic (ROC) curves are a way to visualize this relationship. The plot gives
the sensitivity versus false positive rate for varying thresholds T (e.g. Figure 2).
The ROC-curve is useful and widely used device for assessing and comparing the
value of tests [5, 7]. The proportion of the whole area of the graph which lies below
the ROC-curve is a one-value measure of the accuracy of a test [6]. The higher
this proportion, the better the test. Figure 2 shows the ROC-curves for two simple
classifiers that use only one single indicator. (Which means that we classify a tumor
being malignant when the value of the indicator rises above a certain value.) It is
seen that the CA 125 level has high predictive power as its ROC-curve spans 87.5%
of the total area (left Figure 2). For the age parameter, the ROC-curve spans
only 65.6% (right Figure 2). As indicated by the horizontal line in the plot, a CA
125 level classification will only misclassify 15% of all benign cases to reach a 80%
sensitivity, whereas using only age, one would then misclassify up to 50% of them.
H. Verrelst, Y. Moreau, 1. Vandewalle and D. Timmennan
982
:.... I(
'f
.,
,r
o? ? ?
,J
"
. ,
..
t f
1
t,
'1
U
U
??
,.I
..
0'
??
"
1
Figure 2: The Receiver Operating Characteristic (ROC) curve is the plot of the sensitivity versus the false positive rate of a classifier for varying thresholds used. Only single
indicators (left: CA 125, right: age) are used for these ROC-curlVes. The horizontal line
marks the 80% specificity level.
Since for every set of parameters of the neural network the area under the ROCcurve can be calculated numerically, this one-value measure can also be used for
supervised training, as will be shown in the next Section.
4
4.1
Simulation results
Inputs and architecture
The continuous inputs were standardized by subtracting their mean and dividing
by their standard deviation (both calculated over the entire population). Binary
valued inputs were left unchanged. The desired outputs were labeled 0 for benign
examples, 1 for malignant cases. The data set was split up: 2/3 of both benign and
malignant samples were randomly selected to form the training set. The remaining
examples formed the test set. The ratio of benign to all examples is >. ~ j.
Since the training set is not large, there is a risk of overtraining when too many
parameters are used. We will limit the number of hidden neurons to nh = 3 or 5.
As the CA 125 level measurement is more expensive and time consuming, we will
investigate two different classifiers: one which does use the CA 125 level and one
which does not. The one-hidden-Iayer MLP architectures that are used, are 11-3-1
and 10-5-1. A tanh(.) is taken for the activation function in the hidden layer.
4.2
Training
A first way of training was MSE optimization using the cost function (3) . By taking
~ in this expression, the role of malignant examples is more heavily weighted.
The parameter vector e was randomly initialized (zero mean Gaussian distribution,
standard deviation a = 0.01). Training was done using a quasi-Newton method with
BFGS-update of the Hessian (fminu in Matlab) [8, 9]. To prevent overtraining,
the training was stopped before the MSE on the test set started to rise. Only few
iterations (~ 100) were needed.
>. =
A second way of training was through the use of the area spanned by the ROC-curve
of the classifier and simulated annealing techniques [10]. The area-measure AROC
was numerically calculated for every set of trial parameters: first the sensitivity
and false positive rate were calculated for 1000 increasing values of the threshold
T between 0 and 1, which gave the ROC-curve; secondly the area AROC under the
curve was numerically calculated with the trapezoidal integration rule.
983
Use of a MLP to Predict Malignancy in Ovarian Tumors
We used Boltzmann Simulated Annealing to maximize the ROC-area. At time
k a trial parameter set of the neural network OHl is randomly generated in the
neighborhood of the present set Ok (Gaussian distribution, a = O.OO~. The trial
set 8H1 is always accepted if the area Af.2? 2: Afoc. If Af'?? < Ak OC, Ok+! is
accepted if
A r:g? - A r;oc
(
ROC
)/T.
e
Ak
>Q
with Q a uniformly distributed random variable E [0,1] and Te the temperature. As
cooling schedule we took
Te = 1/(100 + 10k),
so that the annealing was low-temperature and fast-cooling. The optimization was
stopped before the ROC-area calculated for the test set started to decrease. Only
a few hundred annealing epochs were allowed.
4.3
Results
Table 2 states the results for the different approaches. One can see that adding the
CA 125 serum level clearly improves the classifier's performance. Without it, the
ROC-curve spans about 96.5% of the total square area of the plot, whereas with
the CA 125 indicator it spans almost 98%. Also, the two training methods are seen
to give comparable results. Figure 3 shows the ROC-curve calculated for the total
population for the 11-3-1 MLP case, trained with simulated annealing
Table 2
MLP,
MLP,
MLP,
MLP,
10-5-1
10-5-1
11-3-1
11-3-1
Training set
96.7%
96.6%
97.9%
97.9%
MSE
SA
MSE
SA
Test set
96.4%
96.2%
97.4%
97.5%
Total population
96.5%
96.4%
97.7%
97.8%
Table 2: For the two architectures (10-5-1 and 11-3-1) of the MLP and for the gradient
(MSE) and the simulated annealing (SA) optimization techniques, this table gives the
resulting areas under the ROC-curves .
..
07
?0
0. 1
G.2
":I
04
0.'
0,
07
01
0,
,
Figure 3: ROC-curves of 11-3-1 MLP (with CA 125 level indicator), trained with simulated
annealing. The curve, calculated for the total population, spans 97.8% of the total region.
All patients were examined by two gynecologists, who gave their subjective impressions and also classified the ovarian tumors into (probably) benign and malignant.
Histopathological examinations of the tumors afterwards showed these gynecologists
H. Vendst, Y. Moreau, 1. Vandewalle and D. Timmerman
984
to have a sensitivity up to 98% and a false positive rate of 13% and 12% respectively. As can be seen in Figure 3, the 11-3-1 MLP has a similar performance. For
a sensitivity of 98%, its false positive rate is between 10% and 15%.
5
Conclusion
In this paper we have discussed the development of a Multi-Layer Perceptron neural
network classifier for use in preoperative differentiation between benign and malignant ovarian tumors. To assess the performance and for training the classifiers, the
concepts of sensitivity and specificity were introduced and combined in Receiver
Operating Characteristic curves. Based on objective observations available to every gynecologist, the neural network is able to make reliable predictions with a
discriminating performance comparable to that of experienced gynecologists.
Acknowledgments
This research work was carri e d out at the ESAT laboratory and the Interdisciplinary Center of Neural Networks
ICNN of the Katholieke Universiteit Leuven, in the following frameworks : the Belgian Programme on Interuniversity Poles of Attraction, initiated by the Belgian State, Prime Minister 's Office for Science, Technology and
Culture (IUAP P4-02 and IUAP P4-24), a Concerted Action Project MIPS (Model based Information Processing
Systems) of the Flemish Community and the FWO (Fund for Scientific Research - Flanders) project G.0262.97 :
Learning and Optimization: an Interdisciplinary Approach . The scientific responsibility rests with its authors .
References
[1] Bast R. C., Jr., Klug T .L. St. John E., et aI, "A radioimmunoassay using a monoclonal
antibody to monitor the course of epithelial ovarian cancer," N. Engl. J . Med., Vol.
309, pp. 883-888, 1983
[2] Timmerman D ., Bourne T ., Tailor A., Van Assche F .A., Vergote 1., "Preoperative
differentiation between benign and malignant adnexal masses," submitted
[3] Tailor A., Jurkovic D ., Bourne T.H., Collins W.P., Campbell S., "Sonographic prediction of malignancy in adnexal masses using multivariate logistic regression analysis,"
Ultrasound Obstet. Gynaecol. in press, 1997
[4] Jacobs 1., Oram D., Fairbanks J ., et aI. , "A risk of malignancy index incorporating
CA 125, ultrasound and menopausal status for the accurate preoperative diagnosis
of ovarian cancer," Br. J. Obstet. Gynaecol., Vol. 97, pp. 922-929, 1990
[5] Hanley J.A., McNeil B., "A method of comparing the areas under the receiver operating characteristics curves derived from the same cases," Radiology, Vol. 148, pp.
839-843, 1983
[6J
Swets J.A., "Measuring the accuracy of diagnostic systems," Science, Vol. 240, pp.
1285-1293, 1988
[7J Galen R.S., Gambino S., Beyond normality: the predictive value and efficiency of
medical diagnosis, John Wiley, New York, 1975.
[8J Gill P., Murray W ., Wright M., Practical Optimization, Acad. Press, New York, 1981
[9] Fletcher R., Practical methods of optimization, 2nd ed., John Wiley, New York, 1987.
[10J Kirkpatrick S., Gelatt C.D., Vecchi M., "Optimization by simulated annealing," Science, Vol. 220, pp . 621-680, 1983.
[11] Rumelhart D.E., Hinton G.E., Williams R.J ., "Learning representations by backpropagating errors," Nature, Vol. 323, pp. 533-536, 1986.
[12] Bishop C., Artificial Neural Networks for Pattern Recognition, OUP, Oxford, 1996
| 1407 |@word trial:3 proportion:8 nd:1 open:1 verrelst:3 kus90j:2 simulation:1 contraction:1 covariance:1 jacob:1 bourgine:5 solid:3 initial:3 series:1 score:1 ecole:1 kry80j:2 subjective:1 current:2 discretization:2 ixj:1 ka:2 marquardt:1 activation:2 comparing:2 dx:1 john:3 fn:5 numerical:4 benign:18 enables:1 plot:4 update:3 fund:1 discrimination:1 intelligence:2 ficial:1 selected:3 device:1 accordingly:1 xk:20 steepest:1 short:1 lr:4 institution:1 finitedifference:1 ron:2 successive:3 preference:1 c6:1 ohl:1 zii:1 c2:14 differential:3 zkj:1 ik:1 prove:3 combine:1 hjb:2 introduce:3 concerted:1 acquired:1 swets:1 multi:3 discretized:1 bellman:3 discounted:1 ee6:2 actual:1 increasing:1 project:2 estimating:1 bounded:2 linearity:1 moreover:1 erty:1 mass:3 textbook:1 differentiation:3 every:8 act:1 iearning:1 k2:1 hit:1 classifier:12 control:19 medical:1 hamilton:3 bertsekas:1 katholieke:2 positive:19 before:2 engineering:1 local:1 tends:3 limit:2 flemish:1 acad:1 ak:2 oxford:1 initiated:1 fluctuation:1 chose:1 minimally:1 tory:1 examined:3 antigen:1 averaged:2 unique:1 practical:5 acknowledgment:1 vu:1 practice:1 irregularity:1 area:16 specificity:10 selection:1 prentice:1 nb:5 influence:1 writing:1 risk:2 equivalent:2 deterministic:2 demonstrated:1 center:1 go:1 serum:5 starting:1 williams:1 rule:6 attraction:1 spanned:1 u6:3 dw:2 population:7 notion:1 traditionally:1 updated:6 target:1 suppose:4 heavily:2 programming:2 us:2 designing:1 hypothesis:2 secondorder:1 pa:1 associate:1 rumelhart:1 approximated:3 expensive:1 updating:3 trapezoidal:1 recognition:1 cut:1 cooling:2 labeled:1 ft:1 steven:1 role:1 electrical:1 wj:3 region:1 decrease:1 rescaled:1 yk:13 reward:1 dynamic:7 trained:3 singh:1 solving:1 predictive:3 exit:3 efficiency:1 basis:1 easily:1 aii:2 joint:1 various:1 fast:1 kp:3 artificial:2 outcome:4 refined:1 neighborhood:1 whose:2 widely:1 valued:5 say:1 otherwise:2 think:1 radiology:1 noisy:1 superscript:1 differentiable:2 vud:1 took:2 propose:3 subtracting:1 maximal:1 fr:3 p4:2 relevant:2 description:2 mete:1 convergence:9 assessing:1 jed:2 generating:1 converges:1 tk:24 help:2 oo:2 andrew:1 measured:1 sa:3 subregion:2 dividing:1 quotient:2 predicted:2 convention:2 aroc:2 drawback:4 yiyj:1 correct:1 stochastic:13 exploration:1 vx:6 human:1 galen:1 suffices:1 wall:6 icnn:1 secondly:2 adjusted:1 frontier:2 hold:4 mm:1 hall:1 considered:2 wright:1 fletcher:1 scope:1 predict:4 visualize:1 consecutive:1 belgium:3 estimation:2 epithelial:2 tanh:3 fs93j:2 nant:1 tool:2 weighted:1 clearly:1 always:2 gaussian:2 pn:2 ej:6 cr:1 varying:3 barto:1 office:1 sudy:1 derived:3 indicates:1 sense:2 detect:3 dependent:1 nn:1 lj:1 entire:1 hidden:5 quasi:2 france:2 interested:1 misclassified:1 arg:1 classification:6 ill:1 development:3 art:1 integration:1 cube:1 once:3 having:1 bourne:2 future:1 t2:1 few:2 randomly:3 delayed:1 replaced:1 ourselves:1 consisting:1 attempt:1 misclassify:2 mlp:12 fd:3 abdominal:2 possibility:1 highly:1 investigate:1 kirkpatrick:1 kt:3 accurate:1 partial:1 necessary:1 belgian:2 culture:1 iv:10 taylor:1 logarithm:1 initialized:1 desired:4 obst:1 stopped:2 eckhard:1 formalism:1 column:1 classify:1 markovian:1 cover:1 tp:2 goodness:1 engl:1 measuring:1 lattice:1 cost:7 pole:1 deviation:5 subset:2 rare:1 hundred:1 vandewalle:4 too:1 rex:2 reported:1 kn:5 antony:1 combined:2 thoroughly:1 thanks:2 deduced:1 st:2 siam:1 international:1 ie:1 stay:1 sensitivity:14 discriminating:2 off:1 interdisciplinary:2 together:2 jo:1 again:1 thesis:1 satisfied:3 squared:4 management:1 nm:2 interuniversity:1 derivative:1 dimitri:1 li:5 account:1 de:1 bfgs:1 wk:3 satisfy:1 mv:1 vi:1 depends:1 bilateral:1 h1:1 lot:1 morphologic:4 responsibility:1 sup:11 portion:1 universiteit:2 ass:1 formed:1 minister:1 square:2 ir:2 il:1 yves:1 characteristic:9 percept:2 who:2 iml:1 accuracy:3 wendell:1 accurately:1 trajectory:8 tissue:1 classified:1 submitted:1 overtraining:2 reach:1 joos:1 ed:3 definition:1 mun97:1 acquisition:1 pp:6 invasive:1 dm:1 proof:1 jacobi:2 associated:1 oram:1 proved:1 lim:1 color:3 jacobibellman:1 occured:1 organized:1 schedule:1 improves:1 back:3 campbell:1 appears:2 feed:3 ok:2 higher:1 dt:5 supervised:2 response:2 rand:2 done:1 just:1 stage:2 horizontal:2 replacing:2 ei:14 nonlinear:1 christopher:1 assessment:2 incrementally:1 marker:3 propagation:1 logistic:1 aj:3 gray:1 indicated:1 scientific:2 building:1 usage:1 concept:6 true:4 evolution:1 assigned:1 laboratory:1 assay:1 illustrated:1 deal:1 round:1 adjacent:1 inferior:1 backpropagating:1 harold:1 levenberg:1 oc:2 criterion:2 presenting:1 impression:1 complete:1 polytechnique:1 vo:1 tn:4 motion:3 bradtke:1 temperature:2 fttk:2 laij:1 image:1 meaning:3 fi:1 functional:2 mt:1 rl:10 nh:3 thirdly:1 discussed:1 approximates:2 numerically:3 refer:1 measurement:1 cambridge:1 ai:2 leuven:6 mathematics:1 had:3 dot:1 malignancy:6 operating:7 v0:1 deduce:3 brownian:3 closest:1 recent:1 showed:1 multivariate:1 prime:1 verlag:3 certain:4 binary:4 seen:3 gill:1 mr:1 eo:1 vxixj:3 sc2:1 maximize:1 dashed:1 signal:1 bast:1 ii:10 relates:1 rj:1 jrd:1 afterwards:1 seidel:1 ing:1 smooth:3 af:2 oncologist:1 controlled:2 prediction:11 regression:1 patient:9 expectation:1 iteration:3 c1:1 irregular:1 whereas:2 annealing:9 interval:1 rest:1 probably:1 cedex:2 med:1 db:1 regularly:2 flow:2 integer:2 ee:4 presence:5 counting:1 split:1 viability:1 concerned:1 mips:1 variety:1 fori:1 gave:3 architecture:3 imperfect:1 knowing:1 br:1 tkf:3 synchronous:1 whether:1 motivated:2 expression:1 introd:1 assist:1 septum:1 ird:1 peter:1 puw:1 york:4 hessian:1 remark:2 adequate:1 v8:1 matlab:1 useful:1 action:1 detailed:2 clear:1 viscosity:1 discount:2 fwo:1 extensively:1 situated:1 statespace:1 diameter:1 percentage:1 vy:1 dotted:1 cemagref:2 diagnostic:2 ovarian:13 zd:2 diagnosis:2 vol:6 express:1 kloeden:1 group:1 threshold:8 blood:2 monitor:1 prevent:1 supueuh:2 diffusion:4 backward:1 imaging:3 graph:1 mcneil:1 parameterized:1 tailor:2 almost:3 reader:1 vn:2 lake:1 missed:1 antibody:1 decision:3 vex:5 appendix:2 vf:3 comparable:3 entirely:1 bound:2 fl:1 layer:4 convergent:5 scanned:1 infinity:2 bp:1 ftk:1 i1k:1 generates:1 vecchi:1 min:2 span:5 oup:1 majoration:1 jr:1 xixj:1 soner:1 urgent:1 stic:1 taken:1 equation:10 discus:2 malignant:22 needed:2 demographic:2 available:2 gelatt:1 neighbourhood:2 jd:1 standardized:1 running:1 kushner:2 jiei:1 remaining:1 newton:2 lipschitzian:1 hanley:1 uj:2 murray:1 unchanged:1 surgery:1 objective:6 already:1 strategy:1 concentration:1 said:2 gradient:5 dp:4 ultrasonography:1 simulated:7 vd:1 stocha:1 whom:2 reason:1 besides:8 index:1 relationship:1 xixi:1 ratio:1 vee:1 negative:7 fluid:2 rise:2 design:1 policy:1 unknown:2 boltzmann:1 u6u:1 upper:1 observation:3 neuron:2 markov:1 ultrasound:2 finite:7 descent:1 roccurve:1 hinton:1 dirk:1 rn:1 august:2 learningfor:2 community:1 drift:1 introduced:2 kl:1 doppler:3 fleming:1 esat:1 able:2 beyond:2 krylov:1 dynamical:1 below:2 pattern:1 fp:1 challenge:1 fs93:2 program:1 herman:1 including:1 reliable:4 explanation:1 power:1 examination:1 indicator:12 normality:1 scheme:10 pare:1 technology:1 brief:1 lk:1 started:2 occurence:1 epoch:1 kf:1 relative:4 law:2 fully:1 interesting:1 proven:1 versus:3 age:5 validation:1 sufficient:3 principle:1 vij:1 cd:1 cancer:3 ift:1 course:1 last:1 transpose:1 asynchronous:1 soon:1 free:2 aij:3 side:1 bias:2 perceptron:1 wide:1 taking:2 munos:6 kp95j:3 moreau:4 distributed:1 van:1 boundary:5 calculated:8 dimension:1 overcome:1 transition:2 cumulative:1 kzk:1 curve:23 forward:2 collection:1 reinforcement:12 ignores:1 coincide:1 palaiseau:1 programme:1 author:1 approximate:1 compact:1 status:3 technicality:1 satinder:1 mter:1 receiver:7 consuming:1 iayer:2 continuous:14 un:1 table:8 reality:1 nature:1 zk:14 glycoprotein:1 ca:13 traj:1 mse:10 poly:1 cl:2 protocol:1 whole:1 paul:1 allowed:1 referred:1 malvern:1 representative:1 roc:22 tl:1 dynami:1 wiley:2 precision:1 experienced:2 surveyed:1 ueu:2 sub:1 exponential:1 xl:1 kus90:2 lie:1 iuap:2 watkins:1 third:1 flanders:1 ito:1 learns:1 ix:1 theorem:7 rk:1 bishop:1 list:2 dk:1 exists:6 incorporating:1 uction:1 false:11 adding:1 ci:2 phd:1 te:2 kx:1 crea:1 mf:3 platen:1 depicted:1 remi:2 formly:1 insures:2 simply:1 expressed:1 v6:16 preoperative:5 springer:3 satisfies:2 supueu:1 ma:4 prop:1 towards:2 lipschitz:1 absence:2 considerable:1 formative:1 change:1 included:1 clinician:2 uniformly:2 tumor:18 dwk:3 called:2 eeh:1 hospital:2 conservative:1 gauss:1 total:10 accepted:2 select:1 e6:3 iv6:2 internal:1 mark:1 collins:1 dept:2 ex:6 |
448 | 1,408 | New Approximations of Differential
Entropy for Independent Component
Analysis and Projection Pursuit
Aapo Hyvarinen
Helsinki University of Technology
Laboratory of Computer and Information Science
P.O. Box 2200, FIN-02015 HUT, Finland
Email: aapo.hyvarinen<Ohut.fi
Abstract
We derive a first-order approximation of the density of maximum
entropy for a continuous 1-D random variable, given a number of
simple constraints. This results in a density expansion which is
somewhat similar to the classical polynomial density expansions
by Gram-Charlier and Edgeworth. Using this approximation of
density, an approximation of 1-D differential entropy is derived.
The approximation of entropy is both more exact and more robust against outliers than the classical approximation based on
the polynomial density expansions, without being computationally
more expensive. The approximation has applications, for example,
in independent component analysis and projection pursuit.
1
Introduction
The basic information-theoretic quantity for continuous one-dimensional random
variables is differential entropy. The differential entropy H of a scalar random
variable X with density f(x) is defined as
H(X) = - / f(x) log f(x)dx.
(1)
The 1-D differential entropy, henceforth called simply entropy, has important applications such areas as independent component analysis [2, 10] and projection pursuit
[5, 6]. Indeed, both of these methods can be considered as a search for directions
in which entropy is minimal, for constant variance.
Unfortunately, the estimation of entropy is quite difficult in practice. Using definition (1) requires estimation of the density of X, which is recognized to be both
A Hyviirinen
274
theoretically difficult and computationally demanding. Simpler approximations of
entropy have been proposed both in the context of projection pursuit [9] and independent component analysis [1, 2]. These approximations are usually based on
approximating the density f(x) using the polynomial expansions of Gram-Charlier
or Edgeworth [11]. This construction leads to the use of higher-order cumulants,
like kurtosis. However, such cumulant-based methods often provide a rather poor
approximation of entropy. There are two main reasons for this. Firstly, finitesample estimators of higher-order cumulants are highly sensitive to outliers: their
values may depend on only a few, possibly erroneous, observations with large values
[6]. This means that outliers may completely determine the estimates of cumulants,
thus making them useless. Secondly, even if the cumulants were estimated perfectly, they measure mainly the tails of the distribution, and are largely unaffected
by structure near the centre of the distribution [5].
Therefore, better approximations of entropy are needed. To this end, we introduce
in this paper approximations of entropy that are both more exact in the expectation
and have better finite-sample statistical properties, when compared to the cumulantbased approximations. Nevertheless, they retain the computational and conceptual
simplicity of the cumulant-based approach. Our approximations are based on an
approximative maximum entropy method. This means that we approximate the
maximum entropy that is compatible with our measurements of the random variable
X. This maximum entropy, or further approximations thereof, can then be used as
a meaningful approximation of the entropy of X. To accomplish this, we derive a
first-order approximation of the density that has the maximum entropy given a set
of constraints, and then use it to derive approximations of the differential entropy
of X.
2
Applications of Differential Entropy
First, we discuss some applications of the approximations introduced in this paper. Two important applications of differential entropy are independent component analysis (ICA) and projection pursuit. 'In the general formulation of ICA
[2], the purpose is to transform an observed random vector x = (Xl, ... , Xm)T linearly into a random vector s = (81, ... , 8m )T whose components are statistically
as independent from each other as possible. The mutual dependence of the 8i
is classically measured by mutual information. Assuming that the linear transformation is invertible, the mutual information 1(81, ... , 8 m ) can be expressed as
1(81, ... , 8m ) = 2:i H(8i) - H(Xl, ... , Xm) -log Idet MI where M is the matrix defining the transformation s = Mx. The second term on the right-hand side does not
depend on M, and the minimization of the last term is a simple matter of differential
calculus. Therefore, the critical part is the estimation of the 1-D entropies H(8i}:
finding an efficient and reliable estimator or approximation of entropy enables an
efficient and reliable estimation of the ICA decomposition .
In projection pursuit, the purpose is to search for projections of multivariate data
which have 'interesting' distributions [5, 6, 9]. Typically, interestingness is considered equivalent with non-Gaussianity. A natural criterion of non-Gaussianity is
entropy [6, 9], which attains its maximum (for constant variance) when the distribution is Gaussian, and all other distributions have smaller entropies. Because of the
difficulties encountered in the estimation of entropy, many authors have considered
other measures of non-Gaussianity (see [3]) but entropy remains, in our view, the
best choice of a projection pursuit index, especially because it provides a simple
connection to ICA. Indeed, it can be shown [2] that in ICA as well as in projection
pursuit, the basic problem is to find directions in which entropy is minimized for
New Approximations of Differential Entropy
275
constant variance.
3
Why maximum entropy?
Assume that the information available on the density f(x) of the scalar random
variable X is of the form
J
f(x)Gi(x)dx = Ci, for i = 1, ... , n,
(2)
which means in practice that we have estimated the expectations E{Gi(X)} of n
different functions of X. Since we are not assuming any model for the random
variable X, the estimation of the entropy of X using this information is not a
well-defined problem: there exist an infinite number of distributions for which the
constraints in (2) are fulfilled, but whose entropies are very different from each
other. In particular, the differential entropy reaches -00 in the limit where X takes
only a finite number of values.
A simple solution to this dilemma is the maximum entropy method. This means
that we compute the maximum entropy that is compatible with our constraints
or measurements in (2), which is a well-defined problem. This maximum entropy,
or further approximations thereof, can then be used as an approximation of the
entropy of X.
Our approach thus is very different from the asymptotic approach often used in
projection pursuit [3, 5]. In the asymptotic approach, one establishes a sequence of
functions G i so that when n goes to infinity, the information in (2) gives an asymptotically convergent approximation of some theoretical projection pursuit index. We
avoid in this paper any asymptotic considerations, and consider directly the case of
finite information, i.e., finite n. This non-asymptotic approach is justified by the
fact that often in practice, only a small number of measurements of the form (2)
are used, for computational or other reasons.
4
Approximating the maximum entropy density
In this section, we shall derive an approximation of the density of maximum entropy
compatible with the measurements in (2). The basic results of the maximum entropy
method tell us [4] that under some regularity conditions, the density fo(x) which
satisfies the constraints (2) and has maximum entropy among all such densities, is
of the form
(3)
where A and ai are constants that are determined from the Ci, using the constraints
in (2) (i.e., by substituting the right-hand side of (3) for fin (2)), and the constraint
J fo(x)dx = 1. This leads in general to a system of n+ 1 non-linear equations which
is difficult to solve. Therefore, we decide to make a simple approximation of fo. This
is based on the assumption that the density f(x) is not very far from a Gaussian
distribution of the same mean and variance. Such an assumption, though perhaps
counterintuitive, is justified because we shall construct a density expansion (not
unlike a Taylor expansion) in the vicinity of the Gaussian density. In addition, we
can make the technical assumption that f(x) is near the standardized Gaussian
density 'P(x) = exp(-x 2 /2)/..;2ii, since this amounts simply to making X zeromean and of unit variance. Therefore we put two additional constraints in (2),
defined by G n+1 (x) = x, Cn+l = 0 and G n+2(x) = x 2, Cn+2 = 1. To further simplify
A. HyvlJrinen
276
the calculations, let us make another, purely technical assumption: The functions
G i , i = 1, ... , n, form an orthonormal system according to the metric defined by
cp, and are orthogonal to all polynomials of second degree. In other words, for all
i,j=1, ... ,n
!
cp(x)Gi(x)Gj(x)dx =
{
1, if i = j
0, if i i- j
(4)
,
For any linearly independent functions G i , this assumption can always be made
true by ordinary Gram-Schmidt orthonormalization.
Now, note that the assumption of near-Gaussianity implies that all the other ai in
(3) are very small compared to a n +2 ~ -1/2, since the exponential in (3) is not far
from exp( _x 2 /2). Thus we can make a first-order approximation of the exponential
function (detailed derivations can be found in [8]). This allows for simple solutions
for the constants in (3), and we obtain the approximative maximum entropy density,
which we denote by j(x):
n
j(x) = cp(x)(1
+ L CiGi(X))
(5)
i=l
where Ci = E{ G i (X)}. To estimate this density in practice, the Ci are estimated, for
example, as the corresponding sample averages of the Gi(X). The density expansion
in (5) is somewhat similar to the Gram-Charlier and Edgeworth expansions [11].
5
Approximating the differential entropy
An important application of the approximation of density shown in (5) is in approximation of entropy. A simple approximation of entropy can be found by
approximating both occurences of f in the definition (1) by j as defined in
Eq. (5), and using a Taylor approximation of the logarithmic function, which yields
(1 + ?) log(1 + ?) ~ ? + ?2/2. Thus one obtains after some algebraic manipulations
[8]
H(X)
~-
!
j(x) log j(x)dx
~ H(v) - ~
t
(6)
c;
i=l
where H(v) = ~(1 +log(27r)) means the entropy of a standardized Gaussian variable,
and Ci = E{ Gi(X)} as above. Note that even in cases where this approximation
is not very accurate, (6) can be used to construct a projection pursuit index (or
a measure of non-Gaussianity) that is consistent in the sense that (6) obtains its
maximum value, H(v), when X has a Gaussian distribution.
6
Choosing the measuring functions
Now it remains to choose the 'measuring' functions G i that define the information
given in (2). As noted in Section 4, one can take practically any set of linearly independent functions, say Ch i = 1, ... , n, and then apply Gram-Schmidt orthonormalization on the set containing those functions and the monomials xk, k 0,1,2, so as
to obtain the set G i that fulfills the orthogonality assumptions in (4). This can be
done, in general, by numerical integration. In the practical choice of the functions
Gi , the following criteria must be emphasized: First, the practical estimation of
E{Gi(x)} should not be statistically difficult. In particular, this estimation should
not be too sensitive to outliers. Second, the maximum entropy method assumes
=
277
New Approximations of Differential Entropy
that the function fo in (3) is integrable. Therefore, to ensure that the maximum
entropy distribution exists in the first place, the Gi(x) must not grow faster than
quadratically as a function of Ixl, because a function growing faster might lead to
non-integrability of fo [4]. Finally, the Gi must capture aspects of the distribution
of X that are pertinent in the computation of entropy. In particular, if the density
f(x) were known, the optimal function GoPt would clearly be -logf(x), because
-E{log f(X)} gives directly the entropy. Thus, one might use the log-densities of
some known important densities as Gi .
The first two criteria are met if the Gi(x) are functions that do not grow too fast
(not faster than quadratically) when Ixl grows. This excludes, for example, the
use of higher-order polynomials, as are used in the Gram-Charlier and Edgeworth
expansions. One might then search, according to the last criterion above, for logdensities of some well-known distributions that also fulfill the first two conditions.
Examples will be given in the next section. It should be noted, however, that the
criteria above only delimit the space of function that can be used. Our framework
enables the use of very different functions (or just one) as Ch The choice is not
restricted to some well-known basis of a functional space, as in most approaches
[1,2,9]. However, if prior knowledge is available on the distributions whose entropy
is to estimated, the above consideration shows how to choose the optimal function.
7
A simple special case
A simple special case of (5) is obtained if one uses two functions G1 and G2 , which
are chosen so that G1 is odd and (h is even. Such a system of two functions
can measure the two most important features of non-Gaussian 1-D distributions.
The odd function measures the asymmetry, and the even function measures the
bimodality /sparsity dimension (called central hole/central mass concentration in
[3)). After extensive experiments, Cook et al [3] also came to the conclusion that
two such measures (or two terms in their projection pursuit index) are enough for
projection pursuit in most cases. Classically, these features have been measured by
skewness and kurtosis, which correspond to G 1 (x) = x 3 and G2 (x) = X4, but we do
not use these functions for the reasons explained in Section 6.
In this special case, the approximation in (6) simplifies to
where k1 and k2 are positive constants (see [8]), and v is a Gaussian random variable of zero mean and unit variance. Practical examples of choices of Gi that are
consistent with the requirements in Section 6 are the following.
First, for measuring bimodality /sparsity, one might use, according to the recommendations of Section 6, the log-density of the double exponential (or Laplace) distribution: G2a (x) = Ixl. For computational reasons, a smoother version of G2a might also
be used. Another choice would be the Gaussian function, which may be considered
as the log-density of a distribution with infinitely heavy tails: G2b (X) = exp( -x 2/2).
For measuring asymmetry, one might use, on more heuristic grounds, the following function: G1 (x) = x exp (- x 2 /2). which corresponds to the second term in the
projection pursuit index proposed in [3].
Using the above examples one obtains two practical examples of (7):
Ha(X) = H(v) - [k1 (E{X exp( _X2 /2)})2
Hb(X) = H(v) - [k1 (E{X exp( _X2 /2)})2
+ k2(E{IXI} - }2/71")2],
+ k~(E{exp( _X2 /2)}
(8)
- Ji72)2], (9)
A. Hyvllrinen
278
with kl = 36/(8V3 - 9) , k~ = 1/(2 - 6/'rr), and k~ = 24/(16V3 - 27). As above,
H(v) =
+ log(27r)) means the entropy of a standardized Gaussian variable.
These approximations Ha(X) and Hb(X) can be considered more robust and accurate generalizations of the approximation derLved using the Gr~m-Charlier expansion in [9] . Indeed, using the polynomials G 1 (x) = x 3 and G 2 (x) = x4 one
obtains the approximation of entropy in (9), which is in practice almost identical
to those proposed in [1, 2]. Finally, note that the approximation in (9) is very
similar to the first two terms of the projection pursuit index in [3] . Algorithms for
independent component analysis and projection pursuit can be derived from these
approximations, see [7].
!(1
8
Simulation results
To show the validity of our approximations of differential entropy we compared the
approximations Ha and Hb in Eqs (8) and (9) in Section 7, with the one offered
by higher-order cumulants as given in [9]. The expectations were here evaluated
exactly, ignoring finite-sample effects.
First, we used a family of Gaussian mixture densities, defined by
f(x) = J.tcp(x)
+ (1 -
J.t)2cp(2(x - 1))
(10)
where J.t is a parameter that takes all the values in the interval 0 ::; J.t ::; 1. This family includes asymmetric densities of both negative and positive kurtosis. The results
are depicted in Fig. 1. Note that the plots show approximations of negentropies:
the negentropy of X equals H(v) -H(X), where v is again a standardized Gaussian
variable. One can see that both of the approximations Ha and Hb introduced in
Section 7 were considerably more accurate than the cumulant-based approximation.
Second, we considered the following family of density functions :
(11)
where a is a positive constant : and C 1 , C2 are normalization constants that make f Ot
a probability density of unit variance. For different values of a , the densities in this
family exhibit different shapes. For a < 2, one obtains (sparse) densities of positive
kurtosis. For a = 2, one obtains the Gaussian density, and for a > 2, a density of
negative kurtosis. Thus the densities in this family can be used as examples of different symmetric non-Gaussian densities. In Figure 2, the different approximations
are plotted for this family, using parameter values .5 ::; a ~ 3. Since the densities used are all symmetric, the first terms in the approximations were neglected.
Again, it is clear that both of the approximations Ha and Hb introduced in Section 7
were much more accurate than the cumulant-based approximation in [2, 9]. (In the
case of symmetric densities, these two cumulant-based approximations are identical). Especially in the case of sparse densities (or densities of positive kurtosis), the
cumulant-based approximations performed very poorly; this is probably because it
gives too much weight to the tails of the distribution.
References
[1] S. Amari, A. Cichocki, and H.H. Yang. A new learning algorithm for blind source separation. In D . S. Touretzky, M. C. Mozer, and M. E . Hasselmo, editors, Advances in
Neural Information Processing 8 (Proc. NIPS '95), pages 757- 763. MIT Press, Cambridge, MA, 1996.
12] P. Comon. Independent component analysis - a new concept?
36:287- 314, 1994.
Signal Processing,
New Approximationr of Differential Entropy
0025
... :-: ""
- .... "',
"-
...
",
0 01 $
,
\'
"
"
o.
0 6
279
Figure 1: Comparison of different approximations of negentropy, for the family
of mixture densities in (10) parametrized
by JL ranging from 0 to 1. Solid curve:
true negentropy. Dotted curve: cumulantbased approximation. Dashed curve: approximation Ha in (8). Dot-dashed Cu:Lve:
approximation Hb in (9). Our two approximations were clearly better than the
cumulant-based one.
,
.
'.
"
, '
,
-. , '. ,
. .'
, '
"':"
,'......
--
~~ .
""-
....
"
.
-:- ~ ,~--~-.---
Figure 2: Comparison of different approximations of negentropy, for the family of densities
(11) parametrized by Q. On the left, approximations for densities of positive kurtosis (.5 ~
Q < 2) are depicted, and on the right, approximations for densities of negative kurtosis
(2 < Q ~ 3). Solid curve: true negentropy. Dotted curve: cumulant-based approximation.
Dashed curve: approximation Ha in (8). Dot-dashed curve: approximation Hb in (9) .
Clearly, our two approximations were much better than the cumulant-based one, especially
in the case of densities of positive kurtosis.
[3) D. Cook, A. Buja, and J. Cabrera. Projection pursuit indexes based on orthonormal
function expansions. J. of Computational and Graphical Statistics, 2(3) :225-250, 1993.
[4) T. M. Cover and J. A. Thomas. Elements of Information Theory. John Wiley &
Sons, 1991.
[5) J.H. Friedman. Exploratory projection pursuit. J. of the American Statistical Association, 82(397):249-266, 1987.
[6) P.J. Huber. Projection pursuit. The Annals of Statistics, 13(2):435-475, 1985.
[7) A. Hyvarinen. Independent component analysis by minimization of mutual information. Technical Report A46, Helsinki University of Technology, Laboratory of
Computer and Information Science, 1997.
[8) A. Hyviirinen. New approximations of differential entropy for independent component analysis and projection pursuit. Technical Report A47, Helsinki University of
Technology, Laboratory of Computer and Information Science, 1997. Available at
http://www.cis.hut.fi;-aapo.
[9) M.C. Jones and R. Sibson. What is projection pursuit ? J. of the Royal Statistical
Society, ser. A , 150:1-36, 1987.
[10) C. Jutten and J. Herault. Blind separation of sources, part I: An adaptive algorithm
based on neuromimetic architecture. Signal Processing, 24:1-10, 1991.
[11) M. Kendall and A. Stuart. The Advanced Theory of Statistics. Charles Griffin &
Company, 1958.
| 1408 |@word cu:1 version:1 polynomial:6 calculus:1 simulation:1 decomposition:1 solid:2 negentropy:5 dx:5 must:3 john:1 numerical:1 shape:1 enables:2 pertinent:1 plot:1 cook:2 xk:1 tcp:1 provides:1 firstly:1 simpler:1 c2:1 differential:16 introduce:1 theoretically:1 huber:1 ica:5 indeed:3 growing:1 company:1 mass:1 what:1 skewness:1 finding:1 transformation:2 exactly:1 k2:2 ser:1 unit:3 interestingness:1 positive:7 limit:1 might:6 statistically:2 practical:4 practice:5 edgeworth:4 orthonormalization:2 area:1 projection:22 word:1 put:1 context:1 www:1 equivalent:1 go:1 delimit:1 simplicity:1 occurences:1 estimator:2 counterintuitive:1 orthonormal:2 exploratory:1 laplace:1 annals:1 construction:1 exact:2 approximative:2 us:1 ixi:1 element:1 expensive:1 asymmetric:1 observed:1 capture:1 mozer:1 neglected:1 depend:2 dilemma:1 purely:1 completely:1 basis:1 derivation:1 fast:1 tell:1 choosing:1 quite:1 whose:3 heuristic:1 solve:1 say:1 amari:1 statistic:3 gi:12 g1:3 transform:1 sequence:1 rr:1 kurtosis:9 poorly:1 regularity:1 asymmetry:2 requirement:1 double:1 bimodality:2 derive:4 measured:2 odd:2 eq:2 implies:1 met:1 direction:2 generalization:1 secondly:1 practically:1 hut:2 considered:6 ground:1 exp:7 substituting:1 finland:1 purpose:2 estimation:8 proc:1 sensitive:2 hasselmo:1 establishes:1 minimization:2 mit:1 clearly:3 gaussian:14 always:1 rather:1 fulfill:1 avoid:1 derived:2 integrability:1 mainly:1 attains:1 cabrera:1 sense:1 typically:1 among:1 herault:1 integration:1 special:3 mutual:4 equal:1 construct:2 x4:2 identical:2 stuart:1 jones:1 minimized:1 report:2 simplify:1 few:1 friedman:1 highly:1 mixture:2 accurate:4 orthogonal:1 taylor:2 plotted:1 theoretical:1 minimal:1 cover:1 cumulants:5 measuring:4 ordinary:1 monomials:1 gr:1 too:3 accomplish:1 considerably:1 density:46 retain:1 invertible:1 again:2 central:2 containing:1 choose:2 possibly:1 classically:2 henceforth:1 american:1 gaussianity:5 includes:1 matter:1 blind:2 performed:1 view:1 kendall:1 variance:7 largely:1 yield:1 correspond:1 unaffected:1 gopt:1 reach:1 fo:5 touretzky:1 email:1 definition:2 against:1 thereof:2 mi:1 knowledge:1 higher:4 formulation:1 done:1 box:1 though:1 zeromean:1 evaluated:1 just:1 hand:2 jutten:1 perhaps:1 grows:1 effect:1 validity:1 concept:1 true:3 vicinity:1 symmetric:3 laboratory:3 a46:1 noted:2 criterion:5 theoretic:1 cp:4 ranging:1 consideration:2 fi:2 charles:1 functional:1 jl:1 tail:3 association:1 measurement:4 cambridge:1 ai:2 centre:1 dot:2 gj:1 multivariate:1 manipulation:1 came:1 integrable:1 additional:1 somewhat:2 recognized:1 determine:1 v3:2 signal:2 ii:1 smoother:1 dashed:4 technical:4 faster:3 calculation:1 aapo:3 basic:3 expectation:3 metric:1 normalization:1 lve:1 justified:2 addition:1 interval:1 grow:2 source:2 ot:1 unlike:1 probably:1 near:3 logf:1 yang:1 enough:1 hb:7 charlier:5 architecture:1 perfectly:1 simplifies:1 cn:2 algebraic:1 detailed:1 clear:1 amount:1 http:1 exist:1 dotted:2 estimated:4 fulfilled:1 shall:2 sibson:1 nevertheless:1 asymptotically:1 excludes:1 place:1 almost:1 family:8 decide:1 separation:2 griffin:1 convergent:1 encountered:1 orthogonality:1 constraint:8 infinity:1 helsinki:3 x2:3 aspect:1 according:3 poor:1 smaller:1 son:1 making:2 comon:1 outlier:4 restricted:1 explained:1 computationally:2 equation:1 remains:2 discus:1 needed:1 idet:1 end:1 neuromimetic:1 pursuit:21 available:3 apply:1 schmidt:2 thomas:1 standardized:4 assumes:1 ensure:1 graphical:1 k1:3 especially:3 approximating:4 classical:2 society:1 quantity:1 concentration:1 dependence:1 exhibit:1 mx:1 parametrized:2 reason:4 assuming:2 useless:1 index:7 difficult:4 unfortunately:1 cumulantbased:2 negative:3 observation:1 fin:2 finite:5 defining:1 buja:1 introduced:3 kl:1 extensive:1 connection:1 quadratically:2 nip:1 usually:1 xm:2 sparsity:2 reliable:2 royal:1 critical:1 demanding:1 natural:1 difficulty:1 advanced:1 technology:3 cichocki:1 prior:1 finitesample:1 asymptotic:4 interesting:1 ixl:3 degree:1 offered:1 consistent:2 editor:1 heavy:1 compatible:3 last:2 side:2 sparse:2 curve:7 dimension:1 gram:6 author:1 made:1 adaptive:1 far:2 hyvarinen:3 approximate:1 obtains:6 conceptual:1 continuous:2 search:3 why:1 hyviirinen:2 robust:2 ignoring:1 expansion:11 main:1 linearly:3 fig:1 wiley:1 exponential:3 xl:2 erroneous:1 emphasized:1 exists:1 ci:6 hole:1 entropy:58 depicted:2 logarithmic:1 simply:2 infinitely:1 expressed:1 g2:2 scalar:2 recommendation:1 ch:2 corresponds:1 satisfies:1 ma:1 infinite:1 determined:1 called:2 meaningful:1 fulfills:1 cumulant:9 |
449 | 1,409 | Generalized Prioritized Sweeping
David Andre
Nir Friedman
Ronald Parr
Computer Science Division, 387 Soda Hall
University of California, Berkeley, CA 94720
{dandre,nir,parr}@cs.berkeley.edu
Abstract
Prioritized sweeping is a model-based reinforcement learning method
that attempts to focus an agent's limited computational resources to
achieve a good estimate of the value of environment states. To choose effectively where to spend a costly planning step, classic prioritized sweeping uses a simple heuristic to focus computation on the states that are
likely to have the largest errors. In this paper, we introduce generalized
prioritized sweeping, a principled method for generating such estimates
in a representation-specific manner. This allows us to extend prioritized
sweeping beyond an explicit, state-based representation to deal with compact representations that are necessary for dealing with large state spaces.
We apply this method for generalized model approximators (such as
Bayesian networks), and describe preliminary experiments that compare
our approach with classical prioritized sweeping.
1 Introduction
In reinforcement learning, there is a tradeoff between spending time acting in the environment and spending time planning what actions are best. Model-free methods take one
extreme on this question- the agent updates only the state most recently visited. On the
other end of the spectrum lie classical dynamic programming methods that reevaluate the
utility of every state in the environment after every experiment. Prioritized sweeping (PS)
[6] provides a middle ground in that only the most "important" states are updated, according
to a priority metric that attempts to measure the anticipated size of the update for each state.
Roughly speaking, PS interleaves perfonning actions in the environment with propagating
the values of states. After updating the value of state s, PS examines all states t from which
the agent might reach s in one step and assigns them priority based on the expected size of
the change in their value.
A crucial desideratum for reinforcement learning is the ability to scale-up to complex
domains. For this, we need to use compact (or generalizing) representations of the model and
the value function. While it is possible to apply PS in the presence of such representations
(e.g., see [1]), we claim that classic PS is ill-suited in this case. With a generalizing model,
a single experience may affect our estimation of the dynamics of many other states. Thus,
we might want to update the value of states that are similar, in some appropriate sense, to
s since we have a new estimate of the system dynamics at these states. Note that some of
these states might never have been reached before and standard PS will not assign them a
priority at all.
D. Andre, N. Friedman and R Parr
1002
In this paper, we present generalized prioritized sweeping (GenPS), a method that utilizes
a fonnal principle to understand and extend PS and extend it to deal with parametric
representations for both the model and the value function. If GenPS is used with an explicit
state-space model and value function representation, an algorithm similar to the original
(classic) PS results. When a model approximator (such as a dynamic Bayesian network
[2]) is used, the resulting algorithm prioritizes the states of the environment using the
generalizations inherent in the model representation.
2 The Basic Principle
We assume the reader is familiar with the basic concepts of Markov Decision Processes
(MDPs); see, for example, [5]. We use the following notation: A MDP is a 4-tuple,
(S,A,p,r) where S is a set of states, A is a set of actions, p(t I s,a) is a transition
model that captures the probability of reaching state t after we execute action a at state
s, and r( s) is a reward function mapping S into real-valued rewards. In this paper, we
focus on infinite-horizon MDPs with a discount factor ,. The agent's aim is to maximize
the expected discounted total reward it will receive. Reinforcement learning procedures
attempt to achieve this objective when the agent does not know p and r.
A standard problem in model-based reinforcement learning is one of balancing between
planning (Le., choosing a policy) and execution. Ideally, the agent would compute the
optimal value function for its model of the environment each time the model changes. This
scheme is unrealistic since finding the optimal policy for a given model is computationally
non-trivial. Fortunately, we can approximate this scheme if we notice that the approximate
model changes only slightly at each step. Thus, we can assume that the value function
from the previous model can be easily "repaired" to reflect these changes. This approach
was pursued in the DYNA [7] framework, where after the execution of an action, the
agent updates its model of the environment, and then performs some bounded number
of value propagation steps to update its approximation of the value function . Each vaiuepropagation step locally enforces the Bellman equation by setting V(s) ~ maxaEA Q(s, a),
where Q(s,a) = f(s) + ,Ls'ESP(s' I s,a)V(s'), p(s' I s,a) and f(s) are the agent's
approximation of the MDP, and V is the agent's approximation of the value function.
This raises the question of which states should be updated. In this paper we propose the
following general principle:
GenPS Principle: Update states where the approximation of the value
function will change the most. That is, update the states with the largest
Bellmanerror,E(s)
IV(s) -maxaEAQ(s,a)l.
The motivation for this principle is straightforward. The maximum Bellman error can be
used to bound the maximum difference between the current value function, V(s) and the
optimal value function, V"'(s) [9]. This difference bounds the policy loss, the difference
between the expected discounted reward received under the agent's current policy and the
expected discounted reward received under the optimal policy.
To carry out this principle we have to recognize when the Bellman error at a state changes.
This can happen at two different stages. First, after the agent updates its model of the world,
new discrepancies between V (s) and max a Q( s, a) might be introduced, which can increase
the Bellman error at s. Second, after the agent performs some value propagations, V is
changed, which may introduce new discrepancies.
We assume that the agent maintains a value function and a model that are parameterized
by Dv and DM . (We will sometimes refer to the vector that concatenates these vectors
together into a single, larger vector simply as D.) When the agent observes a transition from
state s to s' under action a, the agent updates its environment model by adjusting some
of the parameters in DM. When perfonning value-propagations, the agent updates V by
updating parameters in Dv. A change in any of these parameters may change the Bellman
error at other states in the model. We want to recognize these states without explicitly
=
Generalized Prioritized Sweeping
1003
computing the Bellman error at each one. Formally, we wish to estimate the change in
error, I~E(B) I, due to the most recent change ~() in the parameters.
We propose approximating I~E(8) 1 by using the gradient of the right hand side of the
Bellman equation (i.e. max a Q(8,a). Thus, we have: I~E(s)1 ~ lV'max a Q(8,a) . ~()I
which estimates the change in the Bellman error at state 8 as a function of the change in
Q( 8, a). The above still requires us to differentiate over a max, which is not differentiable.
In general, we want to to overestimate the change, to avoid "starving" states with nonnegligible error. Thus, we use the following upper bound: 1V'(max a Q(8, a)) . ~81 ~
max a IV'Q(s,a). ~81?
We now define the generalized prioritized sweeping procedure. The procedure maintains
a priority queue that assigns to each state 8 a priority,pri( 8). After making some changes, we
can reassign priorities by computing an approximation of the change in the value function.
Ideally, this is done using a procedure that implements the following steps:
procedure update-priorities (&)
for all s E S pri(s) +- pri(s) + max a IV'Q(s, a) . &1.
Note that when the above procedure updates the priority for a state that has an existing
priority, the priorities are added together. This ensures that the priority being kept is an
overestimate of the priority of each state, and thus, the procedure will eventually visit all
states that require updating.
Also, in practice we would not want to reconsider the priority of all states after an update
(we return to this issue below).
Using this procedure, we can now state the general learning procedure:
procedure GenPS 0
loop
perform an action in the environment
update the model; let & be the change in ()
call update-priorities( &)
while there is available computation time
arg max s pri( s)
let smax
perform value-propagation for V(smax); let & be the change in ()
call update-priorities( &)
pri(smax) +- W(smax) - max a Q(smax,a)11
=
Note that the GenPS procedure does not determine how actions are selected. This issue,
which involves the problem of exploration, is orthogonal to the our main topic. Standard
approache!;, such as those described in [5, 6, 7], can be used with our procedure.
This abstract description specifies neither how to update the model, nor how to update the
value function in the value-propagation steps. Both of these depend on the choices made
in the corresponding representation of the model and the value function. Moreover, it is
clear that in problems that involve a large state space, we cannot afford to recompute the
priority of every state in update-priorities. However, we can simplify this computation
by exploiting sparseness in the model and in the worst case we may resort to approximate
methods for finding the states that receive high priority after each change.
3 Explicit, State-based Representation
In this section we briefly describe the instantiation of the generalized procedure when the
rewards, values, and transition probabilities are explicitly modeled using lookup tables. In
this representation, for each state 8, we store the expected reward at 8, denoted by Of(s)' the
estimated value at 8, denoted by Bv (s), and for each action a and state t the number of times
the execution of a at 8 lead to state t, denoted NS,Q.,t. From these transition counts we can
I In general, this will assign the state a new priority of 0, unless there is a self loop. In this case it
will easy to compute the new Bellman error as a by-product of the value propagation step.
D. Andre, N. Friedman and R. Parr
1004
reconstruct the transition probabilities pACt
I 8, a) =
N?. a.! +N~! a .t
O
W t' N ",a ,t' +N" , a , t l
"
'
where NO
8,a,t
are
fictional counts that capture our prior information about the system's dynamics. 2 After each
step in the worJd, these reward and probability parameters are updated in the straightforward
manner. Value propagation steps in this representation set 8Y (t) to the right hand side of
the Bellman equation.
To apply the GenPS procedure we need to derive the gradient of the Bellman equation
for two situations: (a) after a single step in the environment, and (b) after a value update.
In case (a), the model changes after performing action 8~t . In this case, it is easy to
verify that V'Q(s,a) 'do = dOr(t) +
N.}t+N~.a.t (V(t) - 2:~p(t' I s,a)V(t')), and
tt
that V' Q( s', a') . do = 0 if s' =1= s or a' =1= a. Thus, s is the only state whose priority
changes.
In case (b), the value function changes after updating the value of a state t. In this case,
V'Q(s, a) ?do = ,pet I s, a)l1o v (t)' It is easy to see that this is nonzero only ift is reachable
from 8. In both cases, it is straightforward to locate the states where the Bellman error
might have have changed, and the computation of the new priority is more efficient than
computing the Bellman-error. 3
Now we can relate GenPS to standard prioritized sweeping. The PS procedure has the
general form of this application of GenPS with three minor differences. First, after performing a transition s~t in the environment, PS immediately performs a value propagation
for state s, while GenPS increments the priority of s. Second, after performing a value
propagation for state t, PS updates the priority of states s that can reach t with the value
max a p( tis, a) ./1y( t). The priority assigned by GenPS is the same quantity multiplied by
,. Since PS does not introduce priorities after model changes, this multiplicative constant
does not change the order of states in the queue. Thirdly, GenPS uses addition to combine
the old priority of a state with a new one, which ensures that the priority is indeed an upper
bound. In contrast, PS uses max to combine priorities.
This discussion shows that PS can be thought of as a special case of GenPS when the
agent uses an explicit, state-based representation. As we show in the next section, when
the agent uses more compact representations, we get procedures where the prioritization
strategy is quite different from that used in PS. Thus, we claim that classic PS is desirable
primarily when explicit representations are used.
4 Factored Representation
We now examine a compact representation of p( s' I s, a) that is based on dynamic Bayesian
networks (DBNs) [2]. DBNs have been combined with reinforcement learning before in
[8], where they were used primarily as a means getting better generalization while learning.
We will show that they also can be used with prioritized sweeping to focus the agent's
attention on groups of states that are affected as the agent refines its environment model.
We start by assuming that the environment state is described by a set of random variables,
XI, . .. , X n ? For now, we assume that each variable can take values from a finite set
Val(Xi). An assignment of values XI, .? ? , Xn to these variables describes a particular
environment state. Similarly, we assume that the agent's action is described by random
variables AI, ... ,A k . To model the system dynamics, we have to represent the probability
of transitions s~t, where sand t are two assignments to XI, .. . , Xn and a is an assignment
to AI, ... ,A k ? To simplify the discussion, we denote by Yi, .. . , Yn the agent's state after
2Formally, we are using multinomial Dirichlet priors. See, for example, [4] for an introduction to
these Bayesian methods.
3 Although ~~(s .a ) involves a summation over all states, it can be computed efficiently. To see
.. ,a ,t
this, note that the summation is essentially the old value of Q( s, a) (minus the immediate reward)
which can be retained in memory.
Generalized Prioritized Sweeping
1005
the action is executed (e.g., the state t). Thus, p(t I s,a) is represented as a conditional
probability P(YI , ... , Y n I XI, ... ,Xn , AI , ... ,Ak).
A DBN model for such a conditional distribution consists of two components. The
first is a directed acyclic graph where each vertex is labeled by a random variable and in
which the vertices labeled XI, . .. ,Xn and AI, .. . , Ak are roots. This graph speoifies the
factorization of the conditional distribution:
n
P(Yi, ... , Y n I XI'? .. ' X n , AI,???, Ak) =
II P(Yi I Pai),
(1)
i=I
where Pai are the parents of Yi in the graph. The second component of the DBN model is
a description of the conditional probabilities P(Yi I Pai). Together, these two components
describe a unique conditional distribution. The simplest representation of P(Yi I Pai) is a
table that contains a parameter (}i ,y,z = P(Yi = y I Pai = z) for each possible combination
of y E Val(Yi) and z E Val(Pai) (note that z is a joint assignment to several random
variables). It is easy to see that the "density" of the DBN graph determines the number of
parameters needed. In particular, a complete graph, to which we cannot add an arc without
violating the constraints, is equivalent to a state-based representation in terms of the number
of parameters needed. On the other hand, a sparse graph requires few parameters.
In this paper, we assume that the learneris supplied with the DBN structure and only has to
learn the conditional probability entries. It is often easy to assess structure information from
experts even when precise probabilities are not available. As in the state-based representation, we learn the parameters using Dirichlet priors for each multinomial distribution [4].
In this method, we assess the conditional probability (}i,y,z using prior knowledge and the
frequency of transitions observed in the past where Yi = y among those transitions where
Pai = z. Learning amounts to keeping counts Ni ,y,z that record the number of transitions
where Yi = y and Pai = z for each variable Yi and values y E Val(Yi) and z E Val(Pai).
Our prior knowledge is represented by fictional counts Np,y,z. Then we estimate probabil. . USIng
. the 10rmu
~
Ia Ui,y,z
LI
, z+N?,y , z
h
N?~, ' ,z ?
ltles
- N; ,1I N;,.
,z
' W ere
- "
L....,y' N t,y',z
+ NOi,y' ,z?
We now identify which states should be reconsidered after we update the DBN parameters.
Recall that this requires estimating the term V Q( s, a) ./1n. Since!!:.o is sparse, after making
the transition s* ~t*, we have that VQ(s, a) . !!:.O =
2:i a~Q(s:a). ' where yi and zi are the
I,V i '%i
assignments to Yi and Pai, respectively, in s* ~t*. (Recall that s*, a* and t* jointly assign
values to all the variables in the DBN.)
We say that a transition s~t is consistent with an assignment X = x for a vector of
random variables X, denoted (s,a, t) F= (X = x), if X is assigned the value x in s~t.
We also need a similar notion for a partial description of a transition. We say that s and
a are consistent with X = x, denoted (s,a,?) F= (X = x), if there is a t such that
(s, a, t) F= (X = x).
Using this notation, we can show that if (s, a, .) F (Pai = zi), then
8Q(s, a)
=
-: Nt'~-''-':- [8,.?:.,: t(,.J~.: .': p(t I s, a)1l(t) - t(,. ~I=': p(t I s, a)1l(t)]
and if s, a are inconsistent with Pai =
zi, then aa:.(8:a).
""i
=
o.
'Z i
This expression shows that if s is similar to s* in that both agree on the values they assign
to the parents of some Yi (i.e., (s, a*) is consistent with zi), then the priority of s would
change after we update the model. The magnitude of the priority change depends upon both
the similarity of sand s* (i.e. how many of the terms in VQ(s, a) ? !!:'o will be non-zero),
and the value of the states that can be reached from s.
D. Andre, N. Friedman and R. Parr
1006
PSPS+fact<X'ed --- .-
G:nps ??..:.~.._~_..::.:.;. ::::.:....~..:.....:._--I---...?---
2.S
?? oX-'--
~
'3
"
0
~
I.S
?
005
./
,.l
___
/~
o~~~--~~--~~~--~~
1000 2000 3000 4000 sooo 6000 7000 8000 9000 10000
Number of iterations
(b)
(a)
(c)
Figure 1: (a) The maze used in the experiment. S marks the start space, G the goal state, and 1, 2
and 3 are the three flags the agent has to set to receive the reward. (b) The DBN structure that captures
the independencies in this domain. (c) A graph showing the performance of the three procedures on
this example. PS is GenPS with a state-based model, PS+factored is the same procedure but with a
factored model, and GenPS exploits the factored model in prioritization. Each curve is the average
of 5 runs.
The evaluation of 8~(8:a). requires us to sum over a subset of the states -namely, those
1,1Ii
.z,
states t that are consistent with zi. Unfortunately, in the worst case this will be a large
fragment of the state space. If the number of environment states is not large, then this might
be a reasonable cost to pay for the additional benefits of GenPS. However, this might be a
burdensome when we have a large state space, which are the cases where we expect to gain
the most benefit from using generalized representations such as DBN.
In these situations, we propose a heuristic approach for estimating V'Q(s, a)~ without
summing over large numbers of states for computing the change of priority for each possible
state. This can be done by finding upper bounds on or estimates of 8~( 8:a ).. Once we
I,ll, ' %i
have computed these estimates, we can estimate the priority change for each state s. We
use the notation s '"Vi s* if sand s* both agree on the assignment to Pai . If Ci is an upper
bound on (or an estimate of)
18~~~:t:: I, we have that IV'Q(8, a)/loM ~ Li:s~iS. C i .
1
Thus, to evaluate the priority of state s, we simply find how "similar" it is to s*. Note
that it is relatively straightforward to use this equation to enumerate all the states where the
priority change might be large. Finally, we note that the use of a DBN as a model does not
change the way we update priorities after a value propagation step. If we use an explicit
table of values, then we would update priorities as in the previous section. If we use a
compact description of the value function, then we can apply GenPS to get the appropriate
update rule.
S An Experiment
We conducted an experiment to evaluate the effect of using GenPS with a generalizing
model. We used a maze domain similar to the one described in [6]. The maze, shown
in Figure 1(a), contains 59 cells, and 3 binary flags, resulting in 59 x 2 3 = 472 possible
states. Initially the agent is at the start cell (marked by S) and the flags are reset. The
agent has four possible actions, up, down, left, and right, that succeed 80% of the time,
and 20% of the time the agent moves in an unintended perpendicular direction. The i'th
flag is set when the agent leaves the cell marked by i. The agent receives a reward when
it arrives at the goal cell (marked by G) and all of the flags are set. In this situation, any
action resets the game. As noted in [6], this environment exhibits independencies. Namely,
the probability of transition from one cell to another does not depend on the flag settings.
Generalized Prioritized Sweeping
1007
These independencies can be captured easily by the simple DBN shown in Figure I(b) Our
experiment is designed to test the extent to which GenPS exploits the knowledge of these
independencies for faster learning.
We tested three procedures. The first is GenPS, which uses an explicit state-based
model. As explained above, this variant is essentially PS. The second procedure uses a
factored model of the environment for learning the model parameters, but uses the same
prioritization strategy as the first one. The third procedure uses the GenPS prioritization
strategy we describe in Section 4. All three procedures use the Boltzman exploration
strategy (see for example [5]). Finally, in each iteration these procedures process at most
10 states from the priority queue.
The results are shown in Figure l(c). As we can see, the GenPS procedure converged
faster than the procedures that used classic PS. As we can see, by using the factored model
we get two improvements. The first improvement is due to generalization in the model.
This allows the agent to learn a good model of its environment after fewer iterations. This
explains why PS+factored converges faster than PS. The second improvement is due to
the better prioritization strategy. This explains the faster convergence of GenPS.
6 Discussion
We have presented a general method for approximating the optimal use of computational
resources during reinforcement learning . Like classic prioritized sweeping, our method
aims to perform only the most beneficial value propagations. By using the gradient of the
Bellman equation our method generalizes the underlying principle in prioritized sweeping.
The generalized procedure can then be applied not only in the explicit, state-based case,
but in cases where approximators are used for the model. The generalized procedure also
extends to cases where a function approximator (such as that discussed in [3]) is used for
the value function, and future work will empirically test this application of GenPS. We are
currently working on applying GenPS to other types of model and function approximators.
Acknowledgments
We are grateful to Geoff Gordon, Daishi Harada, Kevin Murphy, and Stuart Russell for
discussions related to this work and comments on earlier versions of this paper. This
research was supported in part by ARO under the MURI program "Integrated Approach to
Intelligent Systems," grant number DAAH04-96-I-0341. The first author is supported by a
National Defense Science and Engineering Graduate Fellowship.
References
[I] S. Davies. Multidimensional triangulation and interpolation for reinforcement learning. In
Advances in Neurallnfonnation Processing Systems 9. 1996.
[2] T. Dean and K. Kanazawa. A model for reasoning about persistence and causation. Computationallntelligence, 5:142-150, 1989.
[3] G. J. Gordon. Stable function approximation in dynamic programming. In Proc. 12th
Int. Con! on Machine Learning, 1995.
[4] D. Heckerman. A tutorial on learning with Bayesian networks. Technical Report MSR-TR-9506, Microsoft Research, 1995. Revised November 1996.
[5] L. P. Kaelbling, M. L. Littman and A. W. Moore. Reinforcement learning: A survey. Journal
of Artificial Intelligence Research, 4:237-285, 1996.
[6] A. W. Moore and C. G. Atkeson. Prioritized sweeping-reinforcement learning with less data
and less time. Machine Learning, 13:103-130, 1993.
[7] R. S. Sutton. Integrated architectures for learning, planning, and reacting based on approximating
dynamic programming. In Machine Learning: Proc. 7th Int. Con!, 1990.
[8] P. Tadepalli and D. Ok. Scaling up average reward reinforcement learning by approximating the
domain models and the value function. In Proc. 13th Int. Con! on Machine Learning, 1996.
[9] R. J. Williams and L. C. III Baird. Tight performance bounds on greedy policies based on
imperfect value functions. Technical report, Computer Science, Northeastern University. 1993.
| 1409 |@word msr:1 briefly:1 version:1 middle:1 tadepalli:1 minus:1 tr:1 carry:1 contains:2 fragment:1 unintended:1 past:1 existing:1 current:2 nt:1 refines:1 ronald:1 happen:1 designed:1 update:26 pursued:1 selected:1 leaf:1 fewer:1 intelligence:1 greedy:1 record:1 provides:1 recompute:1 consists:1 combine:2 manner:2 introduce:3 indeed:1 expected:5 roughly:1 planning:4 nor:1 examine:1 bellman:14 discounted:3 estimating:2 notation:3 bounded:1 moreover:1 underlying:1 what:1 finding:3 berkeley:2 every:3 multidimensional:1 ti:1 nonnegligible:1 grant:1 yn:1 overestimate:2 before:2 engineering:1 esp:1 sutton:1 ak:3 reacting:1 interpolation:1 might:8 limited:1 factorization:1 perpendicular:1 graduate:1 directed:1 unique:1 acknowledgment:1 enforces:1 practice:1 implement:1 procedure:27 thought:1 davy:1 persistence:1 get:3 cannot:2 applying:1 equivalent:1 dean:1 straightforward:4 attention:1 williams:1 l:1 survey:1 assigns:2 immediately:1 factored:7 examines:1 rule:1 classic:6 notion:1 increment:1 updated:3 dbns:2 programming:3 prioritization:5 us:9 updating:4 muri:1 labeled:2 observed:1 capture:3 worst:2 ensures:2 russell:1 observes:1 noi:1 principled:1 environment:18 ui:1 reward:12 ideally:2 littman:1 dynamic:9 raise:1 depend:2 grateful:1 tight:1 upon:1 division:1 easily:2 joint:1 geoff:1 represented:2 describe:4 artificial:1 kevin:1 choosing:1 whose:1 heuristic:2 spend:1 valued:1 larger:1 quite:1 reconsidered:1 reconstruct:1 say:2 ability:1 jointly:1 differentiate:1 differentiable:1 propose:3 aro:1 product:1 reset:2 loop:2 achieve:2 description:4 getting:1 probabil:1 exploiting:1 parent:2 convergence:1 p:22 smax:5 generating:1 converges:1 derive:1 propagating:1 minor:1 received:2 c:1 involves:2 direction:1 exploration:2 sand:3 require:1 explains:2 assign:4 generalization:3 preliminary:1 summation:2 hall:1 ground:1 mapping:1 claim:2 parr:5 estimation:1 proc:3 currently:1 visited:1 largest:2 ere:1 aim:2 reaching:1 avoid:1 focus:4 daah04:1 improvement:3 contrast:1 sense:1 burdensome:1 integrated:2 initially:1 issue:2 arg:1 ill:1 among:1 denoted:5 special:1 once:1 never:1 pai:13 stuart:1 prioritizes:1 anticipated:1 discrepancy:2 future:1 report:2 np:2 simplify:2 inherent:1 primarily:2 few:1 gordon:2 intelligent:1 causation:1 recognize:2 national:1 murphy:1 familiar:1 dor:1 microsoft:1 friedman:4 attempt:3 evaluation:1 extreme:1 arrives:1 tuple:1 partial:1 necessary:1 experience:1 orthogonal:1 unless:1 iv:4 old:2 earlier:1 assignment:7 cost:1 kaelbling:1 vertex:2 entry:1 subset:1 harada:1 conducted:1 combined:1 density:1 together:3 reflect:1 choose:1 priority:37 resort:1 expert:1 return:1 li:2 lookup:1 int:3 baird:1 explicitly:2 depends:1 vi:1 multiplicative:1 root:1 reached:2 start:3 maintains:2 ass:2 ni:1 efficiently:1 identify:1 bayesian:5 converged:1 reach:2 andre:4 ed:1 frequency:1 dm:2 con:3 gain:1 adjusting:1 recall:2 knowledge:3 ok:1 violating:1 execute:1 done:2 ox:1 stage:1 hand:3 receives:1 working:1 propagation:11 mdp:2 effect:1 concept:1 verify:1 l1o:1 assigned:2 nonzero:1 moore:2 pri:5 deal:2 ll:1 game:1 self:1 during:1 noted:1 generalized:12 tt:1 complete:1 performs:3 reasoning:1 spending:2 recently:1 multinomial:2 empirically:1 sooo:1 thirdly:1 extend:3 discussed:1 refer:1 ai:5 dbn:10 similarly:1 reachable:1 interleaf:1 fictional:2 similarity:1 stable:1 add:1 neurallnfonnation:1 recent:1 triangulation:1 store:1 binary:1 approximators:3 yi:16 captured:1 fortunately:1 additional:1 determine:1 maximize:1 ii:2 desirable:1 technical:2 faster:4 visit:1 desideratum:1 basic:2 variant:1 essentially:2 metric:1 iteration:3 sometimes:1 represent:1 cell:5 receive:3 addition:1 want:4 fellowship:1 crucial:1 comment:1 inconsistent:1 call:2 presence:1 iii:1 easy:5 affect:1 psps:1 zi:5 architecture:1 imperfect:1 tradeoff:1 expression:1 utility:1 defense:1 reevaluate:1 queue:3 speaking:1 afford:1 action:14 reassign:1 enumerate:1 clear:1 involve:1 amount:1 discount:1 locally:1 simplest:1 specifies:1 supplied:1 tutorial:1 notice:1 estimated:1 affected:1 group:1 independency:4 four:1 neither:1 kept:1 graph:7 sum:1 run:1 parameterized:1 soda:1 extends:1 reader:1 reasonable:1 utilizes:1 decision:1 scaling:1 bound:7 pay:1 daishi:1 bv:1 constraint:1 performing:3 relatively:1 according:1 combination:1 describes:1 slightly:1 beneficial:1 heckerman:1 making:2 dv:2 explained:1 computationally:1 resource:2 equation:6 vq:2 agree:2 eventually:1 count:4 dyna:1 needed:2 know:1 end:1 available:2 generalizes:1 multiplied:1 apply:4 appropriate:2 original:1 dirichlet:2 exploit:2 fonnal:1 approximating:4 classical:2 objective:1 move:1 question:2 added:1 quantity:1 parametric:1 costly:1 strategy:5 exhibit:1 gradient:3 topic:1 extent:1 trivial:1 pet:1 assuming:1 modeled:1 retained:1 executed:1 unfortunately:1 relate:1 reconsider:1 policy:6 perform:3 upper:4 revised:1 markov:1 arc:1 finite:1 november:1 immediate:1 situation:3 precise:1 locate:1 sweeping:17 david:1 introduced:1 namely:2 california:1 beyond:1 below:1 maxaea:1 program:1 max:11 memory:1 unrealistic:1 ia:1 pact:1 scheme:2 mdps:2 nir:2 prior:5 val:5 repaired:1 loss:1 expect:1 acyclic:1 approximator:2 lv:1 agent:29 consistent:4 principle:7 balancing:1 ift:1 changed:2 supported:2 free:1 keeping:1 side:2 understand:1 sparse:2 benefit:2 curve:1 xn:4 transition:14 world:1 maze:3 author:1 made:1 reinforcement:11 atkeson:1 boltzman:1 approximate:3 compact:5 dealing:1 instantiation:1 summing:1 xi:7 spectrum:1 why:1 table:3 learn:3 concatenates:1 ca:1 complex:1 domain:4 main:1 motivation:1 n:1 explicit:8 wish:1 lie:1 third:1 northeastern:1 down:1 specific:1 showing:1 kanazawa:1 effectively:1 ci:1 magnitude:1 execution:3 sparseness:1 horizon:1 suited:1 generalizing:3 simply:2 likely:1 aa:1 determines:1 succeed:1 conditional:7 goal:2 marked:3 prioritized:17 change:29 approache:1 infinite:1 acting:1 flag:6 total:1 formally:2 mark:1 perfonning:2 evaluate:2 tested:1 |
450 | 141 | 232
SPEECH PRODUCTION USING A NEURAL
NETWORK WITH A COOPERATIVE
LEARNING MECHANISM
Mitsuo Komura
Akio Tanaka
International Institute for Advanced Study of Social Information Science,
Fujitsu Limited
140 Miyamoto, Numazu-shi Shizuoka, 410-03 Japan
ABSTRACT
We propose a new neural network model and its learning
algorithm. The proposed neural network consists of four layers
- input, hidden, output and final output layers. The hidden and
output layers are multiple. Using the proposed SICL(Spread
Pattern Information and Cooperative Learning) algorithm, it
is possible to learn analog data accurately and to obtain
smooth outputs. Using this neural network, we have developed
a speech production system consisting of a phonemic symbol
production subsystem and a speech parameter production
subsystem. We have succeeded in producing natural speech
waves with high accuracy.
INTRODUCTION
Our purpose is to produce natural speech waves. In general, speech synthesis
by rule is used for producing speech waves. However, there are some
difficulties in speech synthesis by rule. First, the rules are very complicated.
Second, extracting a generalized rule is difficult. Therefore, it is hard to
synthesize a natural speech wave by using rules. We use a neural network for
producing speech waves. Using a neural network, it is possible to learn speech
parameters without rules. (Instead of describing rules explicitly, selecting a
training data set becomes an important subject.) In this paper, we propose a
new neural network model and its learning algorithm. Using the proposed
neural network, it is possible to learn and produce analog data accurately. We
apply the network to a speech production system and examine the system
performance.
PROPOSED NEURAL NETWORK AND ITS LEARNING
ALGORITHM
We use an analog neuron-like element in a neural network. The element has a
logistic activation function presented by equation (3). As a learning algorithm,
Speech Production Using A Neural Network
the BP(Back Propagation) method is widely used. By using this method it is
possible to learn the weighting coefficients of the units whose target values are
not given directly. However, there are disadvantages. First, there are singular
points at 0 and 1 (outputs of the neuron-like element). Second, finding the
optimum values of learning constants is not easy. We have proposed a new
neural network model and its learning algorithm to solve this problem. The
proposed SICL(Spread Pattern Information and Cooperative Learning) method
has the following features.
(a)The singular points of the BP method are removed. (Outputs are not simply
oor 1.) This improves the convergence rate.
(b)A spread pattern information(SI) learning algorithm is proposed. In the SI
learning algorithm, the weighting coefficients from the hidden layers to the
output layers are fixed to random values. Pattern information is spread over
the memory space of the weighting coefficients. As a result, the network can
learn analog data accurately.
(c)A cooperative learning(CL) algorithm is proposed. This algorithm makes it
possible to obtain smooth and stable output. The CL system is shown in Fig.1
where D(L) is a delay line which delays L time units.
In the following sections, we define a three-layer network, introduce the BP
method, and propose the SICL method .
.. ' .'
k=K
",
.....,..?.?.
....
".'.
-'.
~,;"
,:yi
I';;
,/
Input layar
Hlddan layar.
Output layar.
Final output layar
Figure 1. Cooperative Learning System
(Speech Parameter I Phonemic Symbol Production Subsystem)
233
234
Komura and Tanaka
THREE?LAYER NETWORK
We define a three-layer network that has an input layer, a hidden layer, and
an output layer. The propagation of a signal from the input layer to the hidden
layer is represented as follows.
Uj
=
'EiwIHijXi,
Yj
=
f(Uj -OJ) ,
(1)
where i= 1,2,... ,1; j= 1,2, ... ,J and Xi is an input, Yj the output of the hidden
layer, and OJ a threshold value. The propagation of a signal from the hidden
layer to the output layer is represented as follows.
(2)
where Zk is the output of the output layer and k= 1,2, ... ,K. The activation
function f(u) is defined by
f(u) = (1+exp(-u+O?-l.
Settingy
(3)
= f(u), the derivation off(u) is then given by f' (u) =y(1-y).
BACK PROPAGATION (BP) METHOD
The three-layer BP algorithm is shown as follows. The back propagation error,
80 k(n) , for an output unit is given by
(4)
where n is an iteration number.
The back propagation error, 8Hin), for a hidden unit is given by
8Hin) = ('Ek80k(n) WHOjk) f' (u)<n? .
(5)
The change to the weight from the i-th to thej-th unit, flwIH;,in) is given by
flwIHi/n) =a8H/n)xi(n)
+ jJflwIH;,in-l) .
(6)
The change to the weight from thej-th to the k-th unit, flwH0jk(n) is given by
flwHOjk(n)
= a80k(n) Yin) + jJflwHOjk(n-l) ,
(7)
where a and jJ are learning constants, and have positive values.
SPREAD PATTERN INFORMATION AND
COOPERATIVE LEARNING (SICL) METHOD
The proposed learning algorithm - SICL method is shown as follows. The
propagation of a signal from the input layer to the hidden layer is given by
u/l,n) = 'EiwlHij(l,n-l) xi(n) ,
y/l,n) =f(u/l,n) -oin) ,
(8)
where l is a stage number (l =-LI2, ... ,LI2). The propagation of a signal from the
hidden layer to the output layer is given by
vk(l,n)
='EjwH0jk(n Yin) ,
zk(l,n) = f(vk(l,n) -Ok(n) .
The back propagation error, 80 k(n), for an output unit is given by
(9)
Speech Production Using A Neural Network
(10)
where y is a constant for removing singular points.
The back propagation error, OHj(l,n), for a hidden unit is given by
8HjCI,n)
= (EkoOk(l,n) wH0jk(l)
(Yil,n)(1-yjCl,n? + y) .
(11)
The change to the weight from the i-th to thej-th unit, ll.wIHi/l,n) is given by
AwIHij(l,n) =aOHjCl,n)xln)
+ jJAwIHij(l,n-l) .
(12)
The weight from thej-th to the k-th unit,wHOjk(l) is given by
wH?jk(l)
= Cjk(l),
(13)
where Cjk(l) is a fixed value and a random number with normal distribution. A
final output is a weighting sum of outputs zk(l,n) ,and is given by
ZFk(
n)
=E,Wl zk(l,n).
(14)
A SPEECH PRODUCTION SYSTEM USING THE
PROPOSED NEURAL NETWORK
The block diagram of a speech production system is shown in Fig.2. The system
consists of a phonemic symbol production subsystem and a speech parameter
production subsystem using the proposed neural network.
Phonemic Symbols
Input String
Speech ParameterslAnalog detal
=! ?
Speech Parameter
Production Subsystern
Phonemic SYlnbol
I'Iaduclim SiRy.1n _
@
1 """"?~-=0--1
Synthesized Speech
Automatic
Input Speech S~lg...Lnll~I-------li Speech Psremeter
ExtlClIon S.mystem
Figure 2. The Speech Production System Using the Proposed Neural Network
In the automatic speech parameter extraction subsystem@, speech
parameters are extracted automatically from an input speech signal. Speech
parameters are composed of source parameters( voiced/unvoiced ratio, pitch
and source power) and a vocal tract area function(PARCOR coefficients). The
extracted speech parameters are used as training data of the speech parameter
production subsystem.
In the speech parameter production subsystem?, input is the string of
phonemic symbols. (In the training stage, phonemic symbols are decided by the
teacher using input speech data. After training, phonemic symbols are given
by phonemic symbol production subsystem.) Targets are speech parameters
extracted in @ .
The phonemic symbol production subsystem CD consists of a preprocessor and a
learning system using the proposed neural network. In the preprocessor, a
235
236
Komura and Tanaka
string ofinput characters is converted to a string of phonemic symbols with the
mean length of utterance. The input is the string of phonemic symbols
converted by the preprocessor. In the training stage, the targets are actual
phonemic symbols, namely, the inputs of subsystem @ which are decided by
the teacher.
The output speech parameters are converted into the synthesized speech wave
using the PARCOR synthesizer circuit~.
EXPERIMENTS
We performed two separate experiments.
Experiment I
Automatic Speech Parameter Extraction Subsystem@
Sampling frequency: 8 KHz
Frame length : 20 ms
Frame period: 10 ms
Speech Parameter Production Subsystem @
Input layer: 1,044(36 X 29) units
Hidden layers: 80 units X 9 stages
Output layers: 13 units X 9 stages
Final output layer: 13 units
For source parameters, 3(VIUV, pitch, source power) units are assigned. For a
vocal tract area function, 10(PARCOR coefficients) units are assigned. This
system has an input layer, 9 hidden layers, 9 output layers, and a final output
layer. For each output layer, different target data are assigned. In the final
output layer, these outputs are summed with weighting coefficients.
Phonemic Symbol Production SubsystemCD
Input layer: 1,044(36 X 29) units Hidden layers: 80 units X 9 stages
Output layers: 36 units X 9 stages Final output layer: 36 units
Input Speech Signal No.1
The input speech signal No.1 is r ASAHA YAKU BANGARONI DENPOGA
TODOITA J, that is, Japanese sentence which means that "A telegram was
sent to the bungalow early in the morning.". This signal is a 408 frame (4.08s)
sequence.
Experimental Results
In Fig.3, the targets and the outputs of the speech parameter production
subsystem are shown. The dotted line is the sequence of target values. The
solid line is the sequence of output values after the training stage. In this case,
the learning constant a is equal to 0.7 and jJ to 0.2. After training, the actual
outputs produced by SICL system agree well with the targets.
Fig.4 (a) shows the learning behaviors of the speech parameter production
subsystem. In this case, the input speech signal is No.1. The learning constant
a is equal to 0.7 and P to 0.2. The learning curve based on the SICL and SI
Speech Production Using A Neural Network
methods converged. However, the learning curve based on the BP method did
not converge.
Fig.4 (b) also shows the learning behaviors. The learning constant a is equal to
0.07 and jJ to 0.02. In this case, all of the learning curves converged. These
results show that if we use the SICL or SI method instead of the BP method, we
can obtain better results.
1.0
.\
~.
If '.
I I
0.5
I
0.0
o
1.0
2.0
3.0
4.0 s
a = 0.7,/1 = 0.2
Input speech signal : No.1
Source Power
Figure 3. Targets and Outputs of Speech Parameter Production Subsystem
MSE(log)
MSE(log)
1
1
10-1
10-1
10-2
10-2
10-3
10-3
10- 4
10-4
10-5
10.5
1
10
102
103
(a) a = 0.7 ,jJ = 0.2
(NNI) 1
10
(b) a
102
103
= 0.07 ,p = 0.02
MSE: Mean square error
NNI : Normalized number of iteration
G-?J : SICL method (9stage) G-B: SI method (l = 0)
BP method
Figure 4. Learning Behavior of Speech Parameter Production Subsystem
* *:
Table 1 shows mean square errors of the system using SICL method for a, p. It
should be noted that the domain of convergence is very wide. From these
237
238
Komura and Tanaka
experiments, it is seen that the SICL method almost always allows stable and
smooth output to be obtained.
We also examined the system performance for another input speech signal
which is a 1,60S frame (16.0S s) sequence. The learning curves converged and
were similar to those for the input speech signal No.1
0.01
0.02
0.1
0.2
0.35
0.7
0.07
0.20
0.7
7.9S X 10-5
7.97X10-5
7.8SX 10-5
7.85 X 10-5
7.82x 10-5
7.80X 10-5
7.S2X 10-5
7.S2X10-5
7.S1 X 10- 5
7.S0X 10-5
7.S1 X 10-5
8_26X 10-5
1.03 X 10-4
1.09X 10-4
1.26 X 10-4
1.72 X 10- 4
2.S6X 10- 4
1.72x 10-3
Input speech signal: No.1.
TABLE 1. Mean Square Errors for a, p
Experiment II
Speech Parameter Production Subsystem @
Input layer: 665 (35X 19)units
Hidden layers: 400 units X 9 stages
Output layers: 14 (12 + 2)units X 9 stages
For the first target data(VIUV, power ratio, 10 PARCOR coefficients), 12 units
are . assigned. For the second target data(pitch, source power), 2 units are
assigned.
.
Input Speech Signal No.2
This signal is a 1:),700 frame (137s) sequence shown as follows.
IARAARN,IARAIRN, .... ,1ARAKARN, I ARAKIRN, ..... ,1ARASARN, ..... .
I ARARAN,I!RARA!/, ..... ,1KARARAKN ,1KIRARAKI/, ... ,1SARARASN, .. ...
In experiment II , a learning test for producing an arbitrary combination of the
input phonemes (ai) was done. The learning behavior of subsystem? is shown
in Fig.5. In this case, the fIrst 12 target data items are trained. The other 2
target data it~ms should be trained using another network. The upper line is a
learning curve for the SI method, and the lower line for the SICL method.
Some of the learning curves for the SI method didn't converge. However, the
learning curve for the SICL method did. We can thus say that the SICL method
is very powerful for actual use. (In this case, 400 hidden units is not sufficient
for the size of input data. If more hidden units are used, the MSE will be small.)
CONCLUSION
Speech Production Using A Neural Network
MSE (log) ~---------...,
G-B : SICL method (9stage)
10-1
G-E) : SI method (l= 0)
10-2
a =0.2,
10-3
11 =0.035
Input speech signal: No.2
10-4
1
10
102
NNI
Figure 5. Learning Behavior of Speech Parameter Production Subsystem
From the experiments shown in previous section, it should be noted that using
the SICL (Spread Pattern Information and Cooperative Learning) method
makes it possible to learn speech parameters or phonemic symbols stably and
to produce more natural speech waves than those synthesized by rules. If the
input is the combination of a word and postpositional particle, it is easy to
produce sound for unknown input data using the proposed speech production
system. However, the number of the hidden units and training data will be
great. Therefore we have to make the system learn phonemes among phoneme
sequence. Using this training data makes it possible to produce an arbitrary
sequence of phonemes. In experiment II , phonemic information (VIUV power
ratio, PARCOR coefficients) is trained. Then the range of input window was
set to be 190 ms. For prosodic information (pitch, source power), we must use
another network. Because, if we want to make the system learn prosodic
information, we must set the range of input data wider than that of words.
U sing these strategies, it is possible to produce arbitrary natural speech.
I
Acknowledgments
The authors thank Dr. Tosio Kitagawa, the president of liAS-SIS for his
encouragement.
References
D. E. Rumelhart, J. L. McClelland and PDP Research Group Parallel
Distributed Processing. VOL.l The MIT Press (1987).
T.J.Sejnowski and C.R.Rosenberg Parallel Networks that Learn to Pronounce
English Text. Complex Systems, 1, pp.145-168 (1987).
M.Komura and A.Tanaka Speech Synthesis Using a Neural Network with a
Cooperative Learning Mechanism. IEICE Tech. Rep. MBE88-8 (1988) (in
Japanese).
239
| 141 |@word uj:2 normalized:1 assigned:5 strategy:1 ll:1 solid:1 noted:2 separate:1 thank:1 m:4 generalized:1 selecting:1 rara:1 kitagawa:1 length:2 hin:2 s2x:1 activation:2 synthesizer:1 si:9 must:2 normal:1 exp:1 great:1 ratio:3 difficult:1 lg:1 yil:1 early:1 khz:1 purpose:1 analog:4 synthesized:3 unknown:1 item:1 upper:1 neuron:2 unvoiced:1 ai:1 wl:1 encouragement:1 automatic:3 sing:1 parcor:5 particle:1 mit:1 always:1 frame:5 pdp:1 stable:2 arbitrary:3 rosenberg:1 consists:3 namely:1 vk:2 sentence:1 introduce:1 tech:1 rep:1 behavior:5 tanaka:5 examine:1 yi:1 seen:1 pattern:6 automatically:1 actual:3 hidden:18 window:1 converge:2 period:1 becomes:1 signal:16 ii:3 memory:1 multiple:1 circuit:1 didn:1 among:1 sound:1 x10:1 smooth:3 oj:2 power:7 string:5 natural:5 summed:1 developed:1 difficulty:1 advanced:1 equal:3 finding:1 extraction:2 sampling:1 pitch:4 utterance:1 iteration:2 unit:28 text:1 producing:4 composed:1 positive:1 want:1 diagram:1 singular:3 source:7 consisting:1 mitsuo:1 subject:1 sent:1 examined:1 sufficient:1 extracting:1 limited:1 fujitsu:1 cd:1 range:2 easy:2 production:29 decided:2 acknowledgment:1 pronounce:1 xln:1 yj:2 succeeded:1 english:1 block:1 institute:1 wide:1 area:2 miyamoto:1 distributed:1 curve:7 word:2 vocal:2 author:1 oor:1 disadvantage:1 speech:59 subsystem:20 jj:4 social:1 shi:1 delay:2 mcclelland:1 teacher:2 xi:3 rule:8 dotted:1 international:1 table:2 his:1 learn:9 off:1 li2:2 zk:4 president:1 synthesis:3 vol:1 target:12 group:1 four:1 mse:5 threshold:1 cl:2 japanese:2 complex:1 domain:1 did:2 oin:1 synthesize:1 element:3 rumelhart:1 jk:1 s0x:1 dr:1 spread:6 li:1 japan:1 cooperative:8 converted:3 sum:1 powerful:1 fig:6 coefficient:8 almost:1 explicitly:1 removed:1 performed:1 wave:7 layer:39 weighting:5 complicated:1 parallel:2 nni:3 voiced:1 removing:1 trained:3 preprocessor:3 square:3 accuracy:1 phoneme:4 symbol:14 bp:8 cjk:2 represented:2 accurately:3 produced:1 thej:4 derivation:1 prosodic:2 sx:1 sejnowski:1 converged:3 combination:2 lnll:1 yin:2 simply:1 whose:1 character:1 widely:1 solve:1 say:1 frequency:1 pp:1 s1:2 final:7 wh:1 sequence:7 equation:1 agree:1 extracted:3 improves:1 describing:1 propose:3 mechanism:2 back:6 ok:1 hard:1 change:3 apply:1 done:1 experimental:1 stage:12 convergence:2 optimum:1 produce:6 tract:2 propagation:10 wider:1 morning:1 logistic:1 stably:1 viuv:3 ieice:1 phonemic:16 |
451 | 1,410 | A mathematical model of axon guidance by
diffusible factors
Geoffrey J. Goodhill
Georgetown Institute for Cognitive and Computational Sciences
Georgetown University Medical Center
3970 Reservoir Road
Washington DC 20007
geoff@giccs.georgetown.edu
Abstract
In the developing nervous system, gradients of target-derived diffusible factors play an important role in guiding axons to appropriate targets. In this paper, the shape that such a gradient might
have is calculated as a function of distance from the target and the
time since the start of factor production. Using estimates of the
relevant parameter values from the experimental literature, the
spatiotemporal domain in which a growth cone could detect such
a gradient is derived. For large times, a value for the maximum
guidance range of about 1 mm is obtained. This value fits well
with experimental data. For smaller times, the analysis predicts
that guidance over longer ranges may be possible. This prediction
remains to be tested.
1 Introduction
In the developing nervous system, growing axons are guided to targets that may be
some distance away. Several mechanisms contribute to this (reviewed in TessierLavigne & Goodman (1996?. One such mechanism is the diffusion of a factor
from the target through the extracellular space, creating a gradient of increasing
concentration that axons can sense and follow. In the central nervous system, such
a process seems to occur in at least three cases: the guidance ofaxons from the
trigeminal ganglion to the maxillary process in the mouse (Lumsden & Davies,
1983, 1986), of commissural axons in the spinal cord to the floor plate (TessierLavigne et al., 1988), and ofaxons and axonal branches from the corticospinal tract
to the basilar pons (Heffner et al., 1990). The evidence for this comes from both in
vivo and in vitro experiments. For the latter, a piece of target tissue is embedded in
a three dimensional collagen gel near to a piece of tissue containing the appropriate
G. J Goodhill
160
population of neurons. Axon growth is then observed directed towards the target,
implicating a target-derived diffusible signal. In vivo, for the systems described,
the target is always less than 500 J..lm from the population ofaxons. In vitro, where
the distance between axons and target can readily be varied, guidance is generally
not seen for distances greater than 500 - 1000 J..lm. Can such a limit be explained
in terms of the mathematics of diffusion?
There are two related constraints that the distribution of a diffusible factor must
satisfy to provide an effective guidance cue at a point. Firstly, the absolute concentration of factor must not be too small or too large. Secondly, the fractional change
in concentration of factor across the width of the gradient-sensing apparatus, generally assumed to be the growth cone, must be sufficiently large. These constraints
are related because in both cases the problem is to overcome statistical noise. At
very low concentrations, noise exists due to thermal fluctuations in the number of
molecules of factor in the vicinity of the growth cone (analyzed in Berg & Purcell
(1977?. At higher concentrations, the limiting source of noise is stochastic variation in the amount of binding of the factor to receptors distributed over the growth
cone. At very high concentrations, all receptors will be saturated and no gradient
will be apparent. The closer the concentration is to the upper or lower limits, the
higher the gradient that is needed to ensure detection (Devreotes & Zigmond, 1988;
Tessier-Lavigne & Placzek, 1991). The limitations these constraints impose on the
guidance range of a diffusible factor are now investigated. For further discussion
see Goodhill (1997; 1998).
2
Mathematical model
Consider a source releasing factor with diffusion constant D cm2 /sec, at rate q
moles/sec, into an infinite, spatially uniform three-dimensional volume. Initially,
zero decay of the factor is assumed. For radially symmetric Fickian diffusion in
three dimensions, the concentration C(r, t) at distance r from the source at time t
is given by
q
r
(1)
C(r, t) = - 4
D erfc r;-r:;-;
7r r
v4Dt
(see e.g. Crank (1975?, where erfc is the complementary error function. The percentage change in concentration p across a small distance D.r (the width of the
growth cone) is given by
D.r [
p
= --;:-
r
e-r2/4Dt
1 + J7rDt erfc(r/J4Dt)
1
This function has two perhaps surprising characteristics. Firstly, for fixed r,
(2)
Ipi
decreases with t. That is, the largest gradient at any distance occurs immediately
after the source starts releaSing factor. For large t, Ipi asymptotes at D.r Jr. Secondly,
for fixed t < 00, numerical results show that p is nonmonotonic with r. In particular
it decreases with distance, reaches a minimum, then increases again. The position
of this minimum moves to larger distances as t increases.
The general characteristics of the above constraints can be summarized as follows.
(1) At small times after the start of production the factor is very unevenly distributed. The concentration C falls quickly to almost zero moving away from the
source, the gradient is steep, and the percentage change across the growth cone
p is everywhere large. (2) As time proceeds the factor becomes more evenly distributed. C everywhere increases, but p everywhere decreases. (3) For large times,
C tends to an inverse variation with the distance from the source r, while Ipi tends
A Mathematical Model ofAxon Guidance by Diffusible Factors
161
to 6.r/r independent of all other parameters. This means that, for large times,
the maximum distance over which guidance by diffusible factors is possible scales
linearly with growth cone diameter 6.r.
3 Parameter values
Diffusion constant, D. Crick (1970) estimated the diffusion constant in cytoplasm
for a molecule of mass 0.3 - 0.5 kDa to be about 10- 6 cm2 /sec. Subsequently, a
direct determination of the diffusion constant for a molecule of mass 0.17 kDa in
the aqueous cytoplasm of mammalian cells yielded a value of about 3.3 x 10- 6
cm2 / sec (Mastro et aL, 1984). By fitting a particular solution of the diffusion equation to their data on limb bud determination by gradients of a morphogenetically
active retinoid, Eichele & Thaller (1987) calculated a value of 10- 7 cm2 /sec for this
molecule (mass 348.5 kDa) in embryonic limb tissue. One chemically identified
diffusible factor known to be involved in axon guidance is the protein netrin-1,
which has a molecular mass of about 75 kDa (Kennedy et al., 1994). D should
scale roughly inversely with the radius of a molecule, Le. with the cube root of its
mass. Taking the value of 3.3 x 10- 6 cm2 /sec and scaling it by (170/75,000)1/3
yields 4.0 x 10- 7 cm2 /sec. This paper therefore considers D = 10- 6 cm2 /sec and
D = 10- 7 cm2 /sec.
Rate of production of factor q. This is hard to estimate in vivo: some insight can
be gained by considering in vitro experiments. Gundersen & Barrett (1979) found
a turning response in chick spinal sensory axons towards a nearby pipette filled
with a solution of NGF. They estimated the rate of outflow from their pipette to
be 1 /LI/hour, and found an effect when the concentration in the pipette was as
low as 0.1 nM NGF (Tessier-Lavigne & Placzek, 1991). This corresponds to a q of
3 x 1O- 11 nM/sec. Lohof et al. (1992) studied growth cone turning induced by
a gradient of cell-membrane permeant cAMP from a pipette containing a 20 mM
solution and a release rate of the order of 0.5 pI/sec: q = 10- 5 nM/sec. Below a
further calculation for q is performed, which suggests an appropriate value may
be q = 10- 7 nM/sec.
Growth cone diameter, 6.r. For the three systems mentioned above, the diameter
of the main body of the growth cone is less than 10 /Lm. However, this ignores
filopodia, which can increase the effective width for gradient sensing purposes.
The values of 10 /Lm and 20 /Lm are considered below.
Minimum concentration for gradient detection. Studies of leukocyte chemotaxis
suggest that when gradient detection is limited by the dynamics of receptor binding rather than physical limits due to a lack of molecules of factor, optimal detection occurs when the concentration at the growth cone is equal to the dissociation
constant for the receptor (Zigmond, 1981; Devreotes & Zigmond, 1988). Such studies also suggest that the low concentration limit is about 1% of the dissociation constant (Zigmond, 1981). The transmembrane protein Deleted in Colorectal Cancer
(DeC) has recently been shown to possess netrin-1 binding activity, with an order
of magnitude estimate for the dissociation constant of 10 nM (Keino-Masu et aI,
1996). For comparison, the dissociation constant of the low-affinity NGF receptor
P75 is about 1 nM (Meakin & Shooter, 1992). Therefore, low concentration limits of
both 10- 1 nM and 10- 2 nM will be considered.
Maximum concentration for gradient detection. Theoretical considerations suggest that, for leukocyte chemotaxis, sensitivity to a fixed gradient should fall off
symmetrically in a plot against the log of background concentration, with the peak
at the dissociation constant for the receptor (Zigmond, 1981). Raising the con-
162
G. 1. Goodhill
centration to several hundred times the dissociation constant appears to prevent
axon guidance (discussed in Tessier-Lavigne & Placzek (1991?. At concentrations
very much greater than the dissociation constant, the number of receptors may be
downregulated, reducing sensitivity (Zigm0nd, 1981). Given the dissociation constants above, 100 nM thus constitutes a reasonable upper bound on concentration.
Minimum percentage change detectable by a growth cone, p. By establishing
gradients of a repellent, membrane-bound factor directly on a substrate and measuring the response of chick retinal axons, Baier & Bonhoeffer (1992) estimated p to
be about 1%. Studies of cell chemotaxis in various systems have suggested optimal
values of 2%: for concentrations far from the dissociation constant for the receptor,
p is expected to be larger (Devreotes & Zigmond, 1988). Both p = 1% and p = 2%
are considered below.
4 Results
In order to estimate bounds for the rate of production of factor q for biological
tissue, the empirical observation is used that, for collagen gel assays lasting of the
order of one day, guidance is generally seen over distances of at most 500 I'm
(Lumsden & Davies, 1983,1986; Tessier-Lavigne et al., 1988). Assume first that this
is constrained by the low concentration limit. Substituting the above parameters
(with D = 10- 7 cm2/sec) into equation 1 and specifying that C(500pm, 1 day) =
0.01 nM gives q :::::: 10- 9 nM/ sec. On the other hand, assuming constraint by the
high concentration limit, i.e. C(500pm, 1 day) = 100 nM, gives q :::::: 10- 5 nM/sec.
Thus it is reasonable to assume that, roughly, 10- 9 nM/ sec < q < 10- 5 nM/ sec.
The results discussed below use a value in between, namely q = 10- 7 nM/sec.
The constraints arising from equations 1 and 2 are plotted in figure 1. The cases of
D = 10- 6 cm2/sec and D = 10- 7 cm2/sec are shown in (A,C) and (B,D) respectively. In all four pictures the constraints C = 0.01 nM and C = 0.1 nM are plotted.
In (A,B) the gradient constraint p = 1% is shown, whereas in (C,D) p = 2% is
shown. These are for a growth cone diameter of 10 I'm. The graph for a 2% change
and a growth cone diameter of 20 I'm is identical to that for a 1% change and a
diameter of 10 I'm. Each constraint is satisfied for regions to the left of the relevant line. The line C = 100 nM is approximately coincident with the vertical axis
in all cases. For these parameters, the high concentration limit does not therefore
prevent gradient detection until the axons are within a few microns of the source,
and it is thus assumed that it is not an important constraint.
As expected, for large t the gradient constraint asymptotes at D.r Ir = p, i.e.
r = 1000 I'm for p = 1% and r = 500 I'm for p = 2% and a 10 I'm growth cone. That
is, the gradient constraint is satisfied at all times when the distance from the source
is less than 500 I'm for p = 2% and D.r = 10 I'm. The gradient constraint lines
end to the right because at earlier times p exceeds the critical value over all distances (since the formula for p is non-monotonic with r, there is sometimes another
branch of each p curve (not shown) off the graph to the right). As t increases from
zero, guidance is initially limited only by the concentration constraint. The maximum distance over which guidance can occur increases smoothly with t, reaching
for instance 1500 I'm (assuming a concentration limit of 0.01 nM) after about 2
hours for D = 10- 6 cm2/sec and about 6 hours for D = 10- 7 cm2/sec. However
at a particular time, the gradient constraint starts to take effect and rapidly reduces
the maximum range of guidance towards the asymptotic value as t increases. This
time (for p = 2%) is about 2 hours for D = 10- 6 cm 2/sec, and about one day for
D = 10- 7 cm2 / sec. It is clear from these pictures that although the exact size of
A Mathematical Model ofAxon Guidance by Diffusible Factors
A
163
B
Ci)
>-
'"
Ci)
>-
4.0
~
CD
E
'"
... C=O.OlnM
. C=O.lnM
3.5
i=
CD
E
-
3.5
i=
"
3.0
2.5
2.0
2.0
1.5
1.5
1.0
1.0
0.5
0.5
0.0
0.0
500
1000
...
3.0
p=O.OI
2.5
0
"'-J
4.0
~
1500 2000 2500
Distance (microns)
...
-
C=O.OlnM
C = O.lnM
p= 0.01
....... "
0
500
1000
. .. "
,
",
"
1500 2000 2500
Distance (microns)
D
C
~
Ci)
>-
'"
~
4.0
... C = O.OlnM
., . C = O.lnM
CD
E 3.5
i=
. 3.0
-
p=0.02
'"
4.0
CD
E 3.5
i=
3.0
2.5
2.5
2.0
2.0
1.5
1.5
1.0
1.0
.. ' ., '
0.5
0.0
0
1500 2000
2500
Distance (microns)
... C = O.OlnM
., . C=O.lnM
- P =0.02
~
0.5
!~
0.0
0
500
.........
1000
.. ' ."
.. .' "
" "
1500 2000 2500
Distance (microns)
Figure 1: Graphs showing how the gradient constraint (solid line) interacts with
the minimum concentration constraint (dashed/dotted lines) to limit guidance
range, and how these constraints evolve over time. The top row (A,B) is for p = 1%,
the bottom row (C,D) for p = 2%. The left column (A,C) is for D = 10- 6 cm2 / sec,
the right column (B,D) for D = 10- 7 cm2 /sec. Each constraint is satisfied to the
left of the appropriate curve. It can be seen that for D = 1O- 6 cm2 / s~c the gradient
limit quickly becomes the dominant constraint on maximum guidance range. In
contrast for D = 10- 7 cm2 / sec, the concentration limit is the dominant constraint
at times up to several days. However after this the gradient constraint starts to
take effect and rapidly reduces the maximum guidance range.
164
G. 1. Goodhill
the diffusion constant does not affect the position of the asymptote for the gradient
constraint, it does play an important role in the interplay of constraints while the
gradient is evolving. The effect is however subtle: reducing D from 10- 6 cm2 /sec
to 10- 7 cm 2 /sec increases the time for the C = 0.01 nM limit to reach 2000 J..tm, but
decreases the time for the C = 0.1 nM limit to reach 2000 f..lm.
5 Discussion
Taking the gradient constraint to be a fractional change of at least 2% across a
growth cone of width of 10 f..lm or 20 J..tm yields asymptotic values for the maximum distance over which guidance can occur once the gradient has stabilized
of 500 f..lm and 1000 f..lm respectively. This fits well with both in vitro data, and
the fact that for the systems mentioned in the introduction, the growing axons are
always less than 500 f..lm from the target in vivo. The concentration limits seem
to provide a weaker constraint than the gradient limit on the maximum distances
possible. However, this is very dependent on the value of q, which has only been
very roughly estimated: if q is significantly less than 10- 7 nM/sec, the low concentration limits will provide more restrictive constraints (q may well have different
values in different target tissues). The gradient constraint curves are independent
of q. The gradient constraint therefore provides the most robust explanation for
the observed guidance limit.
The model makes the prediction that guidance over longer distances than have
hitherto been observed may be possible before the gradient has stabilized. In the
early stages following the start of factor production the concentration falls off more
steeply, providing more effective guidance. The time at which guidance range is a
maximum defends on the diffusion constant D. For a rapidly diffusing molecule
(D >::::: 1O- 6 cm jsec) this occurs after only a few hours. For a more slowly diffusing
molecule however (D >::::: 10- 7 cm2 jsec) this occurs after a few days, which would
be easier to investigate in vitro. In vivo, molecules such as netrin-1 may thus be
large because, during times immediately following the start of production by the
source, there could be a definite benefit (i.e. steep gradient) to a slowly-diffusing
molecule. Also, it is conceivable that Nature has optimized the start of production of factor relative to the time that guidance is required in order to exploit an
evolving gradient for extended range. This could be especially important in larger
animals, where axons may need to be guided over longer distances in the developing embryo.
Bibliography
Baier, H. & Bonhoeffer, F. (1992). Axon guidance by gradients of a target-derived
component. Science, 255,472-475.
Berg, H.C. and Purcell, E.M. (1977). Physics of chemoreception. Biophysical Journal,
20,193-219.
Crick, F. (1970). Diffusion in embryogenesis. Nature, 255,420-422.
Crank, J. (1975). The mathematics of diffusion, Second edition. Oxford, Clarendon.
Devreotes, P.N. & Zigmond, S.H. (1988). Chemotaxis in eukaryotic cells: a focus on
leukocytes and Dictyostelium. Ann. Rev. Cell. Bioi., 4, 649-686.
Eichele, G. & Thaller, C. (1987). Characterization of concentration gradients of a
morphogenetically active retinoid in the chick limb bud. J. Cell. Bioi., 105, 19171923.
A Mathematical Model ofAxon Guidance by Diffusible Factors
165
Goodhill, G.J. (1997). Diffusion in axon guidance. Eur. J. Neurosci., 9,1414-1421.
Goodhill, G.J. (1998). Mathematical guidance for axons. Trends. Neurosci., in press.
Gundersen, R.W. & Barrett, J.N. (1979). Neuronal chemotaxis: chick dorsal-root
axons tum toward high concentrations of nerve growth factor. Science, 206,
1079-1080.
Heffner, C.D., Lumsden, AG.5. & O'Leary, D.D.M. (1990). Target control of collateral extension and directional growth in the mammalian brain. Science, 247,
217-220.
Keino-Masu, K., Masu, M., Hinck, L., Leonardo, E.D., Chan, S.5.-Y., Culotti, J.G. &
Tessier-Lavigne, M. (1996). Deleted in Colorectal Cancer (DCC) encodes a netrin
receptor. Cell, 87, 175-185.
Kennedy, T.E., Serafini, T., de al Torre, J.R. & Tessier-Lavigne, M. (1994). Netrins are
diffusible chemotropic factors for commissural axons in the embryonic spinal
cord. Cell, 78,425-435.
Lohof, A.M., Quillan, M., Dan, Y, & Poo, M-m. (1992). Asymmetric modulation of
cytosolic cAMP activity induces growth cone turning. J. Neurosci., 12, 12531261.
Lumsden, AG.s. & Davies, A.M. (1983). Earliest sensory nerve fibres are guided to
peripheral targets by attractants other than nerve growth factor. Nature, 306,
786-788.
Lumsden, A.G.5. & Davies, AM. (1986). Chemotropic effect of specific target epithelium in the developing mammalian nervous system. Nature, 323, 538-539.
Mastro, AM., Babich, M.A, Taylor, W.D. & Keith, AD. (1984). Diffusion of a small
molecule in the cytoplasm of mammalian cells. Proc. Nat. Acad. Sci. USA, 81,
3414-3418.
Meakin, S.O. & Shooter, E.M. (1992). The nerve growth family of receptors. Trends.
Neurosci., 15, 323-331.
Tessier-Lavigne, M. & Placzek, M. (1991). Target attraction: are developing axons
guided by chemotropism? Trends Neurosci., 14,303-310.
Tessier-Lavigne, M. & Goodman, C.S. (1996). The molecular biology of axon guidance. Science, 274, 1123-1133.
Tessier-Lavigne, M., Placzek, M., Lumsden, A.G.5., Dodd, J. & Jessell, T.M. (1988).
Ch~motropic guidance of developing axons in the mammalian central nervous system. Nature, 336, 775-778.
Tranquillo, R.T. & Lauffenburger, D.A. (1987). Stochastic model of leukocyte
chemosensory movement. J. Math. BioI., 25,229-262.
Zigmond, S.H. (1981). Consequences of chemotactic peptide receptor modulation
for leukocyte orientation. J. Cell. BioI., 88, 644-647.
| 1410 |@word seems:1 cm2:20 solid:1 surprising:1 must:3 readily:1 numerical:1 shape:1 asymptote:3 plot:1 cue:1 nervous:5 meakin:2 characterization:1 provides:1 math:1 contribute:1 firstly:2 ipi:3 mathematical:6 direct:1 epithelium:1 fitting:1 dan:1 expected:2 roughly:3 growing:2 brain:1 considering:1 increasing:1 becomes:2 mass:5 hitherto:1 cm:3 ag:2 growth:22 control:1 medical:1 before:1 apparatus:1 limit:18 tends:2 acad:1 receptor:11 consequence:1 oxford:1 establishing:1 fluctuation:1 modulation:2 approximately:1 might:1 studied:1 suggests:1 specifying:1 corticospinal:1 limited:2 range:9 directed:1 definite:1 dodd:1 empirical:1 evolving:2 significantly:1 davy:4 road:1 protein:2 suggest:3 center:1 poo:1 immediately:2 lumsden:6 insight:1 attraction:1 population:2 variation:2 limiting:1 target:17 play:2 exact:1 substrate:1 trend:3 giccs:1 mammalian:5 asymmetric:1 predicts:1 observed:3 role:2 bottom:1 cord:2 region:1 decrease:4 movement:1 transmembrane:1 mentioned:2 dynamic:1 geoff:1 various:1 effective:3 nonmonotonic:1 apparent:1 larger:3 interplay:1 biophysical:1 relevant:2 rapidly:3 tessier:9 tract:1 basilar:1 keith:1 defends:1 come:1 guided:4 radius:1 torre:1 stochastic:2 subsequently:1 biological:1 secondly:2 extension:1 attractants:1 chemotactic:1 mm:2 sufficiently:1 considered:3 lm:10 substituting:1 early:1 purpose:1 proc:1 peptide:1 largest:1 always:2 rather:1 reaching:1 earliest:1 derived:4 release:1 focus:1 steeply:1 contrast:1 detect:1 sense:1 camp:2 am:2 dependent:1 initially:2 embryogenesis:1 orientation:1 animal:1 constrained:1 cube:1 equal:1 once:1 washington:1 identical:1 biology:1 constitutes:1 collagen:2 few:3 filopodia:1 detection:6 investigate:1 saturated:1 analyzed:1 aqueous:1 closer:1 collateral:1 filled:1 taylor:1 plotted:2 guidance:31 theoretical:1 instance:1 column:2 earlier:1 measuring:1 pons:1 uniform:1 hundred:1 too:2 spatiotemporal:1 eur:1 peak:1 sensitivity:2 off:3 chemotaxis:5 physic:1 mouse:1 quickly:2 leary:1 again:1 central:2 nm:23 satisfied:3 containing:2 slowly:2 cognitive:1 creating:1 li:1 de:1 retinal:1 sec:31 summarized:1 satisfy:1 ad:1 piece:2 performed:1 root:2 start:8 vivo:5 oi:1 ir:1 characteristic:2 dissociation:9 yield:2 directional:1 kennedy:2 tissue:5 reach:3 against:1 involved:1 commissural:2 con:1 radially:1 fractional:2 subtle:1 purcell:2 appears:1 nerve:4 clarendon:1 higher:2 dt:1 day:6 follow:1 tum:1 response:2 dcc:1 stage:1 until:1 hand:1 lack:1 perhaps:1 usa:1 effect:5 vicinity:1 spatially:1 symmetric:1 assay:1 during:1 width:4 plate:1 ofaxons:3 consideration:1 recently:1 vitro:5 physical:1 spinal:3 volume:1 discussed:2 ai:1 mathematics:2 pm:2 devreotes:4 dictyostelium:1 moving:1 longer:3 dominant:2 chan:1 seen:3 minimum:5 greater:2 floor:1 impose:1 mole:1 signal:1 dashed:1 branch:2 reduces:2 exceeds:1 determination:2 calculation:1 molecular:2 pipette:4 prediction:2 sometimes:1 cell:10 dec:1 background:1 whereas:1 unevenly:1 source:9 goodman:2 releasing:2 posse:1 induced:1 seem:1 axonal:1 near:1 diffusible:11 symmetrically:1 diffusing:3 affect:1 fit:2 identified:1 tm:2 generally:3 colorectal:2 clear:1 amount:1 induces:1 diameter:6 percentage:3 stabilized:2 dotted:1 estimated:4 arising:1 lnm:4 four:1 deleted:2 prevent:2 diffusion:14 graph:3 cone:17 fibre:1 inverse:1 everywhere:3 micron:5 chick:4 chemically:1 family:1 almost:1 reasonable:2 scaling:1 bound:3 yielded:1 activity:2 occur:3 constraint:29 leonardo:1 bibliography:1 encodes:1 nearby:1 extracellular:1 developing:6 peripheral:1 chemosensory:1 jr:1 smaller:1 across:4 membrane:2 rev:1 lasting:1 explained:1 embryo:1 equation:3 remains:1 detectable:1 mechanism:2 needed:1 kda:4 end:1 lauffenburger:1 limb:3 away:2 appropriate:4 eichele:2 top:1 ensure:1 lavigne:9 exploit:1 restrictive:1 especially:1 erfc:3 move:1 occurs:4 concentration:32 interacts:1 gradient:38 affinity:1 conceivable:1 distance:23 sci:1 evenly:1 baier:2 considers:1 toward:1 assuming:2 bud:2 gel:2 providing:1 steep:2 upper:2 vertical:1 neuron:1 observation:1 coincident:1 thermal:1 extended:1 dc:1 varied:1 namely:1 crank:2 required:1 optimized:1 raising:1 leukocyte:5 hour:5 suggested:1 proceeds:1 below:4 goodhill:7 explanation:1 critical:1 turning:3 heffner:2 inversely:1 picture:2 axis:1 literature:1 georgetown:3 evolve:1 asymptotic:2 relative:1 embedded:1 limitation:1 geoffrey:1 jessell:1 pi:1 cd:4 production:7 embryonic:2 cancer:2 row:2 ngf:3 weaker:1 institute:1 fall:3 taking:2 absolute:1 distributed:3 benefit:1 overcome:1 calculated:2 dimension:1 curve:3 sensory:2 ignores:1 far:1 cytoplasm:3 active:2 assumed:3 shooter:2 reviewed:1 nature:5 molecule:11 robust:1 investigated:1 domain:1 eukaryotic:1 main:1 maxillary:1 linearly:1 neurosci:5 noise:3 edition:1 complementary:1 outflow:1 body:1 neuronal:1 reservoir:1 axon:22 position:2 guiding:1 formula:1 specific:1 showing:1 sensing:2 r2:1 decay:1 barrett:2 evidence:1 exists:1 gained:1 ci:3 magnitude:1 repellent:1 nat:1 easier:1 bonhoeffer:2 smoothly:1 ganglion:1 binding:3 monotonic:1 ch:1 corresponds:1 bioi:4 trigeminal:1 ann:1 towards:3 crick:2 change:7 hard:1 infinite:1 reducing:2 experimental:2 berg:2 latter:1 dorsal:1 tested:1 |
452 | 1,411 | Instabilities in Eye Movement Control: A Model
of Periodic Alternating Nystagmus
ErnstR. Dow
Center for Biophysics and
Computational Biology,
Beckman Institute
University of Illinois at UrbanaChampaign,Urbana, IL 61801.
edow@uiuc.edu
Thomas J. Anastasio
Department of Molecular and Integrative Physiology, Center for Biophysics
and Computational Biology,
Beckman Institute
University of Illinois at UrbanaChampaign, Urbana, IL 61801.
tstasio@uiuc.edu
Abstract
Nystagmus is a pattern of eye movement characterized by smooth rotations of the eye in one direction and rapid rotations in the opposite direction that reset eye position. Periodic alternating nystagmus (PAN) is
a form of uncontrollable nystagmus that has been described as an unstable but amplitude-limited oscillation. PAN has been observed previously only in subjects with vestibulo-cerebellar damage. We describe
results in which PAN can be produced in normal subjects by prolonged
rotation in darkness. We propose a new model in which the neural circuits that control eye movement are inherently unstable, but this instability is kept in check under normal circumstances by the cerebellum.
Circumstances which alter this cerebellar restraint, such as vestibulocerebellar damage or plasticity due to rotation in darkness, can lead to
PAN.
1 INTRODUCTION
Visual perception involves not only an operating visual sensory system, but also the ability to control eye movements. The oculomotor subsystems provide eye movement control. For example, the vestibulo-ocular reflex (VOR) maintains retinal image stability by
making slow-phase eye rotations that counterbalance head rotations, making it possible to
move and see at the same time (Wilson and Melvill Jones, 1979). The VOR makes slowphase eye rotations that are directed opposite to head rotations. When these ongoing
slow-phase eye rotations are interrupted by fast-phase eye rotations that reset eye position, the resulting eye movement pattern is called nystagmus. Periodic alternating nys-
A Model of Periodic Alternating Nystagmus
139
tagmus (PAN) is a congenital or acquired eye movement disorder characterized by uncontrollable nystagmus that alternates direction roughly sinusoidally with a period of 200
s to 400 s (Baloh et al., 1976; Leigh et al., 1981; Furman et aI., 1990). Furman and colleagues (1990) have determined that PAN in humans is caused by lesions of parts of the
vestibulo-cerebellum known as the nodulus and uvula (NU). Lesions to the NU cause
PAN in the dark (Waespe et aI., 1985; Angelaki and Hess, 1995). NU lesions also prevent habituation (Singleton, 1967; Waespe et aI, 1985; Torte et al., 1994), which is a
semi-permanent decrease in the gain (eye velocity / head velocity) of the VOR response
that can be brought about by prolonged low-frequency rotational stimulation in the dark.
Vestibulo-cerebellectomy in habituated goldfish causes VOR dishabituation (Dow and
Anastasio, 1996). Temporary inactivation of the vestibulo-cerebellum in habituated goldfish causes temporary dishabituation and can result in a temporary PAN (Dow and Anastasio, in press). Stimulation of the NU temporarily abolish the VOR response
(Fernandez and Fredrickson, 1964). Cerebellar influence on the VOR may be mediated
by connections between the NU and vestibular nucleus neurons, which have been demonstrated in many species (Dow, 1936; 1938).
We have previously shown that intact goldfish habituate to prolonged low-frequency
(0.01 Hz) rotation (Dow and Anastasio, 1996) and that rotation at higher frequencies
(0.05-0.1 Hz) causes PAN (Dow and Anastasio, 1997). We also proposed a limit-cycle
model of PAN in which habituation or PAN result from an increase or decrease, respectively, of the inhibition of the vestibular nuclei by the NU. This model suggested that
velocity storage, which functions to increase low-frequency VOR gain above the biophysical limits of the semicircular canals (Robinson, 1977;1981), is mediated by a potentially unstable low-frequency resonance. This instability is normally kept in check by
constant suppression by the NU.
2 METHODS
PAN was studied in intact, experimentally naive, comet goldfish (carassius auratus).
Each goldfish was restrained horizontally underwater with the head at the center of a cylindrical tank. Eye movements were measured using the magnetic search coil technique
(Robinson, 1963). For technical details see Dow and Anastasio (1996). The tank was
centered on a horizontal rotating platform. Goldfish were rotated continuously for various durations (30 min to 2 h) in darkness at various single frequencies (0.03 - 0.17 Hz).
Some data have been previously reported (Dow and Anastasio, 1997). All stimuli had
peak rotational velocities of 60 degls. Eye position and rotator (i.e. head) velocity signals
were digitiZed for analysis. Eye position data were digitally differentiated to compute eye
velocity and fast-phases were removed. Data were analyzed and simulated using MATLAB and SIMULINK (The Mathworks, Inc.).
3 RESULTS
Prolonged rotation in darkness at frequencies which produced some habituation in naive
goldfish (0.03-0.17 Hz) could produce a lower-frequency oscillation in slow-phase eye
velocity that was superimposed on the normal VOR response (fig 1). This lowerfrequency oscillation produced a periodic alternating nystagmus (PAN). When PAN occurred, it was roughly sinusoidal and varied in period, amplitude, and onset-time. Habituation could occur simultaneously with PAN (fig IB) or habituation could completely
E. R. Dow and T. J. Anastasio
140
60
A
Initial response
~ 40
-'g
~ 20
~
0
g! -20
Q)
>Q)
-40
~-60~------~--------~~--~--~------~--------~
Q)
; -:8 ~JJJvAvAJ~\VV~AvA~?~AvA~\AvAJv\Av?v?~AJfv\?M~?v?JvA~
i
.r;
60
B
Response after 1 h. rotation
~ 40
-'8
~ 20
~
Q)
>
g!.
Q)
-20
,A
iJ
0
~
~
V
~
v~
A
I?
A
v
~j
y~
l~J
v~
-40
tn-60
t
~ 50~A66A6GGGA66666AGGAAG6A666A6GA6AAAA666AGA666GA6A6
-58 VVvv rv VVIlVlJlJ VVVVVVUVVVrv VVVlJ uvrvvrvwvlv vvrVV VVVt
al
0
200
400
600
800
1000
1
'0
.r;
time (s)
Figure 1: Initial 1000 s 0.05 Hz rotation showing PAN (A). Slow-phase
eye velocity shows that PAN starts almost immediately and there is a
slight reduction in VOR gain after 1000 s. Following 1 h continuous rotation in the same goldfish (B), VOR gain has decreased.
suppress PAN (fig 4). PAN observed at lower frequencies (0.03 and 0.05 Hz) typically
decreased in amplitude as rotation continued.
Previous work has shown that PAN was most likely to occur during prolonged rotations at
frequencies between 0.05 and 0.1 Hz (Dow and Anastasio, 1997). At these frequencies,
habituation also caused a slight decrease in VOR gain (1.3 to 1.8 times, initial gain I final
gain) following 1 h of rotation. At higher frequencies, neither habituation nor PAN were
observed. At lower frequencies (0.03 Hz) PAN could occur before habituation substantially reduced VOR gain (fig 4). PAN, was not observed in naIve goldfish rotated a lower
frequency (0.01 Hz) where VOR gain fell by 22 times due to habituation (Dow and Anastasio, 1997).
A Model of Periodic Alternating Nystagmus
141
4 MODEL
Previously, a non-linear limit cycle model was constructed by Leigh, Robinson, and Zee
(1981; see also Furman, 1989) to simulate PAN in humans. This model included a velocity storage loop with saturation, and a central adaptation loop. This second order system would spontaneously oscillate, producing PAN, if the gain of the velocity storage
loop was greater than 1.
We adjusted Robinson's model to simulate rotation inducible PAN and habituation in the
goldfish. Input to and output from the model (fig 2) represent head and slow-phase eye
velocity, respectively. The time constants of the canal (S'tc/(S'tc+ 1? and velocity-storage
(g.l(S'ts+l? elements were set to the value of the canal time constant as determined experimentally in goldfish ('Cc = '1:s = 3 s) (Hartman and Klinke, 1980). The time constant of
the central adaptation element (l/S'ta) was 10 times longer ('1: a = 30 s). The Laplace variable (s) is complex frequency (s = jro where / is -1 and ro is frequency in rad/s). The gain
of the velocity-storage loop (gs) is 1.05 while that of the central adaptation loop (ga) is 1.
The central adaptation loop represents in part a negative feedback loop onto vestibular
nucleus neurons through inhibitory Purkinje cells of the NU. The vestibulo-cerebellum is
known to modulate the gain of the VOR (Wilson and Melvill Jones, 1979). The static
nonlinearity in the velocity storage loop consists of a threshold (? 0.0225) and a saturation (? 1.25). The threshold was added to model the decay in PAN following termination
of rotation (Dow and Anastasio, 1997), which is not modeled here.
Increases or decreases in the absolute value of ga will cause VOR habituation or PAN,
respectively. However, it was more common for VOR habituation and PAN to occur
simultaneously (fig tB). This behavior could not be reproduced with the lumped model
(fig 2). It would be necessary on one hand to increase ga to decrease overall VOR gain
while, on the other hand, decrease ga to produce PAN. A distributed system would address this problem, with multiple parallel pathways, each having velocity-storage and
adaptive control through the NU. The idea can
head velocity
be illustrated using the simplest distributed syscanal
tem which has 2 lumped models in parallel (not
transfer
shown), each having an independently adjustable
function
L...--'r---'
gao The results from such a two parallel pathway
model are shown in fig 3. In one pathway, ga(h)
was increased to model habituation, and gio)
was decreased to start oscillations. Paradoxically, although the ultimate effect of increasing
ga(h) is to decrease VOR gain, the initial effect
as ga(h) is increased is to increase gain. This is
due to the resonant frequency of the system concentral
tinuously shifting to higher frequencies and temadaptation
porarily matching the frequency of rotation (see
eye velocity
DISCUSSION). Conversely, when ga(o) is deFigure 2: Model used
creased, there is a temporary decrease in gain as
to simulation PAN
the resonant frequency moves away from the
(Dow and Anastasio
frequency of rotation. The two results are com1997). Used with
bined after the gain is reduced by half (fig 3B).
permission.
E. R Dow and T. 1. Anastasio
142
The combined result shows a continual decrease in VOR gain with the oscillations superimposed.
5 DISCUSSION
If the nonlinearities (i.e. threshold
and saturation) in the model are
ignored, linear analysis shows that
the model will be unstable when
[(1 - gs)/ts + gJ'tal is negative,
and will oscillate with a period of
[2m1(tstJga)]. With the initial Q)
parameters, the model is stable 8because the central adaptation ~
loop can compensate for the unstable gain of the velocity storage
loop. The natural frequency of
the system, calculated from the Q) -2L---------------~------~--------~---above equation, is 0.017 Hz. This ~
resonance, which peaks at the
al 4
:::l
. . ?.
resonant frequency but is stiII
>
.....
<
..
...
pronounced at nearby frequencies,
produces an enhancement of the
c
/ _:_
:_
: _ _ _ _ _ _ _ _ _ _........_ _
VOR response. The hypothesis
I:: 0 L_ _ _ _--=::.
that low frequency VOR gain eno
1000
2000
3000
hancement is produced by a potime (sec)
tentially unstable resonance is a
Figure 3: Model at 0.03 Hz. Two
novel feature of our model. The
simulations with differing values of
natural frequency increases with
ga (A) are combined in (B) with the
increases in ga and can alter the
values of ga in (C).
frequency specific enhancement.
Decreases in ga, in addition to decreasing the natural frequency, also cause the model to
become unstable. (If ga is reduced to zero, the model becomes first order and the equations are no longer valid). The ability to get either habituation or PAN by varying only
one parameter suggests that habituation and PAN are a related phenomena
2[---
B
il=--~
~2
k
..
Through the process of habituation, prolonged low frequency rotation (0.01 Hz) in goldfish severely decreased VOR gain, often abruptly and unilaterally (Dow and Anastasio, in
press). The decrease in gain due to habituation can effectively eliminate PAN at the lowest frequency at which PAN was observed (0.03 Hz) as shown in fig 4. In this example
the naive VOR responds symmetrically for the first cycle of rotation. It then becomes
markedly asymmetrical, with a strong, unilateral response in one direction for -10 cycles
followed by another in the opposite direction for -17 cycles. The VOR response abruptly
habituates after that with no PAN. Complete habituation can be simulated by further increases in the value of ga' in the limit-cycle model (fig 2). Unilateral habituation has been
simulated previously with a bilateral network model of the VOR in which the cerebellum
inhibits the vestibular nuclei unilaterally (Dow and Anastasio, in press).
A Model of Periodic Alternating Nystagmus
143
60
40
~
Q)
---
20
"0
~
?0
o
0
.l
bJ
~
A ~tLL
r~ I'~ "'1
CD
>
~-20
Q)
-40
o
200
400
600
time (5)
800
Figure 4: PAN superimposed on the VOR response to continuous rotation
at 0.03 Hz. Upper trace, slow-phase eye velocity (fast phases removed);
lower trace, head velocity (not to scale).
The cerebellum has several circuits which could provide an increase in firing rate of some
Purkinje cells with a concurrent decrease in the firing rate of other Purkinje cells suggested by the model. There are many lateral inhibitory pathways including inhibition of
Purkinje cells to neighboring Purkinje cells (Llimis and Walton, 1990). Therefore, if one
Purkinje cell were to increase its firing rate, this circuitry suggests that neighboring
Purkinje cells would decrease their firing rates. Also, experimental evidence shows that
during habituation, not all vestibular nuclei gradually decrease their firing rate as might be
expected. Kileny and colleagues (1980) recorded from vestibular nucleus neurons during
habituation. They could divide the neurons into 3 roughly equal groups based on response over time: continual decrease, constant followed by a decrease, and increase followed by decrease. The cerebellar circuitry and the single-unit recording data support
multiple, variable levels of inhibition from the NU. How this mechanism may work is
being explored with a more biologically realistic distributed model.
6 CONCLUSION
Our experimental results are consistent with a multi-parallel pathway model of the VOR.
In each pathway an unstable, positive velocity-storage loop is stabilized by an inhibitory,
central adaptation loop, and their interaction produces a low-frequency resonance that
enhances the low-frequency response of the VOR. Prolonged rotation at specific frequencies could produce a decrease in central adaptation VOR gain in some pathways resulting in an unstable, low-frequency oscillation resembling PAN in these pathways. An
144
E. R. Dow and T. J. Anastasio
increase in adaptation loop gain in the other pathways would result in a decrease in VOR
gain resembling habituation. The sum over the VOR pathways would show PAN and
habituation occurring together. We suggest that resonance enhancement and mUltiple
parallel (i.e. distributed) pathways are necessary to model the interrelationship between
PAN and habituation.
Acknowledgments
The work was supported by grant MH50577 from the National Institutes of Health. We
thank M. Zelaya and X. Feng for experimental assistance.
References
Angelaki DE and Hess BJM. J Neurophysl73 1729-1751 (1995).
Baloh RW, Honrubia V and Konrad HR. Brain 99 11-26 (1976).
Dow ER and Anastasio TJ. NeuroReport 71305-1309 (1996).
Dow ER and Anastasio TJ. NeuroReport 82755-2759 (1997).
Dow ER and Anastasio TJ. J. Computat. Neuro. in press.
Dow RS. J Comp Neurol63 527-548 (1936).
Dow RS. J Comp Neurol68 297-305 (1938).
Fernandez C and Fredrickson lM. Acta Otolaryngol Suppl192 52-62 (1964).
Furman JMR, Wall C and Pang D. Brain 113 1425-1439 (1990).
Furman lMR, Hain TC and Paige GO. Biol Cybern 61255-264 (1989).
Hartmann R and Klinke R. Pflugers Archiv 388 111-121 (1980).
Kileny P, Ryu JH, McCabe BF and Abbas PJ. Acta Otolaryngol90 175-183 (1980).
Leigh RJ, Robinson DA and Zee OS. Ann NY A cad Sci 374619-635 (1981).
LIinas RR and Walton KD. Cerebellum. In: Shepherd GM ed. The Synaptic Organization o/the Brain. Oxford: Oxford University Press, 1990: 214-245.
Remmel RS. IEEE Trans Biomed Eng 31 388-390 (1984).
Robinson DA. IEEE Trans Biomed Eng 10 137-145 (1963).
Robinson DA. Exp Brain Res 30447-450 (1977).
Robinson DA. Ann Rev Neurosci 4463-503 (1981).
Singleton GT. Laryngoscope 77 1579-1620 (1967).
Torte MP, Courjon JH, Flandrin JM, et al. Exp Brain Res 99 441-454 (1994).
Waespe W, Cohen Band Raphan T. Science 228 199-202 (1985).
Wilson V and Melvill Jones G. Mammalian Vestibular Physiology, New York: Plenum
Press, 1979.
| 1411 |@word cylindrical:1 bf:1 termination:1 integrative:1 simulation:2 r:3 eng:2 reduction:1 initial:5 cad:1 vor:31 interrupted:1 realistic:1 plasticity:1 half:1 gio:1 constructed:1 become:1 consists:1 pathway:11 acquired:1 expected:1 rapid:1 behavior:1 roughly:3 nor:1 uiuc:2 multi:1 brain:5 decreasing:1 prolonged:7 jm:1 increasing:1 becomes:2 circuit:2 mccabe:1 lowest:1 substantially:1 differing:1 continual:2 eno:1 ro:1 control:5 normally:1 unit:1 grant:1 producing:1 before:1 positive:1 limit:4 severely:1 oxford:2 firing:5 might:1 acta:2 studied:1 ava:2 conversely:1 suggests:2 limited:1 directed:1 acknowledgment:1 spontaneously:1 physiology:2 matching:1 suggest:1 get:1 onto:1 ga:14 subsystem:1 storage:9 influence:1 instability:3 cybern:1 darkness:4 demonstrated:1 center:3 resembling:2 go:1 duration:1 independently:1 disorder:1 immediately:1 continued:1 unilaterally:2 stability:1 underwater:1 laplace:1 plenum:1 gm:1 hypothesis:1 velocity:21 element:2 klinke:2 mammalian:1 observed:5 cycle:6 movement:8 decrease:19 removed:2 digitally:1 completely:1 various:2 fast:3 describe:1 ability:2 hartman:1 final:1 reproduced:1 rr:1 biophysical:1 propose:1 interaction:1 reset:2 adaptation:8 neighboring:2 loop:13 pronounced:1 enhancement:3 walton:2 produce:5 rotated:2 urbanachampaign:2 measured:1 ij:1 strong:1 involves:1 direction:5 centered:1 human:2 uncontrollable:2 wall:1 adjusted:1 normal:3 exp:2 bj:1 lm:1 circuitry:2 beckman:2 concurrent:1 brought:1 inactivation:1 varying:1 wilson:3 tll:1 check:2 superimposed:3 suppression:1 typically:1 eliminate:1 biomed:2 tank:2 overall:1 hartmann:1 resonance:5 platform:1 equal:1 bined:1 having:2 biology:2 represents:1 jones:3 alter:2 tem:1 leigh:3 stimulus:1 simultaneously:2 national:1 phase:9 restraint:1 organization:1 analyzed:1 tj:3 zee:2 necessary:2 divide:1 rotating:1 re:2 increased:2 sinusoidally:1 purkinje:7 reported:1 periodic:7 combined:2 peak:2 together:1 continuously:1 central:7 recorded:1 sinusoidal:1 singleton:2 nonlinearities:1 retinal:1 de:1 sec:1 inc:1 permanent:1 caused:2 fernandez:2 onset:1 mp:1 bilateral:1 furman:5 start:2 maintains:1 parallel:5 nystagmus:10 il:3 abolish:1 pang:1 flandrin:1 produced:4 comp:2 cc:1 ed:1 synaptic:1 colleague:2 frequency:34 ocular:1 static:1 gain:25 amplitude:3 higher:3 ta:1 melvill:3 response:11 hand:2 dow:22 horizontal:1 o:1 effect:2 asymmetrical:1 alternating:7 restrained:1 illustrated:1 cerebellum:7 assistance:1 during:3 lumped:2 konrad:1 complete:1 tn:1 interrelationship:1 image:1 novel:1 common:1 rotation:27 stimulation:2 cohen:1 occurred:1 slight:2 m1:1 ai:3 hess:2 illinois:2 nonlinearity:1 had:1 stable:1 longer:2 operating:1 inhibition:3 gj:1 gt:1 llimis:1 greater:1 period:3 signal:1 semi:1 rv:1 multiple:3 rj:1 smooth:1 technical:1 characterized:2 compensate:1 molecular:1 biophysics:2 neuro:1 circumstance:2 cerebellar:4 represent:1 abbas:1 cell:7 addition:1 decreased:4 fell:1 markedly:1 subject:2 hz:14 shepherd:1 recording:1 habituation:24 symmetrically:1 paradoxically:1 opposite:3 honrubia:1 idea:1 ultimate:1 abruptly:2 unilateral:2 paige:1 york:1 cause:6 oscillate:2 matlab:1 ignored:1 dark:2 band:1 raphan:1 simplest:1 reduced:3 rw:1 computat:1 inhibitory:3 stabilized:1 canal:3 group:1 threshold:3 comet:1 prevent:1 neither:1 pj:1 kept:2 sum:1 almost:1 resonant:3 oscillation:6 followed:3 g:2 occur:4 tal:1 nearby:1 archiv:1 simulate:2 min:1 inhibits:1 department:1 alternate:1 kd:1 pan:40 rev:1 making:2 biologically:1 gradually:1 equation:2 previously:5 mechanism:1 mathworks:1 bjm:1 laryngoscope:1 away:1 differentiated:1 magnetic:1 permission:1 thomas:1 hain:1 feng:1 move:2 added:1 damage:2 responds:1 enhances:1 thank:1 simulated:3 lateral:1 sci:1 unstable:9 modeled:1 rotational:2 potentially:1 trace:2 negative:2 suppress:1 adjustable:1 goldfish:12 upper:1 av:1 neuron:4 urbana:2 semicircular:1 t:2 head:8 digitized:1 varied:1 connection:1 rad:1 temporary:4 ryu:1 nu:10 vestibular:7 robinson:8 address:1 trans:2 suggested:2 pattern:2 perception:1 oculomotor:1 saturation:3 tb:1 including:1 shifting:1 natural:3 hr:1 counterbalance:1 eye:24 mediated:2 naive:4 health:1 nucleus:6 consistent:1 vestibulo:6 cd:1 supported:1 l_:1 jh:2 vv:1 institute:3 absolute:1 distributed:4 feedback:1 calculated:1 valid:1 sensory:1 adaptive:1 search:1 continuous:2 habituates:1 transfer:1 inherently:1 complex:1 da:4 neurosci:1 lesion:3 angelaki:2 rotator:1 fig:11 slow:6 ny:2 position:4 ib:1 congenital:1 anastasio:19 specific:2 showing:1 habituated:2 er:3 explored:1 decay:1 otolaryngol:1 evidence:1 effectively:1 occurring:1 tc:3 likely:1 gao:1 visual:2 horizontally:1 temporarily:1 reflex:1 coil:1 modulate:1 ann:2 experimentally:2 included:1 determined:2 called:1 specie:1 experimental:3 intact:2 support:1 phenomenon:1 ongoing:1 neuroreport:2 biol:1 |
453 | 1,412 | Boltzmann Machine learning using mean
field theory and linear response correction
H.J. Kappen
Department of Biophysics
University of Nijmegen, Geert Grooteplein 21
NL 6525 EZ Nijmegen, The Netherlands
F. B. Rodriguez
Instituto de Ingenieria del Conocimiento & Departamento de Ingenieria Informatica.
Universidad Aut6noma de Madrid, Canto Blanco,28049 Madrid, Spain
Abstract
We present a new approximate learning algorithm for Boltzmann
Machines, using a systematic expansion of the Gibbs free energy to
second order in the weights. The linear response correction to the
correlations is given by the Hessian of the Gibbs free energy. The
computational complexity of the algorithm is cubic in the number
of neurons. We compare the performance of the exact BM learning
algorithm with first order (Weiss) mean field theory and second
order (TAP) mean field theory. The learning task consists of a fully
connected Ising spin glass model on 10 neurons. We conclude that
1) the method works well for paramagnetic problems 2) the TAP
correction gives a significant improvement over the Weiss mean field
theory, both for paramagnetic and spin glass problems and 3) that
the inclusion of diagonal weights improves the Weiss approximation
for paramagnetic problems , but not for spin glass problems.
1
Introduction
Boltzmann Machines (BMs) [1], are networks of binary neurons with a stochastic
neuron dynamics, known as Glauber dynamics. Assuming symmetric connections
between neurons, the probability distribution over neuron states s will become
stationary and is given by the Boltzmann-Gibbs distribution P( S). The Boltzmann
distribution is a known function of the weights and thresholds of the network.
However, computation of P(Sj or any statistics involving P(S), such as mean firing
rates or correlations, requires exponential time in the number of neurons. This is
Boltzmann Machine Learning Using Mean Field Theory
281
due to the fact that P(S) contains a normalization term Z, which involves a sum
over all states in the network, of which there are exponentially many. This problem
is particularly important for BM learning.
Using statistical sampling techiques [2], learning can be significantly improved [1].
However, the method has rather poor convergence and can only be applied to small
networks.
In [3 , 4], an acceleration method for learning in BMs is proposed using mean field
theory by replacing (SjSj) by mimj in the learning rule. It can be shown [5] that
such a naive mean field approximation of the learning rules does not converge in
general. Furthermore, we argue that the correlations can be computed using the
linear response theorem [6].
In [7, 5] the mean field approximation is derived by making use of the properties
of convex functions (Jensen's inequality and tangential bounds). In this paper we
present an alternative derivation which uses a Legendre transformation and a small
coupling expansion [8] . It has the advantage that higher order contributions (TAP
and higher) can be computed in a systematic manner and that it may be applicable
to arbitrary graphical models.
2
Boltzmann Machine learning
The Boltzmann Machine is defined as follows. The possible configurations of the
network can be characterized by a vector s = (S1, .. , Si, .. , sn), where Si = ?1 is the
state of the neuron i, and n the total number of the neurons. Neurons are updated
using Glauber dynamics.
Let us define the energy of a configuration
-E(S) =
s as
1
2L
WijSiSj
+
i,j
L
SiOi.
i
After long times, the probability to find the network in a state s becomes independent of time (thermal equilibrium) and is given by the Boltzmann distribution
1
p(S) = Z exp{ -E(S)}.
(1)
Z = L; exp{ - E( S)} is the partition function which normalizes the probability
distribution.
Learning [1] consists of adjusting the weights and thresholds in such a way that
the Boltzmann distribution approximates a target distribution q(S) as closely as
possible.
A suitable measure of the difference between the distributions p(S) and q(S) is the
K ullback divergence [9]
=
J{
~ q(S) log ;~~.
(2)
s
Learning consists of minimizing
~Wij
J{
= 17( (SiSj)c -
using gradient descent [1]
(SiSj) ),
~()i
= 17( (Si)c -
(Si) ).
The parameter 17 is the learning rate. The brackets (-) and (-) c denote the 'free' and
'clamped' expectation values, respectively.
H. 1. Kappen and F. B. Rodr(guez
282
The computation of both the free and the clamped expectation values is intractible,
because it consists of a sum over all unclamped states. As a result, the BM learning
algorithm can not be applied to practical problems.
3
The mean field approximation
We derive the mean field free energy using the small, expansion as introduced by
Plefka [8]. The energy of the network is given by
E(s, w, h, ,)
for, = 1. The free energy is given by
F(w, (), ,) = -logTr s e- E (s ,w,8,),)
and is a function of the independent variables Wij, ()j and ,. We perform a Legendre
transformation on the variables (}i by introducing mj = - ~:.. The Gibbs free energy
G(w, m, ,)
= F(w, (), ,) + L
()jmj
is now a function of the independent variables mj and Wij, and ()i is implicitly
given by (Si) )' = mi. The expectation 0" is with respect to the full model with
interaction , .
We expand
= G(O) + ,G' (0) + ~,2G"(O) + 0(,3)
G(,)
We directly obtain from [8]
G'(,)
(Eint)"
GUb)
(E;n')~ - {Ei:,,), + (E;n,;;= ~ (s, -
For, = 0 the expectation values
can directly compute:
G(O)
0 )'
m;)),
become the mean field expectations which we
1
21 ~ (
(1 + mi)
log 2(1 + mi) + (1 -
)
md log 1
2(1- md
J
G'(O)
-~ L
Wjjmjmj
ij
G"(O)
Thus
G(l)
1
-L
2 .
(
1
1md
)
log-(l + md + (1- mdlog-(l(1 + md
2
I
"'w .
-~2 L-
IJ
2
m I?m?J
ij
-~ L
ij
w;j(l - m;)(l-
m]) + 0(w 3 f(m))
(3)
Boltvnann Machine Learning Using Mean Field Theory
283
where f(m) is some unknown function of m.
The mean field equations are given by the inverse Legendre transformation
e. - ami
ae _
- tan h - 1 ( mi ) 1 -
""'
L- Wijmj
2
1+ ""'
~ Wjjmd
J
m j2 ),
(4)
J
which we recognize as the mean field equations.
The correlations are given by
a 2F
(SiSj) - (Si) (Sj)
ami
= - oeiooj = oej =
( ao ) -1
am ij
( 02~) -1
am
ij
We therefore obtain from Eq. 3
(Si Sj ) -
with
(A-')oj = Jij
(
1 _1 ml
(Si)
(s j) =
Aij
+ ~ wi.(1 - ml )) -
Wij - 2mimjW;j
(5)
Thus, for given Wij and OJ, we obtain the approximate mean firing rates mj by solving Eqs . 4 and the correlations by their linear response approximations Eqs. 5. The
inclusion of hidden units is straigthforward . One applies the above approximations
in the free and the clamped phase separately [5]. The complexity of the method is
O(n 3 ), due to the matrix inversion.
4
Learning without hidden units
We will assess the accuracy of the above method for networks without hidden units.
Let us define Cij = (SjSj)c - (Si)c (Sj)c' which can be directly computed from the
data. The fixed point equation for D..Oj gives
D..Oi = 0 {:} mj = (Si)c .
(6)
The fixed point equation for D..wij, using Eq. 6, gives
D..wij
0 {:} Aij Cij ' i =F j.
(7)
From Eq. 7 and Eq. 5 we can solve for Wij, using a standard least squares method.
In our case, we used fsolve from Matlab. Subsequently, we obtain ei from Eq. 4.
We refer to this method as the TAP approximation.
=
=
In order to assess the effect of the TAP term, we also computed the weights and
thresholds in the same way as described above, but without the terms of order w 2
in Eqs. 5 and 4. Since this is the standard Weiss mean field expression, we refer to
this method as the Weiss approximation.
The fixed point equations are only imposed for the off-diagonal elements of D..Wjj
because the Boltzmann distribution Eq. 1 does not depend on the diagonal elements
Wij. In [5], we explored a variant of the Weiss approximation, where we included
diagonal weight terms . As is discussed there, if we were to impose Eq. 7 for i = j
as well, we have A = C. HC is invertible, we therefore have A-I = C- 1 . However,
we now have more constraints than variables. Therefore, we introduce diagonal
weights Wii by adding the term Wiimi to the righthandside of Eq. 4 in the Weiss
approximation. Thus,
Wij
,sij
= 1 _ m?
_
(C- 1 ) ..
lJ
I
and OJ is given by Eq. 4 in the Weiss approximation. Clearly, this method is computationally simpler because it gives an explicit expression for the solution of the
weights involving only one matrix inversion.
284
5
H. 1. Kappen and F. B. Rodr(guez
Numerical results
For the target distribution q(s) in Eq. 2 we chose a fully connected Ising spin glass
model with equilibrium distribution
with lij i.i.d. Gaussian variables with mean n~l and variance /~1 ' This model
is known as the Sherrington-Kirkpatrick (SK) model [10]. Depending on the values of 1 and 1 0 , the model displays a para-magnetic (unordered), ferro-magnetic
(ordered) and a spin-glass (frustrated) phase. For 10 = 0, the para-magnetic (spinglass) phase is obtained for 1 < 1 (1 > 1). We will assess the effectiveness of our
approximations for finite n, for 10 = 0 and for various values of 1. Since this is a
realizable task, the optimal KL divergence is zero, which is indeed observed in our
simulations.
We measure the quality of the solutions by means ofthe Kullback divergence. Therefore, this comparison is only feasible for small networks . The reason is that the
computation of the Kullback divergence requires the computation of the Boltzmann
distribution, Eq. 1, which requires exponential time due to the partition function
Z.
We present results for a network of n = 10 neurons . For 10 = 0, we generated
for each value of 0.1 < 1 < 3, 10 random weight matrices l i j. For each weight
matrix, we computed the q(S) on all 2n states. For each of the 10 problems, we
applied the TAP method, the Weiss method and the Weiss method with diagonal
weights. In addition, we applied the exact Boltzmann Machine learning algorithm
using conjugate gradient descent and verified that it gives KL diver?ence equal
to zero, as it should. We also applied a factorized model p(S) = Ili ?"(1 + misd
with mi = (Si)c to assess the importance of correlations in the target distribution.
In Fig. la, we show for each 1 the average KL divergence over the 10 problem
instances as a function of 1 for the TAP method, the Weiss method, the Weiss
method with diagonal weights and the factorized model. We observe that the TAP
method gives the best results, but that its performance deteriorates in the spin-glass
phase (1) 1) .
The behaviour of all approximate methods is highly dependent on the individual
problem instance. In Fig. Ib, we show the mean value of the KL divergence of the
TAP solution, together with the minimum and maximum values obtained on the
10 problem instances.
Despite these large fluctuations , the quality of the TAP solution is consistently
better than the Weiss solution. In Fig. lc, we plot the difference between the TAP
and Weiss solution, averaged over the 10 problem instances.
In [5] we concluded that the Weiss solution with diagonal weights is better than
the standard Weiss solution when learning a finite number of randomly generated
patterns. In Fig. Id we plot the difference between the Weiss solution with and
without diagonal weights. We observe again that the inclusion of diagonal weights
leads to better results in the paramagnetic phase (1 < 1), but leads to worse results
in the spin-glass phase. For 1 > 2, we encountered problem instances for which
either the matrix C is not invertible or the KL divergence is infinite. This problem
becomes more and more severe for increasing 1 . We therefore have not presented
results for the Weiss approximation with diagonal weigths for 1 > 2.
BoltvnannMachine Learning Using Mean Field Theory
285
Comparison mean values
TAP
5r-------~------~------.
4
2
fact
weiss+d
weiss
tap
Q)
1.5
u
c:
mean
min
max
Q)
e>
Q)
1
I
>
/
",
.... '"
1_'"
'5
....J
:.::
--
0.5
J..:y '",
o~
o
.... ; ./
__~...=?~J~______~____~
2
o~-=~~------~----~
o
3
2
J
3
J
Difference WEISS and TAP
Difference WEISS+D and WEISS
1.5
en
':!?
n.
.1 0.5
w
J
:.::
I
:.::
I
en
en
J
V
0.5
J
:.::
~
_ _ _ _~~_ _ _ _----.J
o
2
J
3
0
-0.5
v
/
W
0
-0.5 L-._ _ _ _ _ _
'?
en
en
W
:.::
1
~
1
o
2
3
J
Figure 1: Mean field learning of paramagnetic (J < 1) and spin glass (J > 1)
problems for a network of 10 neurons. a) Comparison of mean KL divergences for
the factorized model (fact), the Weiss mean field approximation with and without
diagonal weights (weiss+d and weiss), and the TAP approximation, as a function
of J. The exact method yields zero KL divergence for all J. b) The mean, minimum and maximum KL divergence of the TAP approximation for the 10 problem
instances , as a function of J. c) The mean difference between the KL divergence
for the Weiss approximation and the TAP approximation, as a function of J. d)
The mean difference between the KL divergence for the Weiss approximation with
and without diagonal weights, as a function of J .
6
Discussion
We have presented a derivation of mean field theory and the linear response correction based on a small coupling expansion of the Gibbs free energy. This expansion
can in principle be computed to arbitrary order. However , one should expect that
the solution of the resulting mean field and linear response equations will become
more and more difficult to solve numerically. The small coupling expansion should
be applicable to other network models such as the sigmoid belief network , Potts
networks and higher order Boltzmann Machines .
The numerical results show that the method is applicable to paramagnetic problems.
This is intuitively clear, since paramagnetic problems have a unimodal probability
distribution, which can be approximated by a mean and correlations around the
mean. The method performs worse for spin glass problems. However , it still gives
a useful approximation of the correlations when compared to the factorized model
which ignores all correlations. In this regime, the TAP approximation improves
286
H. 1. Kappen and F. B. Rodr(guez
significantly on the Weiss approximation. One may therefore hope that higher order
approximation may further improve the method for spin glass problems. Therefore.
we cannot conclude at this point whether mean field methods are restricted to
unimodal distributions. In order to further investigate this issue, one should also
study the ferromagnetic case (Jo > 1, J > 1), which is multimodal as well but less
challenging than the spin glass case.
It is interesting to note that the performance of the exact method is absolutely
insensitive to the value of J. Naively, one might have thought that for highly
multi-modal target distributions, any gradient based learning method will suffer
from local minima. Apparently, this is not the case: the exact KL divergence has
just one minimum, but the mean field approximations of the gradients may have
multiple solutions.
Acknowledgement
This research is supported by the Technology Foundation STW, applied science
division of NWO and the techology programme of the Ministry of Economic Affairs.
References
[1] D. Ackley, G. Hinton, and T. Sejnowski. A learning algorithm for Boltzmann Machines. Cognitive Science, 9:147-169, 1985.
[2] C. Itzykson and J-M. Drouffe. Statistical Field Theory. Cambridge monographs on
mathematical physics. Cambridge University Press. Cambridge, UK, 1989.
[3] C. Peterson and J.R. Anderson. A mean field theory learning algorithm for neural
networks. Complex Systems, 1:995-1019, 1987.
[4] G.E. Hinton. Deterministic Boltzmann learning performs steepest descent in weightspace. Neural Computation, 1:143-150. 1989.
[5] H.J. Kappen and F.B. Rodriguez. Efficient learning in Boltzmann Machines using
linear response theory. Neural Computation, 1997. In press.
[6] G. Parisi. Statistical Field Theory. Frontiers in Physics. Addison-Wesley, 1988.
[7] L.K. Saul, T. Jaakkola, and M.1. Jordan. Mean field theory for sigmoid belief networks. Journal of artificial intelligence research, 4:61-76, 1996.
[8] T. Plefka. Convergence condition of the TAP equation for the infinite-range Ising
spin glass model. Journal of Physics A, 15:1971-1978, 1982.
[9] S. Kullback. Information Theory and Statistics. Wiley, New York, 1959.
[10] D. Sherrington and S. Kirkpatrick. Solvable model of Spin-Glass.
letters, 35:1792-1796, 1975.
Physical review
| 1412 |@word inversion:2 grooteplein:1 simulation:1 kappen:5 configuration:2 contains:1 paramagnetic:7 si:11 guez:3 numerical:2 partition:2 plot:2 stationary:1 intelligence:1 affair:1 steepest:1 simpler:1 mathematical:1 become:3 consists:4 introduce:1 manner:1 indeed:1 weightspace:1 multi:1 increasing:1 becomes:2 spain:1 factorized:4 transformation:3 uk:1 unit:3 local:1 instituto:1 despite:1 id:1 firing:2 fluctuation:1 might:1 chose:1 challenging:1 range:1 averaged:1 practical:1 significantly:2 thought:1 cannot:1 imposed:1 deterministic:1 convex:1 ferro:1 rule:2 geert:1 updated:1 target:4 tan:1 exact:5 us:1 element:2 approximated:1 particularly:1 ising:3 observed:1 ackley:1 ferromagnetic:1 connected:2 monograph:1 complexity:2 wjj:1 dynamic:3 depend:1 solving:1 division:1 multimodal:1 various:1 derivation:2 sejnowski:1 artificial:1 solve:2 statistic:2 advantage:1 parisi:1 interaction:1 jij:1 j2:1 wijsisj:1 canto:1 blanco:1 convergence:2 coupling:3 derive:1 depending:1 ij:6 eq:14 involves:1 closely:1 stochastic:1 subsequently:1 behaviour:1 ao:1 frontier:1 correction:4 around:1 exp:2 equilibrium:2 applicable:3 techology:1 nwo:1 hope:1 clearly:1 gaussian:1 rather:1 sisj:3 jaakkola:1 derived:1 unclamped:1 improvement:1 consistently:1 potts:1 am:2 glass:13 realizable:1 dependent:1 lj:1 hidden:3 expand:1 wij:10 rodr:3 issue:1 field:26 equal:1 sampling:1 tangential:1 randomly:1 divergence:13 recognize:1 individual:1 phase:6 highly:2 investigate:1 severe:1 kirkpatrick:2 bracket:1 nl:1 instance:6 ence:1 introducing:1 plefka:2 para:2 systematic:2 universidad:1 off:1 physic:3 invertible:2 together:1 jo:1 again:1 worse:2 cognitive:1 de:3 unordered:1 apparently:1 contribution:1 ass:4 square:1 oi:1 spin:13 intractible:1 accuracy:1 variance:1 yield:1 ofthe:1 energy:8 mi:5 adjusting:1 improves:2 wesley:1 higher:4 response:7 wei:29 improved:1 modal:1 anderson:1 furthermore:1 just:1 correlation:9 replacing:1 ei:2 rodriguez:2 del:1 quality:2 effect:1 symmetric:1 glauber:2 sherrington:2 performs:2 sigmoid:2 physical:1 exponentially:1 insensitive:1 discussed:1 approximates:1 numerically:1 significant:1 refer:2 cambridge:3 gibbs:5 inclusion:3 inequality:1 binary:1 minimum:4 ministry:1 impose:1 converge:1 full:1 unimodal:2 multiple:1 characterized:1 long:1 biophysics:1 involving:2 variant:1 ae:1 expectation:5 normalization:1 addition:1 separately:1 concluded:1 effectiveness:1 jordan:1 economic:1 whether:1 expression:2 itzykson:1 suffer:1 hessian:1 york:1 matlab:1 useful:1 clear:1 weigths:1 netherlands:1 informatica:1 deteriorates:1 threshold:3 verified:1 sum:2 inverse:1 letter:1 ullback:1 bound:1 display:1 encountered:1 constraint:1 mimj:1 min:1 department:1 poor:1 legendre:3 conjugate:1 wi:1 making:1 s1:1 intuitively:1 restricted:1 sij:1 sjsj:2 computationally:1 equation:7 addison:1 wii:1 observe:2 magnetic:3 alternative:1 graphical:1 md:5 diagonal:13 gradient:4 argue:1 reason:1 assuming:1 minimizing:1 difficult:1 cij:2 nijmegen:2 stw:1 boltzmann:17 unknown:1 perform:1 neuron:12 finite:2 descent:3 thermal:1 hinton:2 arbitrary:2 introduced:1 kl:11 connection:1 tap:19 pattern:1 regime:1 oj:4 max:1 belief:2 suitable:1 solvable:1 improve:1 technology:1 naive:1 lij:1 sn:1 review:1 acknowledgement:1 fully:2 expect:1 interesting:1 foundation:1 principle:1 normalizes:1 supported:1 free:9 aij:2 saul:1 peterson:1 diver:1 ignores:1 bm:5 programme:1 sj:4 approximate:3 implicitly:1 kullback:3 ml:2 conclude:2 sk:1 mj:4 expansion:6 hc:1 complex:1 fig:4 madrid:2 en:5 cubic:1 wiley:1 lc:1 explicit:1 exponential:2 clamped:3 ib:1 theorem:1 departamento:1 jensen:1 explored:1 naively:1 adding:1 importance:1 jmj:1 ez:1 ordered:1 applies:1 frustrated:1 acceleration:1 feasible:1 included:1 infinite:2 ami:2 total:1 ili:1 la:1 absolutely:1 |
454 | 1,413 | Selecting weighting factors in logarithmic
opinion pools
Tom Heskes
Foundation for Neural Networks, University of Nijmegen
Geert Grooteplein 21, 6525 EZ Nijmegen, The Netherlands
tom@mbfys.kun.nl
Abstract
A simple linear averaging of the outputs of several networks as
e.g. in bagging [3], seems to follow naturally from a bias/variance
decomposition of the sum-squared error. The sum-squared error of
the average model is a quadratic function of the weighting factors
assigned to the networks in the ensemble [7], suggesting a quadratic
programming algorithm for finding the "optimal" weighting factors.
If we interpret the output of a network as a probability statement,
the sum-squared error corresponds to minus the loglikelihood or
the Kullback-Leibler divergence, and linear averaging of the outputs to logarithmic averaging of the probability statements: the
logarithmic opinion pool.
The crux of this paper is that this whole story about model averaging, bias/variance decompositions, and quadratic programming
to find the optimal weighting factors, is not specific for the sumsquared error, but applies to the combination of probability statements of any kind in a logarithmic opinion pool, as long as the
Kullback-Leibler divergence plays the role of the error measure. As
examples we treat model averaging for classification models under
a cross-entropy error measure and models for estimating variances.
1
INTRODUCTION
In many simulation studies it has been shown that combining the outputs of several
trained neural networks yields better results than relying on a single model. For
regression problems, the most obvious combination seems to be a simple linear
Selecting Weighting Factors in Logarithmic Opinion Pools
267
averaging of the network outputs. From a bias/variance decomposition of the sumsquared error it follows that the error of the so obtained average model is always
smaller or equal than the average error of the individual models. In [7] simple linear
averaging is generalized to weighted linear averaging, with different weighting factors
for the different networks in the ensemble . A slightly more involved bias/variance
decomposition suggests a rather straightforward procedure for finding "optimal"
weighting factors.
Minimizing the sum-squared error is equivalent to maximizing the loglikelihood of
the training data under the assumption that a network output can be interpreted
as an estimate of the mean of a Gaussian distribution with fixed variance. In
these probabilistic terms, a linear averaging of network outputs corresponds to a
logarithmic rather than linear averaging of probability statements.
In this paper, we generalize the regression case to the combination of probability
statements of any kind. Using the Kullback-Leibler divergence as the error measure, we naturally arrive at the so-called logarithmic opinion pool. A bias/variance
decomposition similar to the one for sum-squared error then leads to an objective
method for selecting weighting factors.
Selecting weighting factors in any combination of probability statements is known
to be a difficult problem for which several suggestions have been made. These
suggestions range from rather involved supra-Bayesian methods to simple heuristics
(see e.g. [1, 6] and references therein). The method that follows from our analysis
is probably somewhere in the middle: easier to compute than the supra-Bayesian
methods and more elegant than simple heuristics.
To stress the generality of our results, the presentation in the next section will be
rather formal. Some examples will be given in Section 3. Section 4 discusses how
the theory can be transformed into a practical procedure.
2
LOGARITHMIC OPINION POOLS
Let us consider the general problem of building a probability model of a variable y
given a particular input x. The "output" y may be continuous, as for example in
regression analysis, or discrete, as for example in classification. In the latter case
integrals over y should be replaced by summations over all possible values of y.
Both x and y may be vectors of several elements; the one-dimensional notation is
chosen for convenience. We suppose that there is a "true" conditional probability
model q(ylx) and have a whole ensemble (also called pool or committee) of experts,
each supplying a probability model Pa(Ylx). p{x) is the unconditional probability
distribution of inputs. An unsupervised scenario, as for example treated in [8], is
obtained if we simply neglect the inputs x or consider them constant.
We define the distance between the true probability q(ylx) and an estimate p(ylx)
to be the Kullback-Leibler divergence
,
J\. (q, p)
== -
J J
dx p(x)
dy q(ylx)log [p(Y1x)]
q(ylx)
If the densities p(x) and q(ylx) correspond to a data set containing a finite number
p of combinations {xlJ , ylJ}, minus the Kullback divergence is, up to an irrelevant
T. Heskes
268
constant, equivalent to the loglikelihood defined as
L(p, {i, Y}) ==
~ L logp(y~ Ix~) .
~
The more formal use of the Kullback-Leibler divergence instead of the loglikelihood
is convenient in the derivations that follow.
Weighting factors Wa are introduced to indicate the reliability of each ofthe experts
fr. In the following we will work with the constraints La Wa = 1, which is used in
some of the proofs, and Wa ? 0 for all experts fr, which is not strictly necessary, but
makes it easier to interpret the weighting factors and helps to prevent overfitting
when weighting factors are optimized (see details below).
We define the average model j)(y/x) to be the one that is closest to the given set of
models:
j)(y/x) == argmin
waJ{(P,Pa) .
p(ylx)
L
a
Introducing a Lagrange mUltiplier for the constraint
find the solution
J dxp(y/x) = 1, we immediately
(1)
with normalization constant
Z(x) =
J
dy II[Pa(Y/X)]W
Q
?
(2)
a
This is the logarithmic opinion pool, to be contrasted with the linear opinion pool,
which is a linear average'ofthe probabilities. In fact, logarithmic opinion pools have
been proposed to overcome some of the weaknesses of the linear opinion pool. For
example, the logarithmic opinion pool is "externally Bayesian" , i.e., can be derived
from joint probabilities using Bayes' rule [2]. A drawback of the logarithmic opinion
pool is that if any of the experts assigns probability zero to a particular outcome,
the complete pool assigns probability zero, no matter what the other experts claim.
This property of the logarithmic opinion pool, however, is only a drawback if the
individual density functions are not carefully estimated. The main problem for both
linear and logarithmic opinion pools is how to choose the weighting factors Wa.
The Kullback-Leibler divergence of the opinion pool p(y/x) can be decomposed into
a term containing the Kullback-Leibler divergences of individual models and an
"ambiguity" term:
(3)
Proof: The first term in (3) follows immediately from the numerator in (1), the
second term is minus the logarithm of the normalization constant Z (x) in (2) which
can, using (1), be rewritten as
Selecting Weighting Factors in Logarithmic Opinion Pools
269
for any choice of y' for which p(y'lx) is nonzero. Integration over y' with probability
measure p(y'lx) then yields (3) .
Since the ambiguity A is always larger than or equal to zero, we conclude that the
Kullback-Leibler divergence of the logarithmic opinion pool is never larger than the
average Kullback-Leibler divergences of individual experts. The larger the ambiguity, the larger the benefit of combining the experts' probability assessments. Note
that by using Jensen's inequality, it is also possible to show that the Kullback-Leibler
divergence of the linear opinion pool is smaller or equal to the average KullbackLeibler divergences of individual experts. The expression for the ambiguity, defined
as the difference between these two, is much more involved and more difficult to
interpret (see e.g. [10]).
The ambiguity of the logarithmic opinion pool depends on the weighting factors
Wa , not only directly as expressed in (3), but also through p(ylx). We can make
this dependency somewhat more explicit by writing
A
= ~ ~ waw/3K(Pa ,P/3) + ~ ~ Wa [K(P,Pa) a/3
K(Pa,p)] .
(4)
a
=
Proof: Equation (3) is valid for any choice of q(ylx). Substitute q(ylx)
p/3(ylx),
multiply left- and righthand side by w/3, and sum over {3. Simple manipulation of
terms than yields the result.
Alas, the Kullback-Leibler divergence is not necessarily symmetric, i.e., in general
K (PI, P2) # K (P2, pd . However, the difference K (PI, P2) - K (p2, pd is an order
of magnitude smaller than the divergence K(PI,P2) itself. More formally, writing
PI (ylx) = [1 +?(ylx )]p2(ylx) with ?(ylx) small, we can easily show that K (PI, P2) is of
order (some integral over) ?2(ylx) whereas K(PI,P2) - K(p2,pd is of order ?3(ylx).
Therefore, if we have reason to assume that the different models are reasonably
close together, we can, in a first approximation, and will, to make things tractable,
neglect the second term in (4) to arrive at
K(q,p)
~ ~ waK(q,Pa) - ~ L
a
wa w/3 [K(Pa,P/3)
+ K(P/3,Pa)] .
The righthand side of this expression is quadratic in the weighting factors
property which will be very convenient later on.
3
(5)
a ,/3
Wa ,
a
EXAMPLES
Regression. The usual assumption in regression analysis is that the output functionally depends on the input x, but is blurred by Gaussian noise with standard
deviation (j. In other words, the probability model of an expert a can be written
Pa(ylx) =
1
[-(y - fa(x))2]
Vj~
exp
2(j2
.
(6)
The function fa(x) corresponds to the network's estimate of the "true" regression
given input x. The logarithmic opinion pool (1) also leads to a normal distribution
with the same standard deviation (j and with regression estimate
T. Heskes
270
In this case the Kullback-Leibler divergence
is symmetric, which makes (5) exact instead of an approximation. In [7], this has
all been derived starting from a sum-squared error measure.
Variance estimation. There has been some recent interest in using neural networks not only to estimate the mean of the target distribution, but also its variance
(see e.g. [9] and references therein). In fact, one can use the probability density (6)
with input-dependent <7( x). We will consider the simpler situation in which an
input-dependent model is fitted to residuals y, after a regression model has been
fitted to estimate the mean (see also [5]). The probability model of expert 0' can
be written
_ ~
[za(x)y2]
Pa (YjX ) ~ exp 2
'
V
where l/za(x) is the experts' estimate of the residual variance given input x. The
logarithmic opinion pool is of the same form with za(x) replaced by
z(x) =
L waza(x) .
Here the Kullback-Leibler divergence
" -
I\ (P,Pa) =
21
J
[
z(x)
z(x)]
dx p(x) za(x) -log za(x) - 1
is asymmetric. We can use (3) to write the Kullback-Leibler divergence of the
opinion pool explicitly in terms of the weighting factors Wa. The approximation (5),
with
?
Ii(Pa,PP)
?
- 1
+ Ii(Pp,Pa) -"2
J
dx p(x)
[za(x) - zp(x)]2
za(x)zp(x)
,
is much more appealing and easier to handle.
Classification. In a two-class classification problem, we can treat y as a discrete
variable having two possible realizations, e.g., y E {-I, I}. A convenient representation for a properly normalized probability distribution is
1
Pa(yjX)
= 1 + exp[-2h a (x)y]
.
In this logistic representation, the logarithmic opinion pool has the same form with
The Kullback-Leibler divergence is asymmetric, but yields the simpler form
to be used in the approximation (5). For a finite set of patterns, minus the loglikelihood yields the well-known cross-entropy error.
Selecting Weighting Factors in Logarithmic Opinion Pools
271
The probability models in these three examples are part of the exponential family.
The mean f 0:, inverse variance zo:, and logit ho: are the canonical parameters. It is
straightforward to show that, with constant dispersion across the various experts,
the canonical parameter of the logarithmic opinion pool is always a weighted average
of the canonical parameters of the individual experts. Slightly more complicated
expressions arise when the experts are allowed to have different estimates for the
dispersion or for probability models that do not belong to the exponential family.
4
SELECTING WEIGHTING FACTORS
The decomposition (3) and approximation (5) suggest an objective method for
selecting weighting factors in logarithmic opinion pools. We will sketch this method
for an ensemble of models belonging to the same class, say feedforward neural
networks with a fixed number of hidden units, where each model is optimized on a
different bootstrap replicate of the available data set.
Su ppose that we have available a data set consisting of P combinations {x/J, y/J }.
As suggested in [3], we construct different models by training them on different
bootstrap replicates of the available data set. Optimizing nonlinear models is often
an unstable process: small differences in initial parameter settings or two almost
equivalent bootstrap replicates can result in completely different models. Neural
networks, for example, are notorious for local minima and plateaus in weight space
where models might get stuck. Therefore, the incorporation of weighting factors,
even when models are constructed using the same pro?cedure, can yield a better generalizing opinion pool. In [4] good results have been reported on several regression
lin
problems. Balancing clearly outperformed bagging, which corresponds to Wo:
with n the number of experts, and bumping, which proposes to keep a single expert.
=
Each example in the available data set can be viewed as a realization of an unknown
probability density characterized by p(x) and q(ylx). We would like to choose the
weighting factors Wo: such as to minimize the Kullback-Leibler divergence K(q, p) of
the opinion pool. If we accept the approximation (5), we can compute the optimal
weighting factors once we know the individual Kullbacks K(q,po:) and the Kullbacks
between different models K(po:, PI3). Of course, both q(ylx) and p(x) are unknown,
and thus we have to settle for estimates.
In an estimate for K(po:,PI3) we can simply replace the average over p(x) by an
average over all inputs x/J observed in the data set:
A similar straightforward replacement for q(ylx) in an estimate for K(q, Po:) is biased, since each expert has, at least to some extent, been overfitted on the data set.
In [4] we suggest how to remove this bias for regression models minimizing sumsquared errors. Similar compensations can be found for other probability models.
Having estimates for both the individual Kullback-Leibler divergences K(q,po:) and
the cross terms K (Po:, PI3), we can optimize for the weighting factors Wo:. Under the
constraints 2:0: Wo: = 1 and Wo: 2: 0 the approximation (5) leads to a quadratic programming problem. Without this approximation, optimizing the weighting factors
becomes a nasty exercise in nonlinear programming.
272
T. Heskes
The solution of the quadratic programming problem usually ends up at the edge of
the unit cube with many weighting factors equal to zero. On the one hand, this is
a beneficial property, since it implies that we only have to keep a relatively small
number of models for later processing. On the other hand, the obtained weighting factors may depend too strongly on our estimates of the individual Kullbacks
K (q, Pet). The following version prohibits this type of overfitting. Using simple
statistics, we obtain a rough indication for the accuracy of our estimates K(q,pet).
This we use to generate several, say on the order of 20, different samples with
estimates {K (q, pI), .. . , K (q, Pn)}. For each of these samples we solve the corresponding quadratic programming problem and obtain a set of weighting factors.
The final weighting factors are obtained by averaging. In the end, there are less
experts with zero weighting factors, at the advantage of a more robust procedure.
Acknowledgements
I would like to thank David Tax, Bert Kappen, Pierre van de Laar, Wim Wiegerinck,
and the anonymous referees for helpful suggestions. This research was supported
by the Technology Foundation STW, applied science division of NWO and the
technology programme of the Ministry of Economic Affairs.
References
[1] J. Benediktsson and P. Swain. Consensus theoretic classification methods.
IEEE Transactions on Systems, Man, and Cybernetics, 22:688-704, 1992.
[2] R. Bordley. A multiplicative formula for aggregating probability assessments.
Management Science, 28:1137-1148,1982.
[3] L. Breiman. Bagging predictors. Machine Learning, 24:123-140, 1996.
[4] T. Heskes. Balancing between bagging and bumping. In M. Mozer, M. Jordan,
and T. Petsche, editors, Advances in Neural Information Processing Eystems
9, pages 466-472, Cambridge, 1997. MIT Press.
[5] T. Heskes. Practical confidence and prediction intervals. In M. Mozer, M. Jordan, and T. Petsche, editors, Advances in Neural Information Processing Eystems 9, pages 176-182, Cambridge, 1997. MIT Press.
[6] R. Jacobs. Methods for combining experts' probability assessments. Neural
Computation, 7:867-888, 1995.
[7] A. Krogh and J. Vedelsby. Neural network ensembles, cross validation, and
active learning. In G. Tesauro, D. Touretzky, and T. Leen, editors, Advances
in Neural Information Processing Eystems 7, pages 231-238, Cambridge, 1995.
MIT Press.
[8] P. Smyth and D. Wolpert. Stacked density estimation. These proceedings, 1998.
[9] P. Williams. Using neural networks to model conditional multivariate densities.
Neural Computation, 8:843-854, 1996.
[10] D. Wolpert. On bias plus variance. Neural Computation, 9:1211-1243, 1997.
| 1413 |@word version:1 middle:1 seems:2 replicate:1 logit:1 grooteplein:1 simulation:1 decomposition:6 jacob:1 minus:4 kappen:1 initial:1 selecting:8 ala:1 dx:3 written:2 remove:1 affair:1 supplying:1 lx:2 simpler:2 constructed:1 mbfys:1 relying:1 decomposed:1 becomes:1 estimating:1 notation:1 what:1 kind:2 interpreted:1 argmin:1 prohibits:1 finding:2 unit:2 local:1 treat:2 aggregating:1 might:1 plus:1 therein:2 sumsquared:3 suggests:1 range:1 practical:2 bootstrap:3 procedure:3 convenient:3 word:1 confidence:1 suggest:2 get:1 convenience:1 close:1 y1x:1 writing:2 optimize:1 equivalent:3 maximizing:1 straightforward:3 williams:1 starting:1 immediately:2 assigns:2 rule:1 geert:1 handle:1 target:1 play:1 suppose:1 exact:1 programming:6 bumping:2 smyth:1 pa:15 element:1 referee:1 asymmetric:2 observed:1 role:1 overfitted:1 mozer:2 pd:3 trained:1 depend:1 division:1 completely:1 easily:1 joint:1 po:6 various:1 derivation:1 zo:1 stacked:1 outcome:1 heuristic:2 larger:4 solve:1 loglikelihood:5 say:2 statistic:1 itself:1 final:1 dxp:1 advantage:1 indication:1 fr:2 j2:1 combining:3 realization:2 tax:1 supra:2 zp:2 help:1 krogh:1 p2:9 indicate:1 implies:1 drawback:2 settle:1 opinion:28 crux:1 anonymous:1 summation:1 strictly:1 normal:1 exp:3 claim:1 estimation:2 outperformed:1 wim:1 nwo:1 weighted:2 rough:1 clearly:1 mit:3 always:3 gaussian:2 rather:4 pn:1 breiman:1 derived:2 ylj:1 properly:1 helpful:1 dependent:2 accept:1 hidden:1 transformed:1 classification:5 proposes:1 integration:1 cube:1 equal:4 construct:1 never:1 having:2 once:1 unsupervised:1 divergence:20 individual:9 replaced:2 consisting:1 replacement:1 interest:1 multiply:1 righthand:2 replicates:2 weakness:1 nl:1 unconditional:1 integral:2 edge:1 necessary:1 logarithm:1 fitted:2 logp:1 introducing:1 deviation:2 predictor:1 swain:1 too:1 kullbackleibler:1 reported:1 dependency:1 density:6 probabilistic:1 pool:30 together:1 squared:6 ambiguity:5 management:1 containing:2 choose:2 expert:19 suggesting:1 de:1 matter:1 blurred:1 explicitly:1 depends:2 later:2 multiplicative:1 bayes:1 complicated:1 minimize:1 accuracy:1 variance:12 ensemble:5 yield:6 correspond:1 ofthe:2 generalize:1 bayesian:3 cybernetics:1 za:7 plateau:1 touretzky:1 pp:2 involved:3 obvious:1 vedelsby:1 naturally:2 proof:3 carefully:1 follow:2 tom:2 leen:1 laar:1 strongly:1 generality:1 sketch:1 hand:2 su:1 nonlinear:2 assessment:3 logistic:1 building:1 normalized:1 true:3 multiplier:1 y2:1 assigned:1 symmetric:2 leibler:17 nonzero:1 numerator:1 generalized:1 stress:1 complete:1 theoretic:1 pro:1 belong:1 interpret:3 functionally:1 cambridge:3 heskes:6 reliability:1 closest:1 multivariate:1 recent:1 optimizing:2 irrelevant:1 tesauro:1 scenario:1 manipulation:1 inequality:1 minimum:1 ministry:1 somewhat:1 ii:3 characterized:1 cross:4 long:1 lin:1 cedure:1 prediction:1 regression:10 normalization:2 whereas:1 interval:1 biased:1 probably:1 elegant:1 thing:1 jordan:2 feedforward:1 economic:1 pi3:3 expression:3 wo:5 ylx:22 netherlands:1 generate:1 canonical:3 estimated:1 discrete:2 write:1 prevent:1 sum:7 inverse:1 arrive:2 family:2 almost:1 dy:2 quadratic:7 constraint:3 incorporation:1 relatively:1 combination:6 belonging:1 beneficial:1 smaller:3 slightly:2 xlj:1 across:1 appealing:1 notorious:1 equation:1 discus:1 committee:1 know:1 tractable:1 end:2 available:4 rewritten:1 petsche:2 pierre:1 ho:1 substitute:1 bagging:4 neglect:2 somewhere:1 objective:2 fa:2 usual:1 distance:1 thank:1 extent:1 unstable:1 consensus:1 reason:1 pet:2 minimizing:2 kun:1 difficult:2 statement:6 nijmegen:2 stw:1 unknown:2 dispersion:2 finite:2 compensation:1 situation:1 nasty:1 bert:1 introduced:1 david:1 optimized:2 waj:1 suggested:1 below:1 pattern:1 usually:1 treated:1 residual:2 technology:2 benediktsson:1 acknowledgement:1 suggestion:3 validation:1 foundation:2 editor:3 story:1 pi:7 balancing:2 course:1 supported:1 bias:7 formal:2 side:2 benefit:1 van:1 overcome:1 valid:1 stuck:1 made:1 programme:1 transaction:1 kullback:18 keep:2 overfitting:2 active:1 conclude:1 continuous:1 reasonably:1 robust:1 necessarily:1 vj:1 main:1 whole:2 noise:1 arise:1 allowed:1 bordley:1 explicit:1 exponential:2 exercise:1 weighting:30 ix:1 externally:1 formula:1 specific:1 jensen:1 magnitude:1 easier:3 entropy:2 generalizing:1 logarithmic:23 wolpert:2 simply:2 ez:1 lagrange:1 expressed:1 applies:1 corresponds:4 conditional:2 viewed:1 presentation:1 replace:1 man:1 contrasted:1 averaging:11 wiegerinck:1 called:2 la:1 formally:1 latter:1 wak:1 |
455 | 1,414 | Visual Navigation in a Robot using
Zig-Zag Behavior
M. Anthony Lewis
Beckman Institute
405 N. Mathews Avenue
University of Illinois
Urbana, lllinois 61801
Abstract
We implement a model of obstacle avoidance in flying insects on a small,
monocular robot. The result is a system that is capable of rapid navigation
through a dense obstacle field. The key to the system is the use of zigzag
behavior to articulate the body during movement. It is shown that this behavior
compensates for a parallax blind spot surrounding the focus of expansion normally found in systems without parallax behavior. The system models the cooperation of several behaviors: halteres-ocular response (similar to VOR),
optomotor response, and the parallax field computation and mapping to motor
system. The resulting system is neurally plausible, very simple, and should be
easily hosted on aVLSI hardware.
1 INTRODUCTION
Srinivasan and Zhang (1993) describe behavioral evidence for two distinct movement
detecting systems in bee: (1) A direction selective pathway with low frequency response
characteristics serving the optomotor response and (2) A non-direction selective movement system with higher frequency response serving functions of obstacle avoidance and
the 'tunnel centering' response where the animal seeks a flight path along the centerline of
a narrow corridor. Recently, this parallel movement detector view has received support
from anatomical evidence in fly (Douglass and Strausfeld, 1996). We are concerned here
with the implications of using non-direction selective movement detectors for tasks such
as obstacle avoidance.
A reasonable model of a non-direction selective pathwa:? would be that this pathway is
computing the absolute value of the optic flow, i.e. s = II[x, Y111 where y are the
components of the optic flow field on the retina at the point [x, y 1 .
x,
What is the effect of using the absolute value of the flow field and throwing away
direction information? In section 2 we analyze the effect of a non-direction selective
movement field. We understand from this analysis that rotational information, and the
limited dynamic range of real sensors contaminates the non-direction selective field and
823
Visual Navigation in a Robot Using Zig-Zag Behavior
probably prevents the use of this technique in an area around the direction heading of the
observer.
One technique to compensate for this 'parallax blind spot' is by periodically changing the
direction of the observer. Such periodic movements are seen in insects as well as lower
vertebrates and it is suggestive that these movements may compensate for this basic
problem.
In Section 3, we describe a robotic implementation using a crude non-direction selective
movement detector based on a rectified temporal derivative of luminosity. Each ' neuron'
in the model retina issues a vote to control the motors of the robot. This system, though
seemingly naively simple, compares favorably with other robotic implementations that
rely on the optic flow or a function of the optic flow (divergence). These techniques
typically require a large degree of spatial temporal averaging and seem computationally
complex. In addition, our model agrees better with with the biological evidence.
Finally, the technique presented here is amenable to implementation in custom aVLSI or
mixed aVLSIIdVLSI chips. Thus it should be possible to build a subminiature visually
guided navigation system with several (one?) low-power simple custom chips.
2 ANALYSIS OF NON-DIRECTION SELECTIVE MOVEMENT
DETECTION SYSTEM
Let us assume a perspective projection
(1)
where A. is the focal length of the lens, X, Y, Z is the position of a point in the environment, and x, y is the projection of that point on the retinal plane. The velocity of the image
of a moving point in the world can be found by differentiating (1) with respect to time:
(2)
If we assume that objects in the environment are fixed in relation to one-a~-other and that
and relative
the observer is moving with relativTtranslational velocity Cv = [vx Vy v
to the environment given in frame c, a point in the
rotational velocity cn = rrox roy roJ
environment has relati~e v~loclty:
J
(3)
Now substituting in (2):
~ ~l: ~Jv.+{:> -~~;' j'n.
=
(4)
and taking the absolute value of the optic flow :
r;; '~l"x- ~(Xy... +..y,x' +1)+ y",l)' +~+y+ ~(-..p' +\)+xY"y+x",l]j
=
where we have made the substitution:
[X/2 Y/~ [(1 IlJ
-t
(5)
(that is, the heading direction).
We can see that the terms involving [rox ro vrozl cannot be separated from the x, y terms. If
we assume that [rox roy roJ = 0 then we can r~arrange the equation as:
M. A. Lewis
824
.-l()
L>
s
in the case of Z translation. If ITzl
_ITZI_
-1Zf-
=0
~ -1(s)
=
lsi
AJ[(X-a)2 + (y_~)2J
(6)
then we have:
1
IZI
=
lsi
AJTx 2 + Ty2
(7)
this corresponds to the case of pure lateral translations. Locusts (as well as some vertebrates) use peering or side to side movements to gauge distances before jumping.
We call the quantity in (6) inverse relative depth. Under the correct circumstances it is
equivalent to the reciprocal of time to contact.
Equation (6) can be restated as: ~-l(s) = gjS where g is a gain factor that depends on the
current direction heading and the position in the retina. This gain factor can be implemented neurally as a shunting inhibition, for example.
This has the following implications. If the observer is using a non-direction sensitive
movement detector then (A) it must rotationally stabilize its eyes (B) it must dynamically
alter the gain of this infonnation in a pathway between the retinal input and motor output
or it must always have a constant direction heading and use constant gain factors.
In real systems there is likely to be imperfection in rotational stabilization of the observer
as well as sensors with limited dynamic range. To understand the effect of these, let us
assume that there is a base-line noise level 0 and we assume that this defines a minimum
threshold substituting s = 0, we can find a level curve for the minimum detectability of
an object, i.e.:
(8)
Thus, for constant depth and for 0 independent of the spatial position on the retina, the
level curve is a circle. The circle increases in radius with increasing distance, and noise,
and decreases with increasing speed. The circle is centered around the direction heading.
The solution to the problem of a 'parallax blind spot' is to make periodic changes of
direction. This can be accomplished in an open loop fashion or, perhaps, in an image
driven fashion as suggested by Sobey (1994).
3 ROBOT MODEL
Figure la is a photograph of the robot model. The robot's base is a Khepera Robot. The
Khepera is a small wheeled robot a little over 2" in diameter and uses differential drive
motors. The robot has been fitted with a solid-state gyro attached to its body. This
gyroscope senses angular velocities about the body axis and is aligned with the axis of the
camera joint. A camera, capable of rotation about an axis perpendicular to the ground
plane, is also attached. The camera has a field of view of about 90 0 and can swing of
?90? . The angle of the head rotation is sensed by a small potentiometer.
For convenience, each visual process is implemented on a separate Workstation (SGI
Indy) as a heavyweight process. Interprocess communication is via PVM distributed
computing library. Using a distributed processing model, behaviors can be dynamically
added and deleted facilitating analysis and debugging.
3.1 ROBOT CONTROL SYSTEM
The control is divided into control modules as illustrated in Fig 2. At the top of the
drawing we see a gaze stabilization pathway. This uses a gyro (imitating a halteres organ)
for stabilization of rapid head movements. In addition, a visual pathway, using direction
selective movement detector (DSMD) maps is used for slower optomotor response. Each
of the six maps uses correlation type detectors (Borst and Egelhaaf, 1989). Each map is
Visual Navigation in a Robot Using Zig-Zag Behavior
825
Figure 1. Physical setup. (A) Modified Khepera Robot with camera and gyro
mounted. (B) Typical obstacle field run experiment.
tuned to a different horizontal velocity (three for left image translations and three for right
image translations).
The lower half of the drawing shows the obstacle avoidance pathway. A crude nondirection selective movement detector is created using a simple temporal derivative. The
use of this as a movement detector was motivated by the desire to eventually replace the
camera front end with a Neuromorphic chip. Temporal derivative chips are readily available (Delbrtick and Mead, 1991).
Next we reason that the temporal derivative gives a crude estimate of the ~bsolute value of
the optic flow. For example if we expect only horizontal flows then: E xX = - E t (Horn
and Shunck, 1981). Here E, is the temporal derivative of the luminosity and gA; is the
spatial derivative. If we sample over a patch of the image, Ex will take on a range of
values. If we take the average rectified temporal derivative over a patch then
Ixl = I-E,I/IExl. Thus the average rectified temporal derivative over a patch will give a
velocity proportional the absolute value of the optic flow.
In order to make the change to motor commands, we use a voting scheme. Each pixel in
the nondirection selective movement detector field (NDSMD) votes on a direction for the
robot. The left pixels for a right turn and the right pixels vote for a left turn. The left and
right votes are summed. In certain experiments described below the difference of the left
and right votes was used to drive the rotation of the robot. In others a symmetry breaking
scheme was used. It was observed that with an object dead ahead of the robot, often the
left and right activation would have high but nearly equal activation. In the symmetry
breaking scheme, the side with the lower activation was further decrease by a factor of
50%. This admittedly ad hoc solution remarkably improved the performance in the nonzig-zag case as noted below.
The zig-zag behavior is implemented as a feedforward command to the motor system and
is modeled as:
Khepera
COZigZag
.
= stn(cot)K
Finally, a constant forward bias is added to each wheel so the robot makes constant
progress. K is chosen empirically but in principle one should be able to derive it using the
analysis in section 2.
As described above, the gaze stabilization module has control of head rotation and the zigzag behavior and the depth from parallax behavior control the movement of the robot's
body. During normal operation, the head may exceed the ?90? envelope defined by the
mechanical system. This problem can be addressed in several ways among them are by
making a body saccade to bring the body under the head or making a head saccade to align
the head with the body. We choose the later approach solely because it seemed to work
better in practice.
826
M. A. Lewis
c
J
-k
e
Gyro
Quick Phase Nystagmus
Gaze Stabilillltion
Cmd
Camera
60x80image
l -_ _--:-_ _ _
L-_~-:_;:
a,-l(s)
Motor
Map
f--__~
NDSMD
Figure 2. ZigZag Navigation model is composed of a gaze stabilization system
(top) and an obstacle avoidance system (bottom). See text.
3.2 BIOLOGICAL INSPIRATION FOR MODEL
Course-grained visual pathways are modeled using inspiration from insect neurobiology.
The model of depth from parallax is inspired by details given in Srinivasan & Zhang
(1993) on work done in bees. Gaze stabilization using a fast channel, mediated by the
halteres organs, and a slow optomotor response is inspired by a description of the blowfly
Calliphora as reviewed by Hengstenberg (1991).
4 EXPERIMENTS
Four types of experimental setups were used. These are illustrated in Fig 3. In setup 1 the
robot must avoid a dense field of obstacles (empty soda cans). This is designed to test the
basic competence of this technique. In setup 2, thin dowels are place in the robot's path.
This tests the spatial resolving capability of the robot. Likewise setup 3 uses a dense
obstacle field with one opening replaced by a lightly textured surface.
Finally, experimental setup 4 uses a single small object (1cm black patch) and tests the
distance at which the robot can 'lock-on' to a target. In this experiment, the avoidance
field is sorted for a maximal element over a given threshold. A target cross is placed at this
maximal element. The closest object should correspond with this maximal element. If a
maximal element over a threshold is identified for a continuous 300ms and the target cross
is on the correct target, the robot is stopped and its distance to the object is measured. The
larger the distance, the better.
5 RESULTS
The results are described briefly here. In the setup 1 without the use of symmetry breaking, the scores were ZigZag: 10 Success, 0 Failures and the non-ZigZag: 4 Success and 6
Failures. With Symmetry Breaking installed the results were: ZigZag: 49 Success, 3
Failures and the non-ZigZag: 44 Success and 9 failures.
In the case palisades test: ZigZag: 22 Success, 4 Failures and the non-ZigZag: 14 Success
and 11 failures.
In the false opening case: ZigZag: 8 Success, 2 Failures and the non-ZigZag: 6 Success
and 4 Failures.
Finally, in the distance-to-Iock setup, a lock was achieved at an average distance 21.3 CM
(15 data points) for zigzag and 9.6 cm (15 data points) for the non-zigzag case.
Visual Navigation in a Robot Using Zig-ZAg Behavior
Dense Obstacle
Field A voidance
~
Palisades Test
?? ?
~o
@
False Opening A voidance
827
Distance-To-Lock
?
Thin Dowels
lightly textured barrier
8
Figure 3. Illustrations of the four experimental setups.
We tentatively conclude that zig-zag behavior should improve performance in robot and
in animal navigation.
6 DISCUSSION
In addition to the robotic implementation presented here, there have been many other
techniques presented in the literature. Most relevant is Sobey (1994) who uses a zigzag
behavior for obstacle avoidance. In this work, optic flow is computed through a process of
discrete movements where 16 frames are collected, the robot stops, and the stored frames
are analyzed for optic flow. The basic strategy is very clever: Always choose the next
move in the direction of an identified object. The reasoning is that since we know the
distance to the object in this direction, we can confidently move toward the object,
stopping before collision. The basic idea of using zig-zag behavior is similar except that
the zig-zagging is driven by perceptual input. In addition, the implementation requires
estimation of the flow field requiring smoothing over numerous images. Finally, Sobey
uses Optic Flow and we use the absolute value of the Optic Flow as suggested by biology.
Franceschini et al (1992) reports an analog implementation that uses elementary movement detectors. A unique feature is the non-uniform sampling and the use of three separate
arrays. One array uses a sampling around the circumference. The other two sampling
systems are mounted laterally on the robot and concentrate in the 'blind-spot' immediately
in front of the robot. It is not clear that the strategy of using three sensor arrays, spatially
separated, and direction selective movement detectors is in accord with the biological
constraints.
Santos-Victor et al (1995) reports a system using optic flow and having lateral facing
cameras. Here the authors were reproducing the centering reflex and did not focus on
avoiding obstacles in front of the robot. Coombs and Roberts (1992,1993) use a similar
technique. Weber et al (1996) describe wall following and stopping in front of an obstacle
using an optic flow measure.
Finally, a number of papers report the use of flow field divergence, apparently first
suggested by Nelson and Aloimonos (1989). This requires the computation of higher
derivatives and requires significant smoothing. Even in this case, there is a problem of a
'parallax hole.' See Fig. 3 of that article, for example. In any case they did not implement
their idea on a mobile robot. However, this approach has been followed up with an
implementation in a robot by Camus et al (1996) reporting good results.
The system described here presents a physical model of insect like behavior integrated on
a small robotic platform. Using results derived from an analysis of optic flow, we concluded that a zig-zag behavior in the robot would allow it to detect obstacles in front of the
robot by periodically articulating the blind spot.
The complexity of the observed behavior and the simplicity of the control is striking. The
robot is able to navigate through a field of obstacles, always searching out a freeway for
movement.
828
M. A. Lewis
The integrated behavior outlined here should be a good candidate for a neuromorphic
implementation. A multichip (or single chip?) system could be envisioned using a relatively simple non-directional 2-d movement detector. Two arrays of perpendicular I-d
array of movement detectors should be sufficient for the optomotor response. This information could then be mapped to a circuit comprised of a few op-amp adder circuits and
then sent to the head and body motors. Even the hal teres organ could be simulated with a
silicon fabricated gyroscope. The results would be an extremely compact robot capable of
autonomous, visually guided navigation.
Finally, from our analysis of optic flow, we can make a reasonable prediction abuut the
neural wiring in flying insects. The estimated depth of objects in the environment depends
on where the object falls on the optic array as well as the ratio of translation to forward
movement. Thus a bee or a fly should probably modulate its incoming visual signal to
account for this time varying interpretation of the scene. We would predict that there
should be motor information, related to the ratio of forward to lateral velocities would be
projected to the non-directional selective motion detector array. This would allow a valid
time varying interpretation of the scene in a zig-zagging animal.
Acknowledgments
The author acknowledges the useful critique of this work by Narendra Ahuja, Mark Nelson, John
Hart and Lucia Simo. Special thanks to Garrick Kremesic and Barry Stout who assisted with the
experimental setup and the modification of the Khepera. The author acknowledges the support of
ONR grant NOOOI49610657. The author also acknowledges the loan of the Khepera from UCLA
(NSF grant CDA-9303148).
References
A. Borst and M. Egelhaaf (1989), Principles of Visual Motion Detection, Trends in Neurosciences,
12(8):297-306
T. Camus, D. Coombs, M. Herman, and T.-H. Hong (1996), "Real-time Single-Workstation Obstacle Avoidance Using Only Wide-Field Flow Divergence", Proceedings of the 13th International
Conference on Pattern Recognition. pp. 323-30 vol.3
D. Coombs and K. Roberts (1992), '''Bee-Bot': Using Peripheral Optical Flow to A void Obstacles"
SPIE Vol 1825, Intelligent Robots and Computer Vision XI, pp 714-721.
I
D. Coombs and K. Roberts (1993), "Centering behavior using peripheral vision", Proc. 1993 IEEE
Computer Society Conf. CVPR pp. 440-5, 16 refs. 1993
T. Delbriick and C. A. Mead (1991), Time-derivative adaptive silicon photoreceptor array. Proc.
SPIE - Int. Soc. Opt. Eng. (USA). vol 1541, pp. 92-9.
J. K. Douglass and N. 1. Strausfeld (1996), Visual Motion-Detection Circuits in Flies: Parallel
Direction- and Non-Direction-Sensitive Pathways between the Medulla and Lobula Plate, J. of
Neuroscience 16(15):4551-4562.
N. Franceschini, J. M. Pichon and C. Blanes (1992), "From Insect Vision to Robot Vision", Phil.
Trans. R. Soc Lond. B. 337, pp 283-294.
R. Hengstenberg (1991), Gaze Control in the Blowfly Calliphora: a Multisensory, Two-Stage
Integration Process, Seminars, in the Neurosciences, Vol3,pp 19-29.
B. K. P. Hom and B. G. Shunck (1981), "Determining Optic Flow", Artificial Intelligence, 17(13):185-204.
R. C. Nelson and J. Y. Aloimonos (1989) Obstacle Avoidance Using Flow Field Divergence, IEEE
Trans. on Pattern Anal. and Mach. Intel. 11(10):1102-1106.
1. Santos-Victor, G. Sandini, F. Curotto and S. Garibaldi (1995), "Divergent Stereo in Autonomous
Navigation: From Bees to Robots," Int. J. of Compo Vis. 14, pp 159-177.
P. J. Sobey (1994), "Active Navigation With a Monocular Robot" BioI. Cybern, 71:433-440
M. V. Srinivasan and S. W. Zhang (1993), Evidence for Two Distinct Movement-Detecting Mechanisms in Insect Vision, Naturwissenschaften, 80, pp 38-41.
K. Weber, S. Venkatash and M.V. Srinivasan (1996), "Insect Inspired Behaviours for the Autonomous Control of Mobile Robots" Proc. of ICPR'96, pp 156-160.
| 1414 |@word briefly:1 open:1 seek:1 sensed:1 eng:1 solid:1 substitution:1 score:1 tuned:1 amp:1 current:1 activation:3 must:4 readily:1 john:1 vor:1 periodically:2 motor:9 designed:1 half:1 intelligence:1 plane:2 reciprocal:1 compo:1 detecting:2 centerline:1 zhang:3 along:1 differential:1 corridor:1 pathway:8 behavioral:1 parallax:8 cot:1 rapid:2 behavior:20 inspired:3 borst:2 little:1 freeway:1 vertebrate:2 increasing:2 xx:1 circuit:3 what:1 santos:2 cm:3 fabricated:1 temporal:8 voting:1 laterally:1 ro:1 mathews:1 normally:1 control:9 grant:2 before:2 installed:1 mach:1 mead:2 critique:1 path:2 solely:1 black:1 potentiometer:1 dynamically:2 khepera:6 limited:2 range:3 perpendicular:2 locust:1 horn:1 camera:7 unique:1 acknowledgment:1 cmd:1 practice:1 implement:2 spot:5 area:1 projection:2 cannot:1 convenience:1 ga:1 wheel:1 clever:1 cybern:1 equivalent:1 map:4 quick:1 circumference:1 phil:1 restated:1 simplicity:1 immediately:1 pure:1 avoidance:9 array:8 searching:1 autonomous:3 target:4 us:9 velocity:7 roy:2 element:4 trend:1 recognition:1 observed:2 bottom:1 module:2 fly:3 movement:26 decrease:2 zig:10 envisioned:1 environment:5 complexity:1 calliphora:2 dynamic:2 flying:2 textured:2 easily:1 joint:1 chip:5 surrounding:1 separated:2 distinct:2 fast:1 describe:3 artificial:1 larger:1 plausible:1 cvpr:1 drawing:2 compensates:1 seemingly:1 hoc:1 maximal:4 aligned:1 loop:1 relevant:1 description:1 sgi:1 empty:1 object:11 derive:1 avlsi:2 coombs:4 measured:1 op:1 received:1 progress:1 indy:1 soc:2 implemented:3 direction:24 guided:2 radius:1 concentrate:1 correct:2 stabilization:6 vx:1 centered:1 require:1 behaviour:1 wall:1 articulate:1 biological:3 elementary:1 opt:1 assisted:1 around:3 ground:1 normal:1 visually:2 wheeled:1 mapping:1 predict:1 substituting:2 narendra:1 arrange:1 estimation:1 proc:3 beckman:1 infonnation:1 sensitive:2 agrees:1 organ:3 gauge:1 sensor:3 always:3 imperfection:1 modified:1 avoid:1 mobile:2 command:2 varying:2 derived:1 focus:2 detect:1 stopping:2 typically:1 integrated:2 relation:1 selective:13 pixel:3 issue:1 among:1 insect:8 animal:3 spatial:4 summed:1 smoothing:2 platform:1 special:1 field:18 equal:1 integration:1 having:1 sampling:3 biology:1 nearly:1 thin:2 alter:1 others:1 report:3 intelligent:1 roj:2 opening:3 retina:4 few:1 composed:1 divergence:4 replaced:1 phase:1 detection:3 custom:2 navigation:11 analyzed:1 sens:1 zigzag:14 implication:2 amenable:1 capable:3 xy:2 jumping:1 simo:1 circle:3 cda:1 fitted:1 stopped:1 obstacle:18 neuromorphic:2 uniform:1 comprised:1 stout:1 medulla:1 front:5 stored:1 periodic:2 thanks:1 international:1 multichip:1 gaze:6 ilj:1 choose:2 rox:2 dead:1 conf:1 derivative:10 naturwissenschaften:1 account:1 retinal:2 stabilize:1 int:2 vi:1 blind:5 depends:2 ad:1 view:2 observer:5 later:1 franceschini:2 analyze:1 apparently:1 parallel:2 capability:1 nystagmus:1 characteristic:1 likewise:1 who:2 correspond:1 directional:2 rectified:3 drive:2 detector:14 centering:3 failure:8 frequency:2 ocular:1 pp:9 spie:2 workstation:2 gain:4 stop:1 higher:2 response:9 izi:1 improved:1 done:1 though:1 angular:1 stage:1 correlation:1 flight:1 horizontal:2 adder:1 defines:1 aj:1 perhaps:1 hal:1 usa:1 effect:3 requiring:1 swing:1 inspiration:2 spatially:1 voidance:2 illustrated:2 wiring:1 during:2 noted:1 m:1 hong:1 plate:1 motion:3 bring:1 reasoning:1 image:6 weber:2 recently:1 rotation:4 physical:2 empirically:1 attached:2 analog:1 interpretation:2 strausfeld:2 significant:1 silicon:2 cv:1 focal:1 outlined:1 illinois:1 moving:2 robot:40 surface:1 inhibition:1 base:2 align:1 closest:1 perspective:1 driven:2 certain:1 onr:1 success:8 accomplished:1 victor:2 seen:1 rotationally:1 minimum:2 barry:1 signal:1 ii:1 gjs:1 neurally:2 resolving:1 cross:2 compensate:2 y_:1 divided:1 hart:1 shunting:1 prediction:1 involving:1 basic:4 circumstance:1 vision:5 accord:1 achieved:1 addition:4 remarkably:1 addressed:1 void:1 concluded:1 envelope:1 probably:2 sent:1 flow:23 seem:1 call:1 feedforward:1 exceed:1 concerned:1 identified:2 idea:2 cn:1 avenue:1 pvm:1 six:1 motivated:1 stereo:1 articulating:1 tunnel:1 useful:1 collision:1 clear:1 hardware:1 diameter:1 lsi:2 vy:1 nsf:1 bot:1 estimated:1 neuroscience:3 serving:2 anatomical:1 discrete:1 detectability:1 vol:3 srinivasan:4 key:1 four:2 threshold:3 lobula:1 deleted:1 changing:1 jv:1 douglas:2 run:1 inverse:1 angle:1 soda:1 striking:1 place:1 reporting:1 reasonable:2 patch:4 hom:1 followed:1 ahead:1 optic:17 constraint:1 throwing:1 scene:2 lucia:1 ucla:1 lightly:2 speed:1 extremely:1 lond:1 optical:1 relatively:1 debugging:1 peripheral:2 icpr:1 making:2 modification:1 imitating:1 computationally:1 monocular:2 equation:2 turn:2 eventually:1 mechanism:1 know:1 end:1 available:1 operation:1 away:1 blowfly:2 palisade:2 slower:1 top:2 lock:3 relati:1 build:1 society:1 contact:1 move:2 added:2 quantity:1 strategy:2 distance:9 separate:2 mapped:1 lateral:3 simulated:1 nelson:3 collected:1 reason:1 toward:1 length:1 modeled:2 illustration:1 rotational:3 ratio:2 setup:10 robert:3 favorably:1 implementation:8 anal:1 zf:1 neuron:1 urbana:1 zagging:2 luminosity:2 neurobiology:1 communication:1 head:8 frame:3 reproducing:1 delbriick:1 ty2:1 competence:1 mechanical:1 narrow:1 aloimonos:2 trans:2 able:2 suggested:3 below:2 pattern:2 confidently:1 herman:1 gyroscope:2 power:1 rely:1 scheme:3 improve:1 eye:1 library:1 numerous:1 axis:3 created:1 acknowledges:3 shunck:2 mediated:1 tentatively:1 text:1 literature:1 stn:1 bee:5 determining:1 relative:2 expect:1 mixed:1 mounted:2 proportional:1 facing:1 ixl:1 degree:1 sandini:1 sufficient:1 article:1 principle:2 translation:5 course:1 cooperation:1 placed:1 heading:5 hengstenberg:2 side:3 bias:1 understand:2 allow:2 institute:1 fall:1 wide:1 taking:1 barrier:1 differentiating:1 absolute:5 distributed:2 curve:2 depth:5 world:1 valid:1 seemed:1 forward:3 made:1 author:4 projected:1 adaptive:1 compact:1 suggestive:1 robotic:4 incoming:1 active:1 conclude:1 xi:1 continuous:1 reviewed:1 channel:1 contaminates:1 symmetry:4 expansion:1 complex:1 anthony:1 did:2 dense:4 noise:2 facilitating:1 ref:1 body:8 egelhaaf:2 fig:3 intel:1 hosted:1 fashion:2 ahuja:1 slow:1 seminar:1 position:3 candidate:1 crude:3 perceptual:1 breaking:4 grained:1 navigate:1 divergent:1 evidence:4 naively:1 false:2 optomotor:5 hole:1 gyro:4 photograph:1 likely:1 visual:10 prevents:1 desire:1 saccade:2 reflex:1 corresponds:1 lewis:4 bioi:1 modulate:1 sorted:1 replace:1 change:2 loan:1 typical:1 except:1 averaging:1 admittedly:1 lens:1 experimental:4 la:1 vote:5 photoreceptor:1 zag:9 multisensory:1 support:2 mark:1 avoiding:1 ex:1 |
456 | 1,415 | A Hippocampal Model of Recognition Memory
Randall C. O'Reilly
Department of Psychology
University of Colorado at Boulder
Campus Box 345
Boulder, CO 80309-0345
oreilly@psych.colorado.edu
Kenneth A. Norman
Department of Psychology
Harvard University
33 Kirkland Street
Cambridge, MA 02138
nonnan@wjh.harvard.edu
James L. McClelland
Department of Psychology and
Center for the Neural Basis of Cognition
Carnegie Mellon University
Pittsburgh, PA 15213
jlm@cnbc.cmu.edu
Abstract
A rich body of data exists showing that recollection of specific information makes an important contribution to recognition memory, which
is distinct from the contribution of familiarity, and is not adequately captured by existing unitary memory models. Furthennore, neuropsychological evidence indicates that recollection is sub served by the hippocampus.
We present a model, based largely on known features of hippocampal
anatomy and physiology, that accounts for the following key characteristics of recollection: 1) false recollection is rare (i.e., participants rarely
claim to recollect having studied nonstudied items), and 2) increasing interference leads to less recollection but apparently does not compromise
the quality of recollection (i.e., the extent to which recollected infonnation veridically reflects events that occurred at study).
1 Introduction
For nearly 50 years, memory researchers have known that our ability to remember specific
past episodes depends critically on the hippocampus. In this paper, we describe our initial
attempt to use a mechanistically explicit model of hippocampal function to explain a wide
range of human memory data.
Our understanding of hippocampal function from a computational and biological perspec-
74
R. C. 0 'Reilly, K. A. Norman and 1. L McClelland
tive is based on our prior work (McClelland, McNaughton, & O'Reilly, 1995; O'Reilly &
McClelland, 1994). At the broadest level, we think that the hippocampus exists in part to
provide a memory system which can learn arbitrary information rapidly without suffering
undue amounts of interference. This memory system sits on top of, and works in conjunction with, the neocortex, which learns slowly over many experiences, producing integrative
representations of the relevant statistical features of the environment. The hippocampus accomplishes rapid, relatively interference-free learning by using relatively non-overlapping
(pattern separated) representations. Pattern separation occurs as a result of 1) the sparseness of hippocampal representations (relative to cortical representations), and 2) the fact
that hippocampal units are sensitive to conjunctions of cortical features - given two cortical patterns with 50% feature overlap, the probability that a particular conjunction of
features will be present in both patterns is much less than 50%.
We propose that the hippocampus produces a relatively high-threshold, high-quality recollective response to test items. The response is "high-threshold" in the sense that studied
items sometimes trigger rich recollection (defined as "retrieval of most or all of the test
probe's features from memory") but lures never trigger rich recollection. The response is
"high-quality" in the sense that, most of the time, the recollection signal consists of part
or all of a single studied pattern, as opposed to a blend of studied patterns. The highthreshold, high-quality nature of recollection can be explained in terms of the conjunctivity of hippocampal representations: Insofar as recollection is a function of whether the
features of the test probe were encountered together at study, lures (which contain many
novel feature conjunctions, even if their constituent features are familiar) are unlikely to
trigger rich recollection; also, insofar as the hippocampus stores feature conjunctions (as
opposed to individual features), features which appeared together at study are likely to
appear together at test. Importantly, in accordance with dual-process accounts of recognition memory (Yonelinas, 1994; Jacoby, Yonelinas, & Jennings, 1996), we believe that
hippocampally-driven recollection is not the sole contributor to recognition memory performance. Rather. extensive evidence exists that recollection is complemented by a "fallback" familiarity signal which participants consult when rich recollection does not occur.
The familiarity signal is mediated by as-yet unspecified areas (likely including the parahippocampal temporal cortex: Aggleton & Shaw, 1996; Miller & Desimone, 1994).
Our account differs substantially from most other computational and mathematical models
of recognition memory. Most of these models compute the "global match" between the
test probe and stored memories (e.g .? Hintzman, 1988; Gillund & Shiffrin, 1984); recollection in these models involves computing a similarity-weighted average of stored memory
patterns. In other memory models, recollection of an item depends critically on the extent
to which the components of the item's representation were linked with that of the study
context (e.g., Chappell & Humphreys, 1994). Critically, recollection in all of these models
lacks the high-threshold, high-quality character of recollection in our model. This is most evident when we consider the effects of manipulations which increase interference (e.g.,
increasing the length of the study list. or increasing inter-item similarity). As interference
increases, global matching models predict increasingly "blurry" recollection (reflecting the
contribution of more items to the composite output vector), while the other models predict
that false recollection of lures will increase. In contrast, our model predicts that increasing
interference should lead to decreased correct recollection of studied test probes, but there
should be no concomitant increase in "erroneous" types of recollection (i.e., recollection
of details which mismatch studied test probes, or rich recollection of lures). This prediction is consistent with the recent finding that correct recollection of studied items decreases
with increasing list length (Yonelinas, 1994). Lastly, although extant data certainly do not
contradict the claim that the veridicality of recollection is robust to interference, we acknowledge that additional, focused experimentation is needed to definitively resolve this
issue.
A Hippocampal Model ofRecognition Memory
/"
-
,\
75
/
I '---;----;
- -- -C~-:;~ _
1
L __ oL::..J'
'-
Figure I: The model. a) Shows the areas and connectivity, and the corresponding columns within
the Input, EC. and CAl (see text). b) Shows an example activity pattern. Note the sparse activity in
the DG and CA3, and intermediate sparseness of the CAL
2
Architecture and Overall Behavior
Figure I shows a diagram of our model, which contains the basic anatomical regions of
the hippocampal formation, as well as the entorhinal cortex (EC), which serves as the
primary cortical input/output pathway for the hippocampus. The model as described below
instantiates a series of hypotheses about the structure and function of the hippocampus and
associated cortical areas, which are based on anatomical and physiological data and other
models as described in O'Reilly and McClelland (1994) and McClelland et al. (1995), but
not elaborated upon significantly here.
The Input layer activity pattern represents the state of the EC resulting from the presentation
of a given item. We assume that the hippocampus stores and retrieves memories by way
of reduced representations in the EC, which have a correspondence with more elaborated
representations in other areas of cortex that is developed via long-term cortical learning.
We further assume that there is a rough topology to the organization of EC, with different
cortical areas and/or sub-areas represented by different slots, which can be thought of as
representing different feature dimensions of the input (e.g_, color, font, semantic features,
etc.). Our EC has 36 slots with four units per slot; one unit per slot was active (with each
unit representing a particular "feature value"). Input patterns were constructed from prototypes by randomly selecting different feature values for a random subset of slots. There
are two functionally distinct layers of the EC, one which receives input from cortical areas and projects into the hippocampus (superficial or ECin ), and another which receives
projections from the CAl and projects back out to the cortex (deep or ECout ). While the
representations in these layers are probably different in their details, we assume that they
are functionally equivalent, and use the same representations across both for convenience.
ECin projects to three areas of the hippocampus: the dentate gyrus (DO), area CA3, and
area CAL The storage of the input pattern occurs through weight changes in the feedforward and recurrent projections into the CA3, and the CA3 to CAl connections. The CA3
and CAl contain the two primary representations of the input pattern, while the DO plays
an important but secondary role as a pattern-separation enhancer for the CA3.
The CA3 provides the primary sparse, pattern-separated, conjunctive representation described above. This is achieved by random, partial connectivity between the EC and CA3,
and a high threshold for activation (i.e., sparseness), such that the few units which are activated in the CA3 (5% in our model) are those which have the most inputs from active EC
units. The odds of a unit having such a high proportion of inputs from even two relatively
similar EC patterns is low, resulting in pattern separation (see O'Reilly & McClelland,
76
R. C. O'Reilly, K. A. Norman and 1. L. McClelland
1994 for a much more detailed and precise treatment of this issue, and the role of the DO
in facilitating pattern separation). While these CA3 representations are useful for allowing
rapid learning without undue interference, the pattern-separation process eliminates any
systematic relationship between the CA3 pattern and the original EC pattern that gave rise
to it. Thus, there must be some means of translating the CA3 pattern back into the language
of the EC. The simple solution of directly associating the CA3 pattern with the corresponding EC pattern is problematic due to the interference caused by the relatively high activity
levels in the EC (around 15%, and 25% in our model). For this reason, we think that the
translation is formed via the CAl, which (as a result of long-term learning) is capable of
expanding EC representations into sparser patterns that are more easily linked to CA3, and
then mapping these sparser patterns back onto the EC.
Our CAl has separate representations of small combinations of slots (labeled columns);
columns can be arbitrarily combined to reproduce any valid EC representation. Thus, representations in CAl are intermediate between the fully conjunctive CA3, and the fully
combinatorial EC. This is achieved in our model by training a single CAl column of 32
units with slightly less than 10% activity levels to be able to reproduce any combination of
patterns over 3 ECin slots (64 different combinations) in a corresponding set of3 ECout
slots. The resulting weights are replicated across columns covering the entire EC (see Figure la). The cost of this scheme is that more CAl units are required (32 vs 12 per column
in the EC), which is nonetheless consistent with the relatively greater expansion of this area
relative to other hippocampal areas as a function of cortical size.
After learning, our model recollects studied items by simply reactivating the original CA3,
CAl and ECout patterns via facilitated weights. With partial or noisy input patterns (and
with interference), these weights and two forms of recurrence (the "short loop" within
CA3, and the "big loop" out to the EC and back through the entire hippocampus) allow
the hippocampus to bootstrap its way into recalling the complete original pattern (pattern
completion). If the EC input pattern corresponds to a nonstudied pattern, then the weights
will not have been facilitated for this particular activity pattern, and the CAl will not be
strongly driven by the CA3. Even if the ECin activity pattern corresponds to two components that were previously studied, but not together (see below), the conjunctive nature of
the CA3 representations will minimize the extent to which recall occurs.
Recollection is operationalized as successful recall of the test probe. This raises the basic
problem that the system needs to be able to distinguish between the ECout activation due to
the item input on ECin (either directly or via the CAl), and that which is due to activation
coming from recall in the CA3-CAl pathway. One solution to this problem, which is
suggested by autocorrelation histograms during reversible CA3 lesions (Mizumori et aI.,
1989), is that the CA3 and CAl are 180 0 out of phase with respect to the theta rhythm.
Thus, when the CA3 drives the CAl, it does so at a point when the CAl units would
otherwise be silent, providing a means for distinguishing between EC and CA3 driven CA 1
activation. We approximate something like this mechanism by simply turning off the ECin
inputs to CAl during testing. We assess the quality of hippocampal recall by comparing the
resulting ECout pattern with the ECin cue. The number of active units that match between
ECin and ECout (labeled C) indicates how much of the test probe was recollected. The
number of units that are active in ECout but not in ECin (labeled E) indicates the extent
to which the model recollected an item other than the test probe.
3
Activation and Learning Dynamics
Our model is implemented using the Leabra framework, which provides a robust mechanism for producing controlled levels of sparse activation in the presence of recurrent activa-
A Hippocampal Model ofRecognition Memory
77
tion dynamics, and a simple, effective Hebbian learning rule (O'Reilly, 1996)1. The activation function is a simple thresholded single-compartment neuron model with continuousvalued spike rate output. Membrane potential is updated by dVd't(t) = T L:c gc (t)gc (Ec Vm(t)), with 3 channels (c) corresponding to: e excitatory input; lleak current; and i inhibitory input. Activation communicated to other cells is a simple thresholded function of
the membrane potential: Yj(t) = 1/ (1 + 'Y[v>n(:)-9J+)' As in the hippocampus (and cortex), all principal weights (synaptic efficacies) are excitatory, while the local-circuit inhibition controls positive feedback loops (i.e., preventing epileptiform activity) and produces
sparse representations. Leabra assumes that the inhibitory feedback has an approximate
set-point (i.e., strong activity creates compensatorially stronger inhibition, and vice-versa),
resulting in roughly constant overall activity levels, with a firm upper bound. Inhibitory
current is given by gi = g~+l + q(gr - g~+l)' where 0 < q < 1 is typically .25, and
8
L:. 9c9c(Ec-8)
..
. '
.
9 =
ct? 8-Ei
for the UnIts With the k th and k + 1 th highest excitatory mputs.
A simple, appropriately normalized Hebbian rule is used in Leabra: f).wij = XiYj - YjWij,
which can be seen as computing the expected value of the sending unit's activity conditional on the receiver's activity (if treated like a binary variable active with probability Yj):
Wij ~ (xiIYj}p' This is essentially the same rule used in standard competitive learning or
mixtures-of-Gaussians.
4
Interference and List-Length, Item Similarity
Here, we demonstrate that the hippocampal recollection system degrades with increasing
interference in a way that preserves its essential high-threshold, high-quality nature. Figure 2 shows the effects of list length and item similarity on our C and E measures. Only
studied items appear in the high C, low E comer representing rich recollection. As length
and similarity increase, interference results in decreased C for studied items (without increased E), but critically there is no change in responding to new items. Interference in our
model arises from the reduced but nevertheless extant overlap between representations in
the hippocampal system as a function of item similarity and number of items stored. To the
extent that increasing numbers of individual CA3 units are linked to mUltiple contradictory
CAl representations, their contribution is reduced, and eventually recollection fails. As for
the frequently obtained finding that decreased recollection of studied items is accompanied
by an increase in overall false alarms, we think this results from subjects being forced to
rely more on the (less reliable) fallback familiarity mechanism.
5 Conjunctivity and Associative Recognition
Now, we consider what happens when lures are constructed by recombining elements of
studied patterns (e.g., study ''window-reason'' and "car-oyster", and test with "windowoyster"). One recent study found that participants are much more likely to claim to recollect studied pairs than re-paired lures (Yonelinas, 1997). Furthermore, data from this study
is consistent with the idea that re-paired lures sometimes trigger recollection of the studied word pairs that were re-combined to generate the lure; when this happens (assuming
that each word occurred in only one pair), the participant can confidently reject the lure.
Our simulation data is consistent with these findings: For studied word pairs, the model
(richly) recollected both pair components 86% of the time. As for re-paired lures, both pair
components were never recalled together, but 16% of the time the model recollected one
of the pair components, along with the component that it was paired with at study. The
I Note that the version of Leabra described here is an update to the cited version, which is currently
being prepared for publication.
78
R. C. O'Reilly, K. A. Nonnan and 1. L McClelland
Figure 2: Effects of list length and similarity on recollection perfonnance. Responses can be categorized according to the thresholds shown, producing three regions: rich recollection (RR), weak
recollection (WR), and misrecollection (MR). Increasing list length and similarity lead to less rich
recollection of studied items (without increasing misrecollection for these items), and do not significantly affect the model's responding to lures.
model responded in a similar fashion to pairs consisting of one studied word and a new
word (never recollecting both pair components together, but recollecting the old item and
the item it was paired with at study 13% of the time). Word pairs consisting of two new
items failed to trigger recollection of even a single pair component. Similar findings were
obtained in our simulation of the (Hintzman, Curran, & Oppy, 1992) experiment involving
recombinations of word and plurality cues.
6 Discussion
While the results presented above have dealt with the presentation of complete probe stimuli for recognition memory tests, our model is obviously capable of explaining cued recall
and related phenomena such as source or context memory by virtue of its pattern completion abilities. There are a number of interesting issues that this raises. For example, we
predict that successful item recollection will be highly correlated with the ability to recall additional information from the study episode, since both rely on the same underlying
memory. Further, to the extent that elderly adults form less distinct encodings of stimuli
(Rabinowitz & Ackerman, 1982), this explains both their impaired recollection on recognition tests (Parkin & Walter, 1992) and their impaired memory for contextual ("source")
details (Schacter et aI., 1991).
In summary, existing mathematical models of recognition memory are most likely incorrect
in assuming that recognition is performed with one memory system. Global matching models may provide a good account of familiarity-based recognition, but they fail to account for
the contributions of recollection to recognition, as discussed above. For example, global
matchil).g models predict that lures which are similar to studied items will always trigger
a stronger signal than dissimilar lures; as such, these models can not account for the fact
that sometimes subjects can reject similar lures with high levels of confidence (due, in our
model, to recollection ofa similar studied item; Brainerd, Reyna, & Kneer, 1995; Hintzman
et aI., 1992). Further, global matching models confound the signal for the extent to which
individual components of the test probe were present at all during study, and signal for the
A Hippocampal Model ofRecognition Memory
79
extent to which they occurred together. We believe that these signals may be separable,
with recollection (implemented by the hippocampus) showing sensitivity to conjunctions
of features, but not the occurrence of individual features, and familiarity (implemented by
cortical regions) showing sensitivity to component occurrence but not co-occurence. This
division of labor is consistent with recent data showing that familiarity does not discriminate well between studied item pairs and lures constructed by conjoining items from two
different studied pairs (so long as the pairings are truly novel) (Yonelinas, 1997), and with
the point, set forth by (McClelland et aI., 1995), that catastrophic interference would occur
if rapid learning (required to learn feature co-occurrences) took place in the neocortical
structures which generate the familiarity signal.
7 References
Aggleton, J. P., & Shaw, C. (1996). Amnesia and recognition memory: are-analysis of psychometric
data. Neuropsychologia, 34, 51.
Brainerd, C. J., Reyna, V. F., & Kneer, R. (1995). False-recognition reversal: When similarity is
distinctive. Journal ofMemory and Language, 34, 157-185.
Chappell, M., & Humphreys, M. S. (1994). An auto-associative neural network for sparse representations: Analysis and application to models of recognition and cued recall. Psychological Review,
101, 103-128.
Gillund, G., & Shiffrin, R. M. (1984). A retrieval model for both recognition and recall. Psychological
Review, 91, 1-67.
Hintzman, D. L. (1988). Judgments of frequency and recognition memory in a multiple-trace memory
model. Psychological Review, 95, 528-551.
Hintzman, D. L., Curran, T., & Oppy, B. (1992). Effects of similiarity and repetition on memory:
Registration without learning. Journal ofExperimental Psychology: Learning. Memory. and Cognition, 18, 667-680.
Jacoby, L. L., Yonelinas, A. P., & Jennings, J. M. (1996). The relation between conscious and
unconscious (automatic) influences: A declaration of independence. In J. D. Cohen, & J. W.
Schooler (Eds.), Scientific approaches to the question ofconsciousness (pp. 13-47). Hi1lsdale, NJ:
Lawrence Erbaum Associates.
McClelland, J. L., MCNaughton, B. L., & O'Reilly, R. C. (1995). Why there are complementary
learning systems in the hippocampus and neocortex: Insights from the successes and failures of
connectionst models of learning and memory. Psychological Review, 102,419-457.
Miller, E. K., & Desimone, R. (1994). Parallel neuronal mechanisms for short-term memory. Science,
263, 520--522.
Mizumori, S. J. Y., McNaughton, B. L., Barnes, C. A., & Fox, K. B. (1989). Preserved spatial coding
in hippocampal CAl pyramidal cells during reversible suppression ofCA3c output: Evidence for
pattern completion in hippocampus. Journal of NeuroSCience, 9( II), 3915-3928.
O'Reilly, R. C. (1996). The leabra model of neural interactions and learning in the neocortex. PhD
thesis, Carnegie Mellon University, Pittsburgh, PA, USA.
O'Reilly, R. C., & McClelland, J. L. (1994). Hippocampal conjunctive encoding, storage, and recall:
Avoiding a tradeoff. Hippocampus, 4(6), 661-682.
Parkin, A. J., & Walter, B. M. (1992). Recollective experience, normal aging, and frontal dysfunction.
Psychology and Aging, 7,290--298.
Rabinowitz, J. C., & Ackerman, B. P. (1982). General encoding of episodic events by elderly adults.
In F. I. M. C. S. Trehub (Ed.), Aging and cognitive processes. Plenum Publishing Corporation.
Schacter, D. L., Kaszniak, A. W., Kihlstrom, J. F., & Valdiserri, M. (1991). The relation between
source memory and aging. Psychology and Aging, 6, 559-568.
Yonelinas, A. P. (1994). Receiver-operating characteristics in recognition memory: Evidence for a
dual-process model. Journal ofExperimental Psychology: Learning. Memory. and Cognition, 20,
1341-1354.
Yonelinas, A. P. (1997). Recognition memory ROCs for item and associative information: The
contribution of recollection and familiarity. Memory and Cognition, 25,747-763.
| 1415 |@word version:2 stronger:2 hippocampus:18 proportion:1 integrative:1 simulation:2 initial:1 contains:1 series:1 selecting:1 efficacy:1 past:1 existing:2 current:2 comparing:1 contextual:1 activation:8 yet:1 conjunctive:4 must:1 update:1 v:1 cue:2 item:31 short:2 provides:2 parkin:2 sits:1 gillund:2 mathematical:2 along:1 constructed:3 amnesia:1 pairing:1 incorrect:1 consists:1 pathway:2 autocorrelation:1 cnbc:1 elderly:2 inter:1 expected:1 rapid:3 roughly:1 behavior:1 frequently:1 ol:1 continuousvalued:1 resolve:1 window:1 increasing:9 project:3 campus:1 underlying:1 circuit:1 what:1 unspecified:1 psych:1 substantially:1 developed:1 finding:4 corporation:1 nj:1 jlm:1 temporal:1 remember:1 ofa:1 veridically:1 control:1 unit:15 appear:2 producing:3 positive:1 accordance:1 local:1 aging:5 encoding:3 lure:15 studied:22 co:3 range:1 neuropsychological:1 testing:1 yj:2 oyster:1 differs:1 communicated:1 bootstrap:1 episodic:1 area:12 physiology:1 composite:1 reilly:12 matching:3 significantly:2 thought:1 projection:2 word:7 reject:2 confidence:1 convenience:1 onto:1 parahippocampal:1 cal:21 storage:2 context:2 influence:1 equivalent:1 center:1 focused:1 of3:1 rule:3 insight:1 importantly:1 mcnaughton:3 updated:1 plenum:1 unconscious:1 trigger:6 colorado:2 play:1 distinguishing:1 curran:2 hypothesis:1 harvard:2 pa:2 element:1 recognition:19 associate:1 predicts:1 labeled:3 role:2 region:3 episode:2 decrease:1 highest:1 environment:1 dynamic:2 raise:2 compromise:1 upon:1 creates:1 division:1 distinctive:1 basis:1 comer:1 easily:1 represented:1 retrieves:1 separated:2 distinct:3 forced:1 describe:1 effective:1 walter:2 mizumori:2 formation:1 firm:1 furthennore:1 otherwise:1 ability:3 gi:1 think:3 noisy:1 associative:3 obviously:1 rr:1 took:1 propose:1 interaction:1 coming:1 ackerman:2 relevant:1 loop:3 rapidly:1 shiffrin:2 forth:1 constituent:1 impaired:2 produce:2 cued:2 recurrent:2 completion:3 sole:1 strong:1 implemented:3 involves:1 anatomy:1 correct:2 human:1 translating:1 explains:1 plurality:1 biological:1 around:1 normal:1 lawrence:1 cognition:4 predict:4 dentate:1 claim:3 mapping:1 combinatorial:1 currently:1 infonnation:1 epileptiform:1 schooler:1 sensitive:1 contributor:1 vice:1 repetition:1 reflects:1 weighted:1 rough:1 always:1 rather:1 publication:1 conjunction:6 indicates:3 contrast:1 suppression:1 sense:2 schacter:2 unlikely:1 entire:2 typically:1 relation:2 reproduce:2 wij:2 issue:3 dual:2 overall:3 undue:2 spatial:1 never:3 having:2 represents:1 nearly:1 hippocampally:1 stimulus:2 few:1 randomly:1 dg:1 preserve:1 individual:4 familiar:1 phase:1 consisting:2 attempt:1 recalling:1 organization:1 highly:1 g_:1 certainly:1 mixture:1 truly:1 activated:1 desimone:2 capable:2 partial:2 experience:2 perfonnance:1 fox:1 old:1 re:4 psychological:4 increased:1 column:6 cost:1 ca3:25 subset:1 rare:1 successful:2 gr:1 stored:3 wjh:1 combined:2 cited:1 mechanistically:1 sensitivity:2 systematic:1 off:1 vm:1 together:7 extant:2 connectivity:2 thesis:1 opposed:2 slowly:1 cognitive:1 account:6 potential:2 accompanied:1 coding:1 caused:1 depends:2 tion:1 performed:1 apparently:1 linked:3 competitive:1 participant:4 parallel:1 elaborated:2 contribution:6 minimize:1 formed:1 ass:1 compartment:1 responded:1 largely:1 characteristic:2 miller:2 judgment:1 dealt:1 weak:1 critically:4 served:1 researcher:1 drive:1 explain:1 synaptic:1 ed:2 failure:1 nonetheless:1 frequency:1 pp:1 james:1 broadest:1 associated:1 treatment:1 richly:1 enhancer:1 recall:9 color:1 car:1 oreilly:1 reflecting:1 back:4 response:4 box:1 strongly:1 furthermore:1 lastly:1 receives:2 ei:1 overlapping:1 lack:1 reversible:2 fallback:2 quality:7 rabinowitz:2 scientific:1 believe:2 usa:1 effect:4 contain:2 normalized:1 norman:3 adequately:1 semantic:1 during:4 dysfunction:1 recurrence:1 covering:1 rhythm:1 hippocampal:17 evident:1 complete:2 demonstrate:1 neocortical:1 novel:2 cohen:1 operationalized:1 discussed:1 occurred:3 functionally:2 mellon:2 cambridge:1 versa:1 ai:4 automatic:1 language:2 cortex:5 similarity:9 inhibition:2 etc:1 operating:1 something:1 recent:3 driven:3 manipulation:1 store:2 binary:1 arbitrarily:1 success:1 yonelinas:8 captured:1 seen:1 additional:2 greater:1 mr:1 accomplishes:1 signal:8 ii:1 multiple:2 hebbian:2 match:2 long:3 retrieval:2 aggleton:2 paired:5 controlled:1 prediction:1 involving:1 basic:2 essentially:1 cmu:1 histogram:1 sometimes:3 achieved:2 cell:2 preserved:1 decreased:3 diagram:1 pyramidal:1 source:3 appropriately:1 eliminates:1 probably:1 subject:2 neuropsychologia:1 jacoby:2 odds:1 consult:1 unitary:1 presence:1 intermediate:2 feedforward:1 insofar:2 affect:1 independence:1 psychology:7 gave:1 architecture:1 topology:1 kirkland:1 associating:1 silent:1 idea:1 prototype:1 tradeoff:1 whether:1 deep:1 jennings:2 useful:1 detailed:1 amount:1 prepared:1 neocortex:3 conscious:1 mcclelland:12 reduced:3 gyrus:1 generate:2 conjoining:1 problematic:1 inhibitory:3 neuroscience:1 per:3 wr:1 anatomical:2 carnegie:2 key:1 four:1 threshold:6 nevertheless:1 chappell:2 thresholded:2 kenneth:1 registration:1 year:1 facilitated:2 place:1 separation:5 layer:3 bound:1 ct:1 distinguish:1 correspondence:1 encountered:1 barnes:1 activity:12 occur:2 dvd:1 separable:1 relatively:6 recombining:1 department:3 according:1 leabra:5 combination:3 instantiates:1 membrane:2 across:2 slightly:1 increasingly:1 character:1 happens:2 randall:1 explained:1 confound:1 boulder:2 interference:15 previously:1 eventually:1 mechanism:4 fail:1 needed:1 serf:1 sending:1 reversal:1 gaussians:1 experimentation:1 probe:10 blurry:1 occurrence:3 shaw:2 original:3 top:1 assumes:1 responding:2 publishing:1 recombination:1 question:1 occurs:3 blend:1 font:1 primary:3 spike:1 degrades:1 separate:1 recollection:44 street:1 extent:8 reason:2 assuming:2 length:7 relationship:1 concomitant:1 providing:1 trace:1 rise:1 allowing:1 upper:1 neuron:1 acknowledge:1 precise:1 gc:2 arbitrary:1 tive:1 pair:13 required:2 extensive:1 connection:1 recalled:1 recollect:2 adult:2 able:2 suggested:1 ofexperimental:2 below:2 pattern:38 mismatch:1 appeared:1 confidently:1 including:1 memory:36 reliable:1 event:2 overlap:2 treated:1 rely:2 turning:1 representing:3 scheme:1 theta:1 mediated:1 auto:1 occurence:1 text:1 prior:1 understanding:1 review:4 relative:2 fully:2 interesting:1 consistent:5 translation:1 excitatory:3 summary:1 free:1 allow:1 wide:1 explaining:1 definitively:1 sparse:5 feedback:2 dimension:1 cortical:10 valid:1 rich:9 preventing:1 replicated:1 ec:25 approximate:2 contradict:1 global:5 active:5 receiver:2 pittsburgh:2 why:1 learn:2 nature:3 robust:2 superficial:1 expanding:1 ca:1 channel:1 expansion:1 big:1 alarm:1 suffering:1 facilitating:1 lesion:1 categorized:1 body:1 complementary:1 neuronal:1 psychometric:1 roc:1 fashion:1 sub:2 fails:1 explicit:1 similiarity:1 learns:1 humphreys:2 erroneous:1 familiarity:9 specific:2 showing:4 list:6 physiological:1 virtue:1 evidence:4 exists:3 essential:1 false:4 recollecting:2 phd:1 entorhinal:1 sparseness:3 sparser:2 simply:2 likely:4 failed:1 labor:1 corresponds:2 complemented:1 ma:1 declaration:1 hintzman:5 slot:8 conditional:1 presentation:2 change:2 principal:1 contradictory:1 secondary:1 discriminate:1 catastrophic:1 la:1 rarely:1 arises:1 dissimilar:1 frontal:1 phenomenon:1 avoiding:1 correlated:1 |
457 | 1,416 | Correlates of Attention in a Model of
Dynamic Visual Recognition*
.
Rajesh P. N. Rao
Department of Computer Science
University of Rochester
Rochester, NY 14627
rao@cs.rochester.edu
Abstract
Given a set of objects in the visual field, how does the the visual system learn
to attend to a particular object of interest while ignoring the rest? How are
occlusions and background clutter so effortlessly discounted for when recognizing a familiar object? In this paper, we attempt to answer these questions in the context of a Kalman filter-based model of visual recognition that
has previously proved useful in explaining certain neurophysiological phenomena such as endstopping and related extra-classical receptive field effects in the visual cortex. By using results from the field of robust statistics,
we describe an extension of the Kalman filter model that can handle multiple
objects in the visual field. The resulting robust Kalman filter model demonstrates how certain forms of attention can be viewed as an emergent property of the interaction between top-down expectations and bottom-up signals. The model also suggests functional interpretations of certain attentionrelated effects that have been observed in visual cortical neurons. Experimental results are provided to help demonstrate the ability of the model
to perform robust segmentation and recognition of objects and image sequences in the presence of varying degrees of occlusions and clutter.
1 INTRODUCTION
The human visual system possesses the remarkable ability to recognize objects despite the
presence of distractors and occluders in the field of view. A popular suggestion is that an "attentional spotlight" mediates this ability to preferentially process a relevant object in a given
scene (see [5, 9] for reviews). Numerous models have been proposed to simulate the control of
this ..focus of attention" [10, 11, 15]. Unfortunately, there is inconclusive evidence for the existence of an explicit neural mechanism for implementing an attentional spotlight in the visual
*This research was supported by NIH/PHS research grant 1-P41-RR09283. I am grateful to Dana
Ballard for many useful discussions and suggestions. Author's current address: The Salk Institute, CNL,
10010 N. Torrey Pines Road, La Jolla, CA 92037. E-mail: rao@salk. edu .
.??..\.'..
'
Correlates ofAttention in a Model ofDynamic Visual Recognition
81
cortex. Thus, an important question is whether there are alternate neural mechanisms which
don't explicitly use a spotlight but whose effects can nevertheless be interpreted as attention.
In other words, can attention be viewed as an emergent property of a distributed network of
neurons whose primary goal is visual recognition?
In this paper, we extend a previously proposed Kalman filter-based model of visual recognition
[13, 12] to handle the case of multiple objects, occlusions, and clutter in the visual field. We
provide simulation results suggesting that certain forms of attention can be viewed as an emergent property of the interaction between bottom-up signals and top-down expectations during
visual recognition. The simulation results demonstrate how "attention" can be switched between different objects in a visual scene without using an explicit spotlight of attention.
2 A KALMAN FILTER MODEL OF VISUAL RECOGNITION
We have previously introduced a hierarchical Kalman filter-based model of visual recognition
and have shown how this model can be used to explain neurophysiological effects such as endstopping and neural response suppression during free-viewing of natural images [ 12, 13 ]. The
Kalman filter [7] is essentially a linear dynamical system that attempts to mimic the behavior
of an observed natural process. At any time instant t, the filter assumes that the internal state
of the given natural process can be represented as a k x 1 vector r(t). Although not directly
accessible, this internal state vector is assumed to generate ann x 1 measurable and observable
output vector I(t) (for example, an image) according to:
I(t)
= Ur(t) + n(t)
(1)
where U is ann x k generative (or measurement) matrix, and n(t) is a Gaussian stochastic
noise process with mean zero and a covariance matrix given by E = E[nilT] (E denotes the
expectation operator and T denotes transpose).
In order to specify how the internal state r changes with time, the Kalman filter assumes that
the process of interest can be modeled as a Gauss-Markov random process [1]. Thus, given
the state r(t- 1) at time instant t - 1, the next state r(t) is given by:
r(t)
= Vr(t- 1) + m(t- 1)
(2)
where Vis the state transition (or prediction) matrix and m is white Gaussian noise with mean
m = E[m] and covariance II= E[(m- m)(m- m)T].
Given the generative model in Equation 1 and the dynamics in Equation 2, the goal is to optimally estimate the current internal state r( t) using only the measurable inputs I( t). An optimization function whose minimization yields an estimate of r is the weighted least-squares
criterion:
=
J
(I- Ur)TE- 1 (I- Ur) + (r- r f M- 1 (r- r)
(3)
where r(t) is the mean of the state vector before measurement of the input data I(t) and M =
E[(r - r)(r - r)T] is the corresponding covariance matrix. It is easy to show [1] that J is
simply the sum of the negative log-likelihood of generating the data I given the stater, and
the negative log of the prior probability of the stater. Thus, minimizing J is equivalent to
maximizing the posterior probability p(rji) of the stater given the input data.
=
0 and solving for the minimum
The optimization function J can be minimized by setting ~J
valuer of the stater (note thatr equals the mean ofr afte~ measurement of I). The resultant
Kalman .filter equation is given by:
?
=
+ N(t)UTE(t)- 1 (I(t)- Ur(t))
(4)
= W(t- 1) + m(t- 1)
(5)
= (UTE(t)- 1 U + M(t)- 1 )-1 is a "normalization" matrix that maintains the
r(t)
r(t)
r(t)
where N(t)
covariance of the stater after measurement ofl. The matrix M, which is the covariance before
R. P. N. Rao
82
measurement of I, is updated as M (t) = V N (t - 1) VT +II (t - 1). Thus, the Kalman filter
predicts one step into the future using Equation 5, obtains the next sensory input I(t), and then
co~ects its Jj.rediction r(9 u~ing the sensory resi~ual e~or (I(t) --: Ur(t)) a~d ~e Kalman
gam N(t)U ~(t)- 1 ? This yields the corrected est1mate r(t) (Equation 4), wh1ch IS then used
to make the next state predictionf(t + 1).
The measurement (or generative) matrix U and the state transition (or prediction) matrix V
used by the Kalman filter together encode an internal nwdel of the observed dynamic process. As suggested in [13], it is possible to learn an internal model of the input dynamics
from observed data. Let u and v denote the vectorized forms of the matrices U and V respectively. For exam,fle, the n ~ k generative matrix U can be collapsed into an nk x 1 vector
u
[U1 U 2 ??. un] where U' denotes the ith row of U. Note that (I- Ur)
(I- Ru)
where R is then x nk matrix given by:
=
R
=[
r;.. r~
~...
.
0
0
=
l
(6)
rT
By minimizing an optimization function similar to J [13], one can derive a Kalman filter-like
..learning rule" for the generative matrix U:
u(t)
=
ii(t)
+ Nu(t)R(t)T~(t)- 1 (I(t)- R(t)ii(t))- aNu(t)ii(t)
(7)
where ii(t) = u(t- 1), Nu(t) = (Nu(t- 1)- 1 + R(t)T~(t)- 1 R(t) + ai)- 1 , and I is the
nk x nk identity matrix. The constant a determines the decay rate of ii.
As in the case of U, an estimate of the prediction matrix V can be obtained via the following
learning rule for v [13]:
v(t)
=
=
v(t)
+ Nv(t)R(tf M(t)- 1 [r(t + 1)- r(t + 1)]- f3Nv(t)v(t)
=
(8)
wherev(t)
v(t-1), Nv(t)
(Nv(t-1)- 1 +R(t)T M(t)- 1 R(t)+f3I)- 1 andRisak X k 2
matrix analogous to R (Equation 6) but with rT =
The constant /3 determines the decay
rate for v while I denotes the k 2 x k 2 identity matrix. Note that in this case, the estimate ofV is
corrected using the prediction residual error (r( t + 1) - r( t + 1)), which denotes the difference
between the actual state and the predicted state. One unresolved issue is the specification of
values for r(t) (comprising R(t)) in Equation 7 and r(t + 1) in Equation 8. The ExpectationMaximization (EM) algorithm [4] suggests that in the case of static stimuli (f(t) = r(t 1)), one may use r(t) = r which is the converged optimal state estimate for the given static
r( tl N), which is
input. In the case of dynamic stimuli, the EM algorithm prescribes r( t)
the optimal temporally snwothed state estimate [1] for timet (5 N), given input data for each
of the time instants 1, ... , N. Unfortunately, the smoothed estimate requires knowledge of
future inputs and is computationally quite expensive. For the experimental results, we used
the on-line estimates r(t) when updating the matrices and during training.
rr.
=
u
v
3 ROBUST KALMAN FILTERING
The standard derivation of the Kalman filter minimizes Equation 3 but unfortunately does not
specify how the covariance ~ is to be obtained. A common choice is to use a constant matrix
or even a constant scalar. Making ~ constant however reduces the Kalman filter estimates to
standard least-squares estimates,? which are highly susceptible to outliers or gross errors i.e.
data points that lie far away from the bulk of the observed or predicted data [6]. For example,
in the case where I represents an input image, occlusions and clutter will cause many pixels in
I to deviate significantly from corresponding pixels in the predicted image Ur. The problem
83
Correlates ofAJtention in a Model ofDynamic Visual Recognition
Gating
Matrix
Sensory
Residual
G
-
Feedforward
Matrix
uT
I- ltd
Input I
-
Nonnalization
Inhibition
Itd=Ur
Top-Down Prediction
of Expeeted I nput
N
Robust
Kalman Filter
Estimate
Prediction
r
Matrix
r0
"
v
-
~
Feedback
Matrix
Predicted State r
u
Figure 1: Recurrent Network Implementation of the Robust Kalman Filter. The gating matrix G is
a non-linear function of the current residual error between the input I and its top-down prediction ur. G
effectively filters out any high residuals, thereby preventing outliers in input data I from influencing the
robust Kalman filter estimate r. Note that the entire filter can be implemented in a recurrent neural network, with U, UT, and V represented by the synaptic weights of neurons with linear activation functions
and G being implemented by a set of threshold non-linear neurons with binary outputs.
of outliers can be tackled using robust estimation procedures [6] such as M-estimation, which
involves minimizing a function of the form:
n
(9)
i=1
where Ii and Ui are the ith pixel and ith row of I and U respectively, and p is a function that
increases less rapidly than the square. This reduces the influence oflarge residual errors (which
correspond to outliers) on the optimization of J', thereby "rejecting" the outliers. A special
case of the above function is the following weighted least squares criterion:
J' =(I-:- Urf S(I- Ur)
(10)
where Sis a diagonal matrix whose diagonal entries Si,i determine the weight accorded to the
corresponding pixel error (Ii - Uir). A simple but attractive choice for these weights is the
non-linear function given by Si,i =min {1, c/(Ii- Uir) 2 }, wherecisa threshold parameter.
To understand the behavior of this function, note that S effectively clips the ith summand in
J' (Equation 10 above) to a constant value c whenever the ith squared residual (Ii - Uir?
exceeds the threshold c; otherwise, the summand is set equal to the squared residual.
By substituting E- 1 = Sin the optimization function J (Equation 3), we can rederive the
following robust Kalman filter equation:
r(t) = r(t) + N(t)UTG(t)(I- Ur(t))
(11)
where r(t) = W(t- 1)) + iii(t- 1), N(t) = (UTG(t)U + M(t)- 1 ) - 1 , M(t) = V N(t1)VT + II(t -1), and G(t) is ann x n diagonal matrix whose diagonal entries at time instant
t are given by:
Gi,i _ { 0 if (Ii(t) - Uir(t)) 2 > c(t)
1 otherwise
G can be regarded as the sensory residual gain or "gating" matrix, which determines the (binary) gain on the various components of the incoming sensory residual error vector. By effectively filtering out any high residuals, G allows the Kalman filter to ignore the corresponding
outliers in the input I, thereby enabling it to robustly estimate the stater. Figure 1 depicts an
implementation of the robust Kalman filter in the form of a recurrent network of linear and
threshold non-linear neurons. In particular, the feedforward, feedback and prediction neurons
possess linear activation functions while the gating neurons implementing G compute binary
outputs based on a threshold non-linearity.
R. P.N.Rao
84
Training Objects
Robust Estimate
Input Image
(b)
(a)
Input Image
Outliers
Robust
Estimate I
Robust
Estimate 2
Outliers
I
Least Squares
Estimate
I.
'
(c)
Figure 2: Correlates of Attention during Static Recognition. (a) Images of size 105 x 65 used to train
a robust Kalman filter network. The generative matrix U was 6825 x 5. (b) Occlusions and background
clutter are treated as outliers (white regions in the third image, depicting the diagonal of the gating matrix
G). This allows the network to "attend to" and recognize the training object, as indicated by the accurate
reconstruction (middle image) of the training image based on the final robust state estimate. (c) In the
more interesting case of the training objects occluding each other, the network converges to one of the
objects (the "dominant" one in the image - in this case, the object in the foreground). Having recognized
one object, the second object is attended to and recognized by taking the complement of the outliers (diagonal of G) and repeating the robust filtering process (third and fourth images). The fifth image is the
image reconstruction obtained from the standard (least squares derived) Kalman filter estimate, showing
an inability to resolve or recognize either of the two objects.
4 VISUAL ATIENTION IN A SIMULATED NETWORK
The gating matrix G allows the Kalman filter network to "selectively attend.. to an object while
treating the remaining components of the sensory input as outliers. We demonstrate this capability of the network using three different examples. In the first example, a network was trained
on static grayscale images of a pair of 3D objects (Figure 2 (a)}. For learning static inputs, the
prediction matrix Vis unnecessary since we may use r(t) = r(t -1) and M(t) = N(t -1).
After training, the network was tested on images containing the training objects with varying
degrees of occlusion and clutter (Figure 2 (b) and (c)). The outlier threshold c was initialized
to the sum of the mean plus k standard deviations of the current distribution of squared residual
errors (Ii - Uir )2 ? The value of k was gradually decreased during each iteration in order to
allow the network to refine its robust estimate by gradually pruning away the outliers as it converges to a single object estimate. After convergence, the diagonal of the matrix G contains
zeros in the image locations containing the outliers and ones in the remaining locations. As
shown in Figure 2 (b), the network was successful in recognizing the training object despite
occlusion and background clutter.
More interestingly, the outliers (white) produce a crude segmentation of the occluder and background clutter, which can subsequently .be used to focus "attention" on these previously ignored objects and recover their identity. In particular, an outlier mask m can be defined by
taking the complement of the diagonal of G (i.e. m i = 1- Gi,i). By replacing the diagonal of
G with m in Equation tt 1 and repeating the estimation process, the network can "attend to"
1 Although not implemented here, this "shifting of attentional focus" can be automated using a model
of neuronal fatigue and active decay (see, for example, [3)).
Correlates ofAUention in a Model of Dynamic Visual Recognition
laput
85
~
~~~~~
.. J~r5' ~
-~
.
Oulllen
-----
~/'
~
---?--~-
(a)
??Ill
. . . I I???
?-
~
(b)
Inputs
Predictions
Predictions
Outliers
Outliers
(c)
(d)
Figure 3: Correlates of Attention during Dynamic Recognition. (a) A network was trained on a cyclic
image sequence of gestures (top), each image of size 75 x 75, with U and V of size 5625 x 15 and 15 x 15
respectively. The panels below show how the network can ignore various fonns of occlusion and clutter
(outliers}, "attending to" the sequence of gestures that it has been trained on. The outlier threshold c
was computed as the mean plus 0.3 standard deviati9ns of the current distribution of squared residual
errors. Results shown are those obtained after 5 cycles of exposure to the occluded images. (b) Three
image sequences used to train a network. (c) and (d) show the response of the network to ambiguous
stimuli comprised of images containing both a horizontal and a vertical bar. Note that the network was
trained on a horizontal bar moving downwards and?a vertical bar moving rightwards (see (b)) but not both
simultaneously. Given ambiguous stimuli containing both these stimuli, the network interprets the input
differently depending on the initial "priming" input. When the initial input is a vertical bar as in (c), the
network interprets the sequence as a vertical bar moving rightwards (with some minor artifacts due to
the other training sequences). On the other hand, when the initial input is a horizontal bar as in (d), the
sequence is interpreted as a horizontal bar moving downwards, not paying "attention" to the extraneous
vertical bars, which are now treated as outliers.
the image region(s) that were previously ignored as outliers. Such a two-step serial recognition process is depicted in Figure 2 (c). The network first recognizes the "doniinant" object,
which was generally observed to be the object occupying a larger area of the input image or
possessing regions with higher contrast.? The outlier mask m is subsequently used for "switching attention" and extracting the identity of the second object (lower arrow). Figure 3 shows
examples of attention during recognition of dynamic stimuli. In particular, Figure 3 (c) and
(d) show how the same image sequence can be interpreted in two different ways depending on
which part of the stimulus is "attended to," which in tum depends on the initial priming input.
5 CONCLUSIONS
The simulation results indicate that certain? experimental observations that have previously
been interpreted using the metaphor of an attentional spotlight can also arise as a result of competition and cooperation during visual recognition within networks of linear and non-linear
86
R. P. N. Rao
neurons. Although not explicitly designed to simulate attention, the robust Kalman filter networks nevertheless display some of the essential characteristics of visual attention, such as
the preferential processing of a subset of the input signals and the consequent "switching" of
attention to previously ignored stimuli. Given multiple objects or conflicting stimuli in their
receptive fields (Figures 2 and 3), the responses of the feedforward, .feedback, and prediction
neurons in the simulated network were modulated according to the current object being "attended to." The modulation in responses was mediated by the non-linear gating neurons G,
taking into account both bottom-up signals as well top-down feedback signals. This suggests a
network-level interpretation of similar forms of attentional response modulation in the primate
visual cortex [2, 8, 14], with the consequent prediction that the genesis of attentional modulation in such cases may not necessarily lie within the recorded neurons themselves but within
the distributed circuitry that these neurons are an integral part of.
References
[1] A.B. Bryson and Y.-C. Ho. Applied Optimtzl Control. New York: John Wiley, 1975.
[2] L. Chelazzi, E.K. Miller, J. Duncan, and R. Desimone. A neural basis for visual search
in inferior temporal cortex. Nature, 363:345-347, 1993.
[3] P. Dayan. An hierarchical model of visual rivalry. In M. Mozer, M. Jordan, and
T. Petsche, editors, Advances in Neural Information Processing Systems 9, pages 48-54.
Cambridge, MA: MIT Press, 1997.
[4] A.P. Dempster, N.M. Laird, andD.B. Rubin. Maximum likelihood from incomplete data
via the EM algorithm. J. Royal Statistical Society Series B, 39:1-38, 1977.
[5] R. Desimone and J. Duncan. Neural mechanisms of selective visual attention. Annual
Review of Neuroscience, 18:193-222,1995.
[6] P.J. Huber. Robust Statistics. New York: John Wiley, 1981.
[7] R.E. Kalman. A new approach to linear filtering and prediction theory. Trans. ASME J.
Basic Eng., 82:35-45, 1960.
[8] J. Moran and R. Desimone. Selective attention gates visual processing in the extrastriate
cortex. Science, 229:782-784, 1985.
[9] W.T. Newsome. Spotlights, highlights and visual awareness. Current Biology, 6(4):357360, 1996.
[10] E. Niebur and C. Koch. Control of selective visual attention: Modeling the "where" pathway. In D. Touretzky, M. Mozer, and M. Hasselmo, editors, Advances in Neural lnformtztion Processing Systems 8, pages 802-808. Cambridge, MA: MIT Press, 1996.
[11] B.A. Olshausen, D.C. Van Essen, and C.H. Anderson. A neurobiological model of visual attention and invariant pattern recognition based on dynamic routing of information.
Journal of Neuroscience, 13:4700-4719, 1993.
[12] R.P.N. Rao and D.H. Ballard. The visual cortex as a hierarchical predictor. Technical
Report 96.4, National Resource Laboratory for the Study of Brain and Behavior, Department of Computer Science, University of Rochester, September 1996.
[13] R.P.N. Rao and D.H. Ballard. Dynamic model of visual recognition predicts neural response properties in the visual cortex. Neural Computation, 9(4):721-763, 1997.
[14] S. Treue and J.H.R. Maunsell. Attentional modulation of visual motion processing in
cortical areas MT and MST. Nature, 382:539-541, 1996.
[15] J .K. Tsotsos, S.M. Culhane, W. Y.K. Wai, Y. Lai, N. Davis, and F. Nuflo. Modeling visual
attention via selective tuning. Artificial Intelligence, 78:507~545, 1995.
| 1416 |@word middle:1 simulation:3 covariance:6 eng:1 attended:3 thereby:3 extrastriate:1 initial:4 cyclic:1 contains:1 series:1 interestingly:1 current:7 activation:2 si:3 john:2 mst:1 treating:1 designed:1 generative:6 intelligence:1 ith:5 location:2 ect:1 pathway:1 huber:1 mask:2 behavior:3 themselves:1 brain:1 occluder:1 discounted:1 resolve:1 actual:1 metaphor:1 provided:1 linearity:1 panel:1 interpreted:4 minimizes:1 temporal:1 demonstrates:1 control:3 grant:1 fle:1 maunsell:1 before:2 t1:1 attend:4 occluders:1 influencing:1 switching:2 despite:2 modulation:4 plus:2 suggests:3 co:1 procedure:1 area:2 significantly:1 word:1 road:1 nonnalization:1 operator:1 context:1 collapsed:1 influence:1 measurable:2 equivalent:1 maximizing:1 exposure:1 attention:22 rule:2 attending:1 regarded:1 handle:2 analogous:1 updated:1 recognition:18 expensive:1 updating:1 predicts:2 bottom:3 observed:6 region:3 cycle:1 gross:1 mozer:2 dempster:1 ui:1 occluded:1 dynamic:10 prescribes:1 grateful:1 solving:1 trained:4 ofv:1 basis:1 emergent:3 differently:1 represented:2 various:2 derivation:1 train:2 culhane:1 describe:1 artificial:1 whose:5 quite:1 larger:1 cnl:1 otherwise:2 ability:3 statistic:2 gi:2 torrey:1 laird:1 final:1 sequence:8 timet:1 rr:1 reconstruction:2 interaction:2 unresolved:1 relevant:1 rapidly:1 thatr:1 competition:1 convergence:1 produce:1 generating:1 converges:2 object:28 help:1 derive:1 exam:1 recurrent:3 depending:2 minor:1 expectationmaximization:1 paying:1 implemented:3 c:1 predicted:4 involves:1 indicate:1 filter:27 stochastic:1 subsequently:2 human:1 routing:1 viewing:1 implementing:2 extension:1 effortlessly:1 koch:1 itd:1 circuitry:1 pine:1 substituting:1 estimation:3 hasselmo:1 tf:1 occupying:1 weighted:2 minimization:1 mit:2 gaussian:2 varying:2 treue:1 encode:1 derived:1 focus:3 likelihood:2 contrast:1 suppression:1 bryson:1 am:1 dayan:1 entire:1 selective:4 comprising:1 pixel:4 issue:1 ill:1 extraneous:1 special:1 field:7 equal:2 having:1 biology:1 represents:1 r5:1 foreground:1 mimic:1 minimized:1 future:2 stimulus:9 oflarge:1 summand:2 report:1 simultaneously:1 recognize:3 national:1 familiar:1 occlusion:8 attempt:2 interest:2 highly:1 essen:1 accurate:1 rajesh:1 integral:1 desimone:3 preferential:1 incomplete:1 initialized:1 modeling:2 rao:8 newsome:1 deviation:1 entry:2 subset:1 predictor:1 comprised:1 recognizing:2 successful:1 optimally:1 answer:1 accessible:1 together:1 squared:4 recorded:1 containing:4 rji:1 suggesting:1 account:1 explicitly:2 vi:2 depends:1 view:1 recover:1 maintains:1 capability:1 rochester:4 square:6 characteristic:1 miller:1 yield:2 correspond:1 rejecting:1 niebur:1 converged:1 explain:1 touretzky:1 whenever:1 synaptic:1 wai:1 resultant:1 static:5 gain:2 proved:1 popular:1 distractors:1 knowledge:1 ut:2 segmentation:2 tum:1 endstopping:2 higher:1 response:6 specify:2 anderson:1 hand:1 horizontal:4 replacing:1 artifact:1 indicated:1 olshausen:1 effect:4 laboratory:1 white:3 attractive:1 sin:1 during:8 inferior:1 ambiguous:2 davis:1 criterion:2 fatigue:1 asme:1 tt:1 demonstrate:3 motion:1 image:26 possessing:1 nih:1 common:1 resi:1 functional:1 mt:1 extend:1 interpretation:2 spotlight:6 measurement:6 cambridge:2 ai:1 tuning:1 ute:2 specification:1 moving:4 cortex:7 inhibition:1 dominant:1 posterior:1 jolla:1 certain:5 binary:3 vt:2 ofr:1 minimum:1 r0:1 recognized:2 determine:1 signal:5 ii:14 multiple:3 reduces:2 ing:1 exceeds:1 technical:1 gesture:2 lai:1 serial:1 prediction:14 basic:1 fonns:1 essentially:1 expectation:3 iteration:1 normalization:1 ofl:1 background:4 decreased:1 extra:1 rest:1 posse:2 nv:3 jordan:1 extracting:1 presence:2 feedforward:3 iii:1 easy:1 automated:1 f3i:1 interprets:2 whether:1 ltd:1 york:2 cause:1 jj:1 ignored:3 useful:2 generally:1 clutter:9 repeating:2 ph:1 clip:1 generate:1 neuroscience:2 bulk:1 nevertheless:2 threshold:7 rediction:1 tsotsos:1 sum:2 fourth:1 duncan:2 tackled:1 display:1 refine:1 chelazzi:1 annual:1 scene:2 u1:1 simulate:2 nuflo:1 min:1 uir:5 department:2 according:2 alternate:1 em:3 ur:11 making:1 primate:1 outlier:23 gradually:2 invariant:1 computationally:1 equation:13 resource:1 previously:7 mechanism:3 accorded:1 gam:1 hierarchical:3 away:2 rr09283:1 petsche:1 robustly:1 ho:1 gate:1 existence:1 top:6 assumes:2 denotes:5 remaining:2 recognizes:1 instant:4 classical:1 society:1 question:2 nput:1 receptive:2 primary:1 rt:2 diagonal:9 september:1 attentional:7 simulated:2 mail:1 ru:1 kalman:27 modeled:1 minimizing:3 preferentially:1 unfortunately:3 susceptible:1 negative:2 implementation:2 perform:1 vertical:5 neuron:12 observation:1 markov:1 enabling:1 genesis:1 smoothed:1 introduced:1 complement:2 pair:1 conflicting:1 mediates:1 nu:3 trans:1 address:1 suggested:1 bar:8 dynamical:1 below:1 pattern:1 royal:1 shifting:1 natural:3 ual:1 treated:2 residual:12 numerous:1 temporally:1 mediated:1 deviate:1 review:2 prior:1 highlight:1 suggestion:2 interesting:1 filtering:4 dana:1 remarkable:1 switched:1 awareness:1 degree:2 vectorized:1 rubin:1 editor:2 row:2 cooperation:1 supported:1 free:1 transpose:1 allow:1 understand:1 institute:1 explaining:1 taking:3 fifth:1 rivalry:1 distributed:2 van:1 feedback:4 cortical:2 transition:2 sensory:6 author:1 preventing:1 far:1 correlate:6 pruning:1 observable:1 obtains:1 ignore:2 neurobiological:1 active:1 incoming:1 assumed:1 unnecessary:1 don:1 grayscale:1 un:1 search:1 learn:2 ballard:3 robust:19 ca:1 nature:2 ignoring:1 depicting:1 priming:2 necessarily:1 arrow:1 noise:2 arise:1 neuronal:1 tl:1 depicts:1 downwards:2 ny:1 salk:2 vr:1 wiley:2 explicit:2 lie:2 crude:1 rederive:1 third:2 down:5 gating:7 showing:1 moran:1 decay:3 consequent:2 evidence:1 inconclusive:1 essential:1 effectively:3 p41:1 te:1 anu:1 nk:4 depicted:1 simply:1 neurophysiological:2 visual:35 scalar:1 determines:3 ma:2 viewed:3 goal:2 identity:4 ann:3 change:1 corrected:2 experimental:3 la:1 gauss:1 occluding:1 selectively:1 internal:6 inability:1 modulated:1 tested:1 phenomenon:1 rightwards:2 |
458 | 1,417 | S-Map: A network with a simple
self-organization algorithm for generative
topographic mappings
Kimmo Kiviluoto
Laboratory of Computer and
Information Science
Helsinki University of Technology
P.O. Box 2200
FIN-02015 HUT, Espoo, Finland
Kimmo.KiviluotoChut.fi
Erkki Oja
Laboratory of Computer and
Information Science
Helsinki University of Technology
P.O. Box 2200
FIN-02015 HUT, Espoo, Finland
Erkki.OjaChut.fi
Abstract
The S-Map is a network with a simple learning algorithm that combines the self-organization capability of the Self-Organizing Map
(SOM) and the probabilistic interpretability of the Generative Topographic Mapping (GTM). The simulations suggest that the SMap algorithm has a stronger tendency to self-organize from random initial configuration than the GTM. The S-Map algorithm
can be further simplified to employ pure Hebbian learning, without changing the qualitative behaviour of the network.
1
Introduction
The self-organizing map (SOM; for a review, see [1]) forms a topographic mapping
from the data space onto a (usually two-dimensional) output space. The SOM has
been succesfully used in a large number of applications [2]; nevertheless, there are
some open theoretical questions, as discussed in [1, 3]. Most of these questions
arise because of the following two facts: the SOM is not a generative model, i.e. it
does not generate a density in the data space, and it does not have a well-defined
objective function that the training process would strictly minimize.
Bishop et al. [3] introduced the generative topographic mapping (GTM) as a solution
to these problems. However, it seems that the GTM requires a careful initialization
to self-organize. Although this can be done in many practical applications, from a
theoretical point of view the GTM does not yet offer a fully satisfactory model for
natural or artificial self-organizing systems.
K. Kiviluoto and E. Oja
550
In this paper, we first briefly review the SOM and GTM algorithms (section 2); then
we introduce the S-Map, which may be regarded as a crossbreed of SOM and GTM
(section 3); finally, we present some simulation results with the three algorithms
(section 4), showing that the S-Map manages to combine the computational simplicity and the ability to self-organize of the SOM with the probabilistic framework
of the GTM.
2
2.1
SOM and GTM
The SOM algorithm
e
The self-organizing map associates each data vector t with that map unit that has
its weight vector closest to the data vector. The activations 11! of the map units are
given by
~ = {I, when IIJLi
11,
0, otherwise
- etll < IIJLj - etll,
'Vj
#i
(1)
where JLi is the weight vector of the ith map unit Ci , i = 1, . . . , K. Using these
activations, the SOM weight vector update rule can be written as
K
JLj
:=
JLj
+ tSt L
11: h(Ci' Cj ; j3t)(e t - JLj)
(2)
i=l
Here parameter tSt is a learning rate parameter that decreases with time. The neighborhood function h(Ci' Cj ; j3t) is a decreasing function of the distance between map
units Ci and Cj ; j3t is a width parameter that makes the neighborhood function get
narrower as learning proceeds. One popular choice for the neighborhood function
is a Gaussian with inverse variance j3t.
2.2
The GTM algorithm
In the GTM algorithm, the map is considered as a latent space, from which a
nonlinear mapping to the data space is first defined. Specifically, a point C in the
latent space is mapped to the point v in the data space according to the formula
L
v(C; M)
= MtP(C) = L ?j(C)JLj
(3)
j=l
where tP is a vector consisting of L Gaussian basis functions, and M is a D x L
matrix that has vectors ILj as its columns, D being the dimension of the data space.
The probability density p( C) in the latent space generates a density to the manifold
that lies in the data space and is defined by (3). IT the latent space is of lower
dimension than the data space, the manifold would be singular, so a Gaussian noise
model is added. A single point in the latent space generates thus the following
density in the data space:
j3)D/2 exp [j3
p(eIC;M,j3)= ( 211'
-2'llv(C;M)-eW ]
(4)
where j3 is the inverse of the variance of the noise.
The key point of the GTM is to approximate the density in the data space by
assuming the latent space prior p( C) to consist of equiprobable delta functions that
S-Map
551
form a regular lattice in the latent space. The centers Ci of the delta functions are
called the latent vectors of the GTM, and they are the GTM equivalent to the SOM
map units. The approximation of the density generated in the data space is thus
given by
1 K
p(eIM,m = K
(5)
LP(elCi;M,,8)
i=l
The parameters of the GTM are determined by minimizing the negative log likelihood error
f(M,,8) = -
t. [~ t.
ln
(6)
P ({'IC,;M,,8)]
over the set of sample vectors {e t }. The batch version of the GTM uses the EM
algorithm [4]; for details, see [3]. One may also resort to an on-line gradient descent
procedure that yields the GTM update steps
K
1L~+1 := IL~ +
6t ,8t
L lI:(M t ,,8t)<pj(Ci)[et - v(Ci;M t )]
(7)
i=l
fJH'
:=
i3' + 6' [ ~
t. ~hM',
(J')lle' -
v(C,; MI)II'
-
2~ ]
(8)
where 11: (M,,8) is the GTM counterpart to the SOM unit activation, the posterior
probability p(Cilet ; M,,8) of the latent vector Ci given data vector et :
lIf(M,,8) = p(Cilet ;M,,8)
p(etICi; M,,8)
- L:f.=1 p(etICi'; M,,8)
(9)
exp[-~IIV(Ci;M)
2.3
_e t Il 2 ]
Connections between SOM and GTM
Let us consider a GTM that has an equal number of latent vectors and basis functions!, each latent vector Ci being the center for one Gaussian basis function <Pi(C).
Latent vector locations may be viewed as units of the SOM, and consequently the
basis functions may be interpreted as connection strengths between the units. Let
us use the shorthand notation <P~ == <Pj(Ci). Note that <P~ = <frt, and assume that
the basis functions be normalized so that L:~l <P~
At the zero-noise limit, or when ,8
~ 00,
= L:~1 <P~ = 1.
the softmax activations of the GTM given
in (9) approach the winner-take-all function (1) of the SOM. The winner unit Cc(t)
e
for the data vector t is the map unit that has its image closest to the data vector,
so that the index c( t) is given by
c(t)
= ar~in Ilv(C,) - e'll = ~in
(t, 4>;1';) -
{t
(10)
INote that this choice serves the purpose of illustration only; to use GTM properly,
one should choose much more latent vectors than basis functions.
552
K. Kiviluoto and E. Oja
The GTM weight update step (7) then becomes
IL~+1 := IL~
+ ~t?j(t) ret -
v(c(t); Mt)]
(11)
This resembles the variant of SOM, in which the winner is searched with the rule (10)
and weights are updated as
IL~+1 := ILj +~t?j(t)(et - IL;)
(12)
Unlike the original SOM rules (1) and (2), the modified SOM with rules (10)
and (12) does minimize a well-defined objective function: the SOM distortion measure [5, 6, 7, 1]. However, there is a difference between GTM and SOM learning
rules (11) and (12). With SOM, each individual weight vector moves towards the
data vector, but with GTM, the image ofthe winnerlatent vector v(c(t); M) moves
towards the data vector, and all weight vectors ILj move to the same direction.
For nonzero noise, when 0 < f3 < 00, there is more difference between GTM and
SOM: with GTM, not only the winner unit but activations from other units as well
contribute to the weight update.
3
S-Map
Combining the softmax activations of the GTM and the learning rule of the SOM,
we arrive at a new algorithm: the S-Map.
3.1
The S-Map algorithm
The S-Map resembles a GTM with an equal number of latent vectors and basis
functions. The position of the ith unit on the map is is given by the latent vector
(i; the connection strength of the unit to another unit j is ?), and a weight vector
ILi is associated with the unit. The activation of the unit is obtained using rule (9).
The S-Map weights learn proportionally to the activation of the unit that the weight
is associated with, and the activations of the neighboring units:
1'1+1 ,= 1'; +6'
(t. ?iij:) (e' -1';)
(13)
which can be further simplified to a fully Hebbian rule, updating each weight proportionally to the activation of the corresponding unit only, so that
IL}+l := ILj
+ ~t7]~ (e t -
ILj)
(14)
The parameter f3 value may be adjusted in the following way: start with a small
value, slowly increase it so that the map unfolds and spreads out, and then keep
increasing the value as long as the error (6) decreases. The parameter adjustment
scheme could also be connected with the topographic error of the mapping, as
proposed in [9] for the SOM.
Assuming normalized input and weight vectors, the "dot-product metric" form of
the learning rules (13) and (14) may be written as
1')+1 ,= 1'] + 6'
(t. ?;ij:)
(I -1']1'?)e'
(15)
S-Map
553
and
JL~+1 := 1')
+ 6t1Jj(I - JL)JLf)e t
(16)
respectively; the matrix in the second parenthesis keeps the weight vectors normalized to unit length, assuming a small value for the learning rate parameter 6t [8].
The dot-product metric form of a unit activity is
~' =
(L:~1 ?)I't {t]
L: exp [13 (Lf=1 1>;' JLj) et ]
exp [,8
T
==1
l
(17)
which approximates the posterior probability p('ile t ;M, 13) that the data vector
were generated by that specific unit. This is based on the observation that if the
data vectors {e t } are normalized to unit length, the density generated in the data
space (unit sphere in RD) becomes
..
-1
normallzmg
p({IC,; M,,8) = ( constant)
x exp [ ,8
3.2
K
f;.
(
i
?jl'j )
T
{
]
(18)
S-Map algorithm minimizes the GTM error function in
dot-product metric
The GTM error function is the negative log likelihood, which is given by (6) and is
reproduced here:
(19)
When the weights are updated using a batch version of (15), accumulating the
updates for one epoch, the expected value of the error [4] for the unit (i is
T
= - LpOld(ilet;M,m In[p?ewU:i)pOeW(e t l(i;M,!3)]
E(?rW)
t=1 '
T
=- L
'"
T/~Id.t
1J~ld,t 13
(
t==1
K
-'---==1!K
1>] JLjew
L
) T
(20)
e t + terms not involving the weight vectors
j==1
The change of the error for the whole map after one epoch is thus
K
E(?oew - fOld)
=- L
T
K
L L 1J~ld,t131>;(JLjew - JLjldfe t
i=1 t=1 j=1
= -
p.s
t, (~ ~ q~ld" ?;{t)
,
'"
T
(I - I'lld 1''Jld T)
-'
aT
J
(t, t, q~ld,t' ?j'(" )
'",
~
K
= -136
L[uf
Uj -
(uf JLjld)2] ~ 0
j=1
with equality only when the weights are already in the error minimum.
-'(21)
554
4
K. Kiviluoto and E. Oja
Experimental results
The self-organization ability of the SOM, the GTM, and the S-Map was tested on
an artificial data set: 500 points from a uniform random distribution in the unit
square.
The initial weight vectors for all models were set to random values, and the final
configuration of the map was plotted on top of the data (figure 1). For all the algorithms, the batch version was used. The SOM was trained as recommended in [1] in two phases, the first starting with a wide neighborhood function, the second with
a narrow neighborhood. The GTM was trained using the Matlab implementation
by Svensen, following the recommendations given in [to]. The S-Map was trained in
two ways: using the "full" rule (13), and the simplified rule (14) . In both cases, the
parameter {3 value was slowly increased every epoch; by monitoring the error (6) of
the S-Map (see the error plot in the figure) the suitable value for {3 can be found.
In the GTM simulations, we experimented with many different choices for basis
function width and their number, both with normalized and unnormalized basis
functions. It turned out that GTM is somewhat sensitive to these choices: it had
difficulties to unfold after a random initialization, unless the basis functions were set
so wide (with respect to the weight matrix prior) that the map was well-organized
already in its initial configuration. On the other hand, using very wide basis functions with the GTM resulted in a map that was too rigid to adapt well to the data.
We also tried to update the parameter {3 according to an annealing schedule, as
with the S-Map, but this did not seem to solve the problem.
r ."
"
. ..... : , .. :.
f .
~:
~. .
.. ,
n'
... . . 7 '
..
"
~.
r? ? . .. .. ':
l~ ~..:. ' ' ~'. ,':'
[
. " ,'-.
..
.
I'
.'
. .: :.
:
..
'
:.
:t;jj ,{ .~.. ".. :. .'
f .f.:t+1.
.....
.
il'
.t}'.
hS*
~:_I;r7.
..
r??
I.
.~
,.
.'
: ~"" "
..t'>i
~:
.
: H ~'
? ~? . ';I..r.'
'~.'
.~ .
,.-'
?I.n
~.
Figure 1: Random initialization (top left), SOM (top middle), GTM (top right),
"full" S-Map (bottom left), simplified S-Map (bottom middle), On bottom right,
the S-Map error as a function of epochs is displayed; the parameter {3 was slightly
increased every epoch, which causes the error to increase in the early (unfolding)
phase of the learning, as the weight update only minimizes the error for a given {3.
5
Conclusions
The S-Map and SOM seem to have a stronger tendency to self-organize from random
initialization than the GTM. In data analysis applications, when the GTM can
S-Map
555
be properly initialized, SOM, S-Map, and GTM yield comparable results; those
obtained using the latter two algorithms are also straightforward to interpret in
probabilistic terms. In Euclidean metric, the GTM has the additional advantage
of guaranteed convergence to some error minimum; the convergence of the S-Map
in Euclidean metric is still an open question. On the other hand, the batch G TM
is computationally clearly heavier per epoch than the S-Map, while the S-Map is
somewhat heavier than the SOM.
The SOM has an impressive record of proven applications in a variety of different
tasks, and much more experimenting is needed for any alternative method to reach
the same level of practicality. SOM is also the basic bottom-up procedure of selforganization in the sense that it starts from a minimum of functional principles
realizable in parallel neural networks. This makes it hard to analyze, however.
A probabilistic approach like the GTM stems from the opposite point of view by
emphasizing the statistical model, but as a trade-off, the resulting algorithm may
not share all the desirable properties of the SOM. Our new approach, the S-map,
seems to have succeeded in inheriting the strong self-organization capability of the
SOM, while offering a sound probabilistic interpretation like the GTM.
References
[1] T. Kohonen, Self-Organizing Maps. Springer Series in Information Sciences 30,
Berlin Heidelberg New York: Springer, 1995.
[2] T. Kohonen, E. Oja, O. Simula, A. Visa, and J. Kangas, "Engineering applications of the self-organizing map," Proceedings of the IEEE, vol. 84, pp. 13581384, Oct. 1996.
[3] C. M. Bishop, M. Svensen, and C. K. I. Williams, "GTM: A principled alternative to the self-organizing map," in Advances in Neural Information Processing
Systems (to appear) (M. C. Mozer, M. I. Jordan, and T . Petche, eds.), vol. 9,
MIT Press, 1997.
[4] A. P. Dempster, N. M. Laird, and D. B. Rubin, "Maximum likelihood from incomplete data via the EM algorithm," Journal of the Royal Statistical Society,
vol. B 39, no. 1, pp. 1-38, 1977.
[5] S. P. Luttrell, "Code vector density in topographic mappings," Memorandum
4669, Defense Research Agency, Malvern, UK, 1992.
[6] T. M. Heskes and B. Kappen, "Error potentials for self-organization," in Proceedings of the International Conference on Neural Networks (ICNN'99), vol. 3,
(Piscataway, New Jersey, USA), pp. 1219-1223, IEEE Neural Networks Council, Apr. 1993.
[7] S. P. Luttrell, "A Bayesian analysis of self-organising maps," Neural Computation, vol. 6, pp. 767-794, 1994.
[8] E. Oja, "A simplified neuron model as a principal component analyzer," Journal of Mathematical Biology, vol. 15, pp. 267-273, 1982.
[9] K. Kiviluoto, "Topology preservation in self-organizing maps," in Proceedings
of the International Conference on Neural Networks (ICNN'96), vol. 1, (Piscataway, New Jersey, USA), pp. 294-299, IEEE Neural Networks Council, June
1996.
[10] M. Svensen, The GTM toolbox - user's guide. Neural Computing Research
Group / Aston University, Birmingham, UK, 1.0 ed., Oct. 1996. Available at
URL http://neural-server . aston. ac. uk/GTM/MATLAB..Impl. html.
| 1417 |@word h:1 version:3 briefly:1 jlf:1 stronger:2 seems:2 middle:2 selforganization:1 open:2 simulation:3 tried:1 ld:4 kappen:1 initial:3 configuration:3 series:1 t7:1 offering:1 activation:10 yet:1 written:2 plot:1 update:7 generative:4 ith:2 record:1 contribute:1 location:1 organising:1 mathematical:1 qualitative:1 shorthand:1 combine:2 introduce:1 expected:1 decreasing:1 increasing:1 becomes:2 notation:1 interpreted:1 minimizes:2 ret:1 every:2 impl:1 uk:3 unit:27 appear:1 organize:4 engineering:1 limit:1 id:1 initialization:4 resembles:2 succesfully:1 practical:1 lf:1 procedure:2 unfold:1 regular:1 suggest:1 ilet:1 get:1 onto:1 accumulating:1 equivalent:1 map:48 center:2 straightforward:1 williams:1 starting:1 simplicity:1 pure:1 rule:11 regarded:1 jli:1 memorandum:1 updated:2 user:1 us:1 associate:1 jlj:5 simula:1 updating:1 bottom:4 connected:1 decrease:2 trade:1 principled:1 mozer:1 dempster:1 agency:1 iiv:1 trained:3 basis:11 gtm:46 jersey:2 artificial:2 neighborhood:5 solve:1 distortion:1 otherwise:1 ability:2 topographic:6 laird:1 final:1 reproduced:1 advantage:1 product:3 eic:1 luttrell:2 neighboring:1 turned:1 combining:1 kohonen:2 organizing:8 convergence:2 ac:1 svensen:3 ij:1 strong:1 direction:1 behaviour:1 icnn:2 adjusted:1 strictly:1 hut:2 considered:1 ic:2 exp:5 mapping:7 finland:2 early:1 ilv:1 purpose:1 birmingham:1 sensitive:1 council:2 unfolding:1 mit:1 clearly:1 gaussian:4 i3:1 modified:1 june:1 properly:2 likelihood:3 experimenting:1 sense:1 realizable:1 rigid:1 html:1 espoo:2 lif:1 softmax:2 equal:2 f3:2 biology:1 r7:1 employ:1 equiprobable:1 oja:6 resulted:1 individual:1 phase:2 consisting:1 organization:5 fjh:1 succeeded:1 unless:1 incomplete:1 euclidean:2 initialized:1 plotted:1 theoretical:2 increased:2 column:1 ar:1 tp:1 lattice:1 uniform:1 too:1 density:8 international:2 probabilistic:5 off:1 ilj:5 choose:1 slowly:2 smap:1 resort:1 li:1 potential:1 view:2 analyze:1 start:2 capability:2 parallel:1 minimize:2 il:8 square:1 variance:2 yield:2 ofthe:1 bayesian:1 manages:1 monitoring:1 cc:1 reach:1 ed:2 pp:6 associated:2 mi:1 popular:1 cj:3 organized:1 schedule:1 done:1 box:2 hand:2 nonlinear:1 usa:2 normalized:5 counterpart:1 equality:1 laboratory:2 satisfactory:1 nonzero:1 ll:1 self:18 width:2 unnormalized:1 image:2 fi:2 functional:1 mt:1 winner:4 jl:3 discussed:1 interpretation:1 approximates:1 interpret:1 rd:1 heskes:1 analyzer:1 had:1 dot:3 impressive:1 closest:2 posterior:2 server:1 minimum:3 additional:1 somewhat:2 recommended:1 preservation:1 ii:1 full:2 desirable:1 sound:1 stem:1 hebbian:2 adapt:1 offer:1 long:1 sphere:1 parenthesis:1 j3:4 variant:1 ile:1 involving:1 basic:1 metric:5 annealing:1 singular:1 unlike:1 seem:2 jordan:1 variety:1 topology:1 opposite:1 tm:1 heavier:2 defense:1 url:1 york:1 cause:1 jj:1 matlab:2 proportionally:2 rw:1 generate:1 http:1 eim:1 delta:2 per:1 vol:7 group:1 key:1 nevertheless:1 tst:2 changing:1 pj:2 inverse:2 arrive:1 comparable:1 guaranteed:1 fold:1 activity:1 strength:2 helsinki:2 erkki:2 generates:2 uf:2 according:2 piscataway:2 slightly:1 em:2 lld:1 visa:1 lp:1 inote:1 ln:1 computationally:1 needed:1 serf:1 available:1 batch:4 alternative:2 original:1 top:4 practicality:1 uj:1 society:1 objective:2 move:3 question:3 added:1 already:2 gradient:1 distance:1 mapped:1 berlin:1 manifold:2 assuming:3 length:2 code:1 index:1 illustration:1 minimizing:1 negative:2 implementation:1 observation:1 neuron:1 fin:2 descent:1 displayed:1 kangas:1 introduced:1 toolbox:1 connection:3 narrow:1 proceeds:1 usually:1 interpretability:1 royal:1 suitable:1 natural:1 jld:1 difficulty:1 scheme:1 technology:2 aston:2 hm:1 review:2 prior:2 epoch:6 fully:2 proven:1 rubin:1 principle:1 pi:1 share:1 llv:1 guide:1 lle:1 wide:3 dimension:2 unfolds:1 simplified:5 approximate:1 keep:2 latent:15 learn:1 heidelberg:1 som:34 vj:1 inheriting:1 did:1 apr:1 spread:1 whole:1 noise:4 arise:1 malvern:1 iij:1 position:1 lie:1 formula:1 emphasizing:1 bishop:2 specific:1 showing:1 experimented:1 consist:1 ci:11 adjustment:1 recommendation:1 springer:2 oct:2 viewed:1 narrower:1 consequently:1 careful:1 towards:2 change:1 hard:1 specifically:1 determined:1 principal:1 called:1 ili:1 tendency:2 experimental:1 ew:1 searched:1 latter:1 tested:1 |
459 | 1,418 | An Incremental Nearest Neighbor
Algorithm with Queries
Joel Ratsaby?
N.A.P. Inc.
Hollis, New York
Abstract
We consider the general problem of learning multi-category classification from labeled examples. We present experimental results for
a nearest neighbor algorithm which actively selects samples from
different pattern classes according to a querying rule instead of the
a priori class probabilities. The amount of improvement of this
query-based approach over the passive batch approach depends on
the complexity of the Bayes rule. The principle on which this algorithm is based is general enough to be used in any learning algorithm which permits a model-selection criterion and for which the
error rate of the classifier is calculable in terms of the complexity
of the model.
1
INTRODUCTION
We consider the general problem of learning multi-category classification from labeled examples. In many practical learning settings the time or sample size available
for training are limited. This may have adverse effects on the accuracy of the resulting classifier. For instance, in learning to recognize handwritten characters typical
time limitation confines the training sample size to be of the order of a few hundred
examples. It is important to make learning more efficient by obtaining only training
data which contains significant information about the separability of the pattern
classes thereby letting the learning algorithm participate actively in the sampling
process. Querying for the class labels of specificly selected examples in the input
space may lead to significant improvements in the generalization error (cf. Cohn,
Atlas & Ladner, 1994, Cohn, 1996). However in learning pattern recognition this
is not always useful or possible. In the handwritten recognition problem, the computer could ask the user for labels of selected patterns generated by the computer
-The author's coordinates are: Address: Hamered St. #2, Ra'anana, ISRAEL. Eaail:
jer@ee. technion. ac.il
An Incremental Nearest Neighbor Algorithm with Queries
613
however labeling such patterns are not necessarily representative of his handwriting
style but rather of his reading recognition ability. On the other hand it is possible to let the computer (learner) select particular pattern classes, not necessarily
according to their a priori probabilities, and then obtain randomly drawn patterns
according to the underlying unknown class-conditional probability distribution. We
refer to such selective sampling as sample querying. Recent theory (cf. Ratsaby,
1997) indicates that such freedom to select different classes at any time during the
training stage is beneficial to the accuracy of the classifier learnt. In the current
paper we report on experimental results for an incremental algorithm which utilizes
this sample-querying procedure .
2
THEORETICAL BACKGROUND
We use the following setting: Given M distinct pattern classes each with a class
conditional probability density fi (x), 1 ::; i ::; M, x E IRd, and a priori probabilities
Pi, 1 ::; i ::; M. The functions fi (x), 1 ::; i ::; M, are assumed to be unknown while
the Pi are assumed to be known or easily estimable as is the case oflearning character
recognition. For a sample-size vector m = [m1' ... , mM] where L~1 mj = m denote
by (m = {( x j , Yj ) } 1 a sample of labeled examples consisting of mi example from
pattern class i where Yj, 1 ::; j ::; m, are chosen not necessarily at random from
{I, 2 .... , M}, and the corresponding Xj are drawn at random i.i.d. according to the
class conditional probability density fy) (x). The expected misclassification error of
a classifier c is referred to as the loss of c and is denoted by L( c ). It is defined as the
probability of miselassification of a randomly drawn x with respect to the underlying
mixture probability density function f(x) = L~l pdi(X). The loss is commonly
represented as L(c) = El{x :c(x);iy(x)}, where l{xEA} is the indicator function of a set
A, expectation is taken with respect to the joint probability distribution fy (x )p(y)
where p(y) is a discrete probability distribution taking values Pi over 1 ::; i ::; M,
while y denotes the label of the class whose distribution fy(x) was used to draw x.
T=
The loss L(e) may also be written as L(e) = Lf!l PiEi1{c(x);ii} where Ei denotes
expectation with respect to fi(X) . The pattern recognition problem is to learn based
on
the optimal classifier, also known as the Bayes classifier, which by definition
has minimum loss whkh we denote by L * .
em
A multi-category classifier c is represented as a vector e(x) = [el(x), . .. , CM(X)]
of boolean classifiers, where Ci (x) = 1 if e( x) = i, and Ci (x) = 0 otherwise, 1 ::;
i ::; M. The loss L(e) of a multi-category classifier c may then be expressed as
the average of the losses of its component classifiers, i.e., L(e) = L~l PiL(ei)
where for a boolean classifier ei the loss is defined as L(ed = Ed{c.(x);il}' As
an estimate of L(e) we define the empirical loss Lm(c} = L~l p;Lm.(e) where
Lm,(c) = ~, Lj :Y1=i l{c(x);ii} which may also can be expressed as Lm,(ci)
=
~,
Lj:YJ=i l{c.(x);il}'
The family of all classifiers is assumed to be decomposed into a multi-structure
5
51 X 52 X .. , X 5 M , where 5 j is a nested structure (cf. Vapnik, 1982) of
boolean families Bk). ' ji = 1,2, ... , for 1 ::; i ::; M, i.e., 51
Bkl , Bk 2 , ?? ? ,Bk)1 ' ... ,
52 = BkllBk2,???,Bk12'oo" up to 5M = Bk1 ,Bk2 ,? .. ,Bk) M ' ? ' " where ki, E71+
denotes the VC-dimension of BkJ , and Bk 1 , ~ Bk1,+I' 1 ::; i ::; M . For any fixed
positive integer vector j E 7l ~ consider the class of vector classifiers 1i k(j) =
Bk1 1
x .Bk
x? .. X Bk 1M == 1ik where we take the liberty in dropping the multil2
index j and write k instead of k(j) . Define by (h the subfamily of 1ik consisting
=
=
J. Ratsaby
614
of classifiers C that are well-defined , i.e., ones whose components Ci, 1 S; i S; M
satisfy Ut!I{X: Ci(X)
I}
IRd and {x : Ci(X)
l}n{x: Cj(x)
I}
0, for
1 S; i =1= j S; M .
= =
=
= =
From the Vapnik-Chervonenkis theory (cf. Vapnik , 1982, Devroye, Gyorfi & Lugosi , 1996) it follows that the loss of any boolean classifier Ci E Bk ], is , with
high confidence, related to its empirical loss as L( Ci ) S Lm. (Ci) + f( m i, kj ,) where
f(mi' kj ,) = const Jkj,ln mi!mi , 1 S; i S; M , where henceforth we denote by
const any constant which does not depend on the relevant variables in the expression. Let the vectors m = [ml, "" mM] and k == k(j) = [kil , . ??, kjM] in 'lh~.
Define f(m, k) = 2:f'!1 Pif(mi ,kj,) . It follows that the deviation between the empirical loss and the loss is bounded uniformly over all multi-category classifiers in
a class (}k by f(m, k) . We henceforth denote by c k the optimal classifier in (}k , i.e.,
ck = argmincE~h L( c) and Ck = argmincEQk Lm (c) is the empirical loss minimizer
over the class (}k.
The above implies that the classifier Ck has a loss which is no more than L( c~) +
f(m, k) . Denote by k* the minimal complexity of a class (}k which contains the
Bayes classifier. We refer to it as the Bayes complexity and henceforth assume
k: < 00, 1 S; i S; M. If k* was known then based on a sample of size m with a
sample size vector m = [ml , "" mM] a classifier Ck o whose loss is bounded from
above by L * + f( m, k*) may be determined where L * = L( c~o) is the Bayes loss.
This bound is minimal with respect to k by definition of k* and we refer to it as the
minimal criterion. It can be further minimized by selecting a sample of size vector
m* = argmin{ 'WM . "\,,,M
_ _ }f(m,k*). This basically says that more examples
mEaJ+ '6,=1 m,_m
should be queried from pattern classes which require more complex discriminating
rules within the Bayes classifier. Thus sample-querying via minimization of the
minimal criterion makes learning more efficient through tuning the subsample sizes
to the complexity of the Bayes classifier. However the Bayes classifier depends
on the underlying probability distributions which in most interesting scenarios are
unknown thus k* should be assumed unknown . In (Ratsaby, 1997) an incremental
learning algorithm, based on Vapnik's structural risk minimization, generates a
random complexity sequence ken) , corresponding to a sequence of empirical loss
minimizers ck(n) over (}k(n), which converges to k* with increasing time n for learning
problems with a zero Bayes loss. Based on this, a sample-query rule which achieves
the same minimization is defined without the need to know k*. We briefly describe
the main ideas next.
At any time n, the criterion function is c(-, ken)) and is defined over the m-domain
'lhtt? A gradient descent step of a fixed size is taken to minimize the current criterion. After a step is taken , a new sample-size vector men + 1) is obtained and
the difference m( n + 1) - m( n) dictates the sample-query at time n, namely, the
increment in subsample size for each of the M pattern classes. With increasing n
the vector sequence men) gets closer to an optimal path defined as the set which
is comprised of the solutions to the minimization of f( m, k*) under all different
constraints of 2:~1 mi = m, where m runs over the positive integers. Thus for
all large n the sample-size vector m( n) is optimal in that it minimizes the minimal
cri terion f(', k*) for the current total sample size m( n). This consti tutes the samplequerying procedure of the learning algorithm. The remaining part does empirical
loss minimization over the current class (}k(n) and outputs ck(n)" By assumption,
since the Bayes classifier is contained in (}k o , it follows that for all large n, the loss
L(ck(n? S; L* + min{mE~ :2::1 m,=m'(n)} f(m, k*), which is basically the minimal
criterion mentioned above. Thus the algorithm produces a classifier ck(n) with a
An Incremental Nearest Neighbor Algorithm with Queries
615
minimal loss even when the Bayes complexity k* is unknown.
In the next section we consider specific modf'l classes consisting of nearest-nf'ighbor
classifiers on which we implement this incremental learning approach.
3
INCREMENTAL NEAREST-NEIGHBOR
ALGORITHM
Fix and Hodges, cf. Silverman & Jones (1989). introduced the simple but powerful
nearest-neighbor classifier which based on a labeled training sample {(Xj,yd}i!:I'
Xi E m,d, Yi E {I, 2, ... , M}, when given a pattern x, it outputs the label Yj corresponding to the example whose x j is closest to x. Every example in the training
sample is used for this decision (we denote such an example as a prototype) thus
the empirical loss is zero. The condensed nearest-neighbor algorithm (Hart, 1968)
and the reduced nearest neighbor algorithm (G ates, 1972) are procedures which
aim at reducing the number of prototypes while maintaining a zero empirical loss.
Thus given a training sample of size m, after running either of these procedures, a
nearest neighbor classifier having a zero empirical loss is generated based on s ~ m
prototypes. Learning in this manner may be viewed as a form of empirical loss
minimization with a complexity regularization component which puts a penalty
proportional to the number of prototypes.
A cell boundary ej,j of the voronoi diagram (cf. Preparata & Shamos, 1985)
corresponding to a multi-category nearest-neighbor classifier c is defined as the
(d - 1)-dimensional perpendicular-bisector hyperplane <?f the line connecting the
x-component of two prototypes Xi and Xj. For a fixed I E {1, ... ,M}, the collection
of voronoi cell-boundaries based on pairs of prototypes of the form (xi,/), (Xj,q)
where q =1= I, forms the boundary which separates the decision region labeled I from
its complement and represents the boolean nearest-neighbor classifier CI. Denote
by kl the number of suc.h cell-boundaries and denote by SI the number of prototypes from a total of ml examples from pattern class t. The value of kl may be
calculated directly from the knowledge of the SI prototypes, 1 ~ I ~ M, using
various algorithms. The boolean classifier Cl is an element of an infinite class of
boolean classifiers based on partitions of m,d by arrangements of kl hyperplanes of
dimensionality d - 1 where each of the cells of a partition is labeled either 0 or 1.
It follows, cf. Devroye et. al. (1996), that the loss of a multi-category nearestneighbor classifier C which consists of 81 prototypes out of ml examples, 1 ~ I ~ M,
is bounded as L(c) ~ Lm(c) + f(m, k), where the a priori probabilities are taken as
known, m
[mI, ... ,mM)' k
[k I , ... ,kM] and f(m,k)
E~lPlf(ml,kl)' where
f( ml, kz) = const ? d + 1 )kl In ml + (ekd d)d) / mi . Letting k* denote the Bayes
complexity then f(-, k*) represents the minimal criterion.
=
J
=
=
The next algorithm uses the Condense and Reduce procedures in order to generate a
sequence of classifiers ck(n) with a complexity vector k( n) which tends t.o k* as n --+
00. A sample-querying procedure referred to as Greedy Query (GQ) chooses at any
time n to increment the single subsample of pattern class j*(n) where mjO(n) is the
direction of maximum descent of the criterion f(', k( n)) at the current sample-size
vector m( n). For the part of the algorithm which utilizes a Delaunay-Triangulation
procedure we use the fast Fortune's algorithm (cf. 0 'Rourke) which can be used
only for dimensionality d = 2. Since all we are interested is in counting Voronoi
borders between all adjacent Voronoi cells then an efficient computation is possible
also for dimensions d > 2 by resorting to linear programming for computing the
adjacencies of facets of a polyhedron, d. Fukuda (1997).
1. Ratsaby
616
Incremental Nearest Neighbor (INN) Algorithm
Initialization: (Time n = 0)
Let increment-size t::. be a fixed small positive integer. Start with m(O) =
[e, . .. , e], where e is a small posit.ive integer. Draw (m(o) = {(m](o)}?";l where
(m)(O ) consists of mJ(O) randomly drawn i.i.d. examples from pattern class j.
While (number of available examples 2: t::.) Do:
1. Call Procedure CR: chIn ) = CR?(m(n? .
2. Call Procedure GQ: m(n
3. n:= n
+ 1) =
GQ(n).
+ 1.
End While
//Used up all examples .
Output: NN-cIassifier ck(n).
Procedure Condense-Reduce (CR)
Input: Sample (m(n) stored in an array A[] of size m(n).
Initialize: Make only the first example A[I] be a prototype.
//Condense
Do:
ChangeOccl.lred := FALSE.
For i= 1, . .. , m(n):
? Classify A[i] based on available prototypes using the NN-Rule.
? If not correct then
- Let A[i] be a prototype.
- ChangeOcel.lred:= TRUE.
? End If
End For
While ( ChangeOecl.lred).
//Reduce
Do:
ChangeOccl.lred := FALSE.
For i = 1, ... . m(n):
? If A[ i] is a prototype then classify it using the remaining prototypes by the
NN-Rule.
? If correct then
- Make A[i] be not a prototype.
- ChangeOccl.lred := TRUE.
? End If
End For
While ( ChangeOec'Ured) .
Run Delaunay-Triangulation Let k(n) = [k~, ... , kM ], k. denotes the number
of Voronoi-cell boundaries associated with the s, prototypes.
Return (NN-classifier with complexity vector k(n?.
Procedure Greedy-Query (GQ)
Input: Time n.
j*(n)
:=
I
argmaxl~J~M a!) f(m, k(n?1
Im(n)
Draw: t::. new i.i.d. examples from class j?(n). Denote them by ( .
Update Sample: (m]?(n)(n+l) := Cn)?(n)Cn) U (, while (m,Cn+l) := (m,Cn), for
1 :S i:;i: j?(n) :S M .
Return: (m(n)+ t::. eJ.Cn?, where eJ is an all zero vector except 1 at jth element.
An Incremental Nearest Neighbor Algorithm with Queries
3.1
617
EXPERIMENTAL RESULTS
We ran algorithm INN on several two-dimensional (d = 2) multi-category classification problems and compared its generalization error versus total sample size m
with that of batch learning, the latter uses Procedure CR (but not Procedure GQ)
with uniform subsample proportions, i.e., mi = ~, 1 ~ i ~ M.
We ran three classification problems consisting of 4 equiprobable pattern classes
with a zero Bayes loss. The generalization curves represent the average of 15 independent learning runs of the empirical error on a fixed size test set. Each run (both
for INN and Batch learning) consists of 80 independent experiments where each
differs by 10 in the sample size used for training where the maximum sample size is
800. We call an experiment a success if INN results in a lower generalization error
than Batch. Let p be the probability of INN beating Batch. We wish to reject the
hypothesis H that p
~ which says that INN and Batch are approximately equal
in performance. The results are displayed in Figure 1 as a series of pairs, the first
picture showing the pattern classes of the specific problem while the second shows
the learning curves for the two learning algorithms. Algorithm INN outperformed
the simple Batch approach with a reject level of less than 1%, the latter ignoring the
inherent Bayes complexity and using an equal subsample size for each of the pattern
classes. In contrast, the INN algorithm learns, incrementally over time, which of
the classes are harder to separate and queries more from these pattern classes.
=
References
Cohn D., Atlas L., Ladner R. (1994), Improving Generalization with Active Learning. Machine Learning, Vol 15, p.201-221.
Devroye L., Gyorfi L. Lugosi G. (1996). "A Probabilistic Theory of Pattern Recognition", Springer Verlag.
Fukuda K. (1997). Frequently Asked Questions in Geometric Computation.
Technical report, Swiss Federal Institute of technology, Lausanne. Available at
ftp://ftp.ifor.ethz.ch/pub/fukuda/reports.
Gates, G. W. (1972) The Reduced Nearest Neighbor Rule. IEEE Trans.
Theo., p.431-433.
Info.
Hart P. E. (1968) The Condensed Nearest Neighbor Rule. IEEE Trans. on Info.
Thea., Vol. IT-14, No.3.
O'rourke J . (1994). "Computational Geometry in C". Cambridge University Press.
Ratsaby, J. (1997) Learning Classification with Sample Queries.
Electrical Engineering Dept., Technion, CC PUB #196.
Available at URL
http://www.ee.technion.ac.il/jer/iandc.ps.
Rivest R. L., Eisenberg B. (1990), On the sample complexity of pac-learning using
random and chosen examples. Proceedings of the 1990 Workshop on Computational
Learning Theory, p. 154-162, Morgan Kaufmann, San Maeto, CA.
B. W. Silverman and M. C. Jones . E. Fix and J. 1. Hodges (1951): An important contribution to nonparametric discriminant analysis and density estimationcommentary on Fix and Hodges (1951). International statistical review, 57(3),
p.233-247, 1989.
Vapnik V.N., (1982), "Estimation of Dependences Based on Empirical Data",
Springer-Verlag. Berlin.
618
J. Ratsaby
n.2
~"
__- - - I ' - - - - " " I - - - - r - - - - '
n.1
~,
il '
\
. ~ 0.1
\
,
.~ ~. I
~
\.
',.
8
o.0K
~":-.
!t~
0.06
0.04
I
I
100
1:-0
0.0
PoltcmCI3.~S I
0
100
PaltcmCla.... 2
PoltcmCl3.'i.'i.l
'..00
200
fb-
'""'"
~OO
",
400
..: 500
' .
600
"'-,
,'
.
700
ROO
Tolal number of """"",I....
Balch
l'a"cmCJa.~s4
I(ICl(
. ..... "<..,
.
:!OO
....... INN
0.25
O.
~ 0.17
..,;
.~ 0.15
.~ O. IJ
;;
Iic 0.1
C 0.011
-~""''' ............. ....... -.....:""," .
0.05
(.<>,)
l'aU<:m('I_1
PaucmCl_2
0.03
l'aUc:mCIass.~
Cit(
0
l'aUcmCl-.l
100
2110
JOn
400
SIlO
TnUl numhcr of """"",lea
Balch
-(r
CoX
w'Ic
.?~1~<
: .,
I(
IC(
. , ';;'{
.Itc>fC"
i.8cl"?
0.26
\?qllc
.
~ 0.24
,
~
I(~ ? I
~PC( icc
~
,(. xi/(o( I(
0
, ><'V-'~t.
'S .:: (# (
~"'x
0
O.J
0.2
., ....
'~ 'J
1 - (II,
INN
~-
t.f.
~::?.t"<i
....
~:I(
_-.:-c
~
100
PanemCI ..... 1
",,,,. PanemClass2
PananCI..
1(( PanemCl.....4
....u
~.>...-,...".
~
I~
. ~ 0.22
."?
J
0.2
"ii
0.1
?
200
0. 16
0. 14
0.12
0.1
0
400
ToI.tl numher of cxamplCli
Batch
-- INN
Figure 1. Three different Pattern Classification Problems and Learning
Curves of the INN-Algorithm compared to Batch Learning.
(1110
70n
1100
| 1418 |@word cox:1 briefly:1 proportion:1 km:2 thereby:1 harder:1 contains:2 series:1 selecting:1 chervonenkis:1 pub:2 terion:1 current:5 si:2 written:1 partition:2 atlas:2 update:1 greedy:2 selected:2 jkj:1 hyperplanes:1 ik:2 calculable:1 consists:3 manner:1 expected:1 ra:1 frequently:1 multi:9 decomposed:1 increasing:2 underlying:3 bounded:3 rivest:1 israel:1 cm:1 argmin:1 minimizes:1 every:1 nf:1 classifier:35 positive:3 engineering:1 tends:1 path:1 yd:1 lugosi:2 approximately:1 initialization:1 au:1 nearestneighbor:1 lausanne:1 limited:1 perpendicular:1 gyorfi:2 practical:1 yj:4 implement:1 lf:1 silverman:2 differs:1 swiss:1 procedure:13 empirical:12 reject:2 dictate:1 confidence:1 get:1 selection:1 put:1 risk:1 www:1 rule:8 array:1 his:2 coordinate:1 increment:3 user:1 programming:1 us:2 hypothesis:1 element:2 recognition:6 labeled:6 electrical:1 region:1 ran:2 mentioned:1 complexity:13 asked:1 depend:1 argmaxl:1 learner:1 easily:1 joint:1 jer:2 represented:2 various:1 distinct:1 fast:1 describe:1 query:11 labeling:1 shamos:1 whose:4 iic:1 ive:1 say:2 roo:1 otherwise:1 ability:1 fortune:1 sequence:4 inn:12 gq:5 relevant:1 p:1 produce:1 incremental:9 converges:1 ftp:2 oo:3 ac:2 ij:1 nearest:16 implies:1 liberty:1 direction:1 posit:1 correct:2 vc:1 adjacency:1 require:1 fix:3 generalization:5 im:1 mm:4 ic:2 lm:7 achieves:1 estimation:1 outperformed:1 condensed:2 label:4 minimization:6 federal:1 always:1 aim:1 rather:1 ck:10 ej:3 cr:4 improvement:2 pif:1 polyhedron:1 indicates:1 contrast:1 pdi:1 cri:1 el:2 minimizers:1 voronoi:5 nn:4 lj:2 selective:1 selects:1 condense:3 interested:1 classification:6 denoted:1 priori:4 initialize:1 equal:2 having:1 sampling:2 represents:2 jones:2 jon:1 minimized:1 report:3 preparata:1 inherent:1 few:1 equiprobable:1 randomly:3 recognize:1 geometry:1 consisting:4 freedom:1 joel:1 mixture:1 pc:1 closer:1 lh:1 theoretical:1 minimal:8 instance:1 classify:2 boolean:7 facet:1 oflearning:1 deviation:1 hundred:1 technion:3 comprised:1 uniform:1 stored:1 learnt:1 chooses:1 st:1 density:4 international:1 ured:1 discriminating:1 probabilistic:1 connecting:1 iy:1 hodges:3 henceforth:3 style:1 return:2 actively:2 inc:1 satisfy:1 depends:2 wm:1 bayes:14 start:1 pil:1 contribution:1 minimize:1 il:5 accuracy:2 kaufmann:1 balch:2 handwritten:2 basically:2 cc:1 bk1:3 ed:2 definition:2 associated:1 mi:9 handwriting:1 icl:1 ask:1 knowledge:1 ut:1 dimensionality:2 rourke:2 cj:1 stage:1 hand:1 ei:3 cohn:3 incrementally:1 effect:1 true:2 kjm:1 regularization:1 adjacent:1 during:1 auc:1 criterion:8 chin:1 passive:1 fi:3 kil:1 ji:1 m1:1 significant:2 refer:3 cambridge:1 queried:1 tuning:1 resorting:1 delaunay:2 closest:1 recent:1 triangulation:2 bkj:1 scenario:1 verlag:2 success:1 yi:1 morgan:1 minimum:1 ii:4 technical:1 hart:2 itc:1 expectation:2 represent:1 cell:6 lea:1 background:1 consti:1 diagram:1 toi:1 integer:4 call:3 ee:2 structural:1 counting:1 enough:1 xj:4 reduce:3 idea:1 prototype:16 cn:5 expression:1 url:1 penalty:1 ird:2 york:1 useful:1 amount:1 nonparametric:1 ifor:1 s4:1 category:8 ken:2 reduced:2 generate:1 http:1 cit:1 discrete:1 write:1 dropping:1 vol:2 drawn:4 run:4 powerful:1 family:2 utilizes:2 draw:3 decision:2 ki:1 bound:1 constraint:1 generates:1 min:1 according:4 beneficial:1 ate:1 em:1 character:2 separability:1 ratsaby:7 taken:4 ln:1 know:1 letting:2 end:5 available:5 permit:1 batch:9 lhtt:1 gate:1 denotes:4 remaining:2 cf:8 running:1 maintaining:1 fukuda:3 const:3 arrangement:1 question:1 dependence:1 gradient:1 separate:2 berlin:1 participate:1 me:1 fy:3 discriminant:1 numher:1 devroye:3 index:1 info:2 unknown:5 ladner:2 descent:2 displayed:1 y1:1 bk:8 introduced:1 namely:1 pair:2 complement:1 kl:5 trans:2 address:1 pattern:22 beating:1 reading:1 misclassification:1 indicator:1 technology:1 picture:1 kj:3 review:1 geometric:1 icc:1 eisenberg:1 loss:27 interesting:1 limitation:1 men:2 proportional:1 querying:6 versus:1 principle:1 whkh:1 bk2:1 pi:3 jth:1 theo:1 institute:1 neighbor:15 taking:1 boundary:5 dimension:2 calculated:1 curve:3 kz:1 fb:1 author:1 commonly:1 collection:1 san:1 ml:7 active:1 assumed:4 xi:4 mj:2 learn:1 ca:1 ignoring:1 obtaining:1 improving:1 necessarily:3 complex:1 suc:1 domain:1 cl:2 main:1 border:1 subsample:5 representative:1 referred:2 ighbor:1 tl:1 wish:1 learns:1 specific:2 showing:1 thea:1 pac:1 workshop:1 vapnik:5 false:2 ci:10 fc:1 subfamily:1 expressed:2 contained:1 springer:2 ch:1 nested:1 minimizer:1 conditional:3 viewed:1 bkl:1 adverse:1 typical:1 determined:1 uniformly:1 reducing:1 hyperplane:1 infinite:1 except:1 total:3 experimental:3 bisector:1 estimable:1 select:2 latter:2 confines:1 ethz:1 dept:1 |
460 | 1,419 | Competitive On-Line Linear Regression
V. Vovk
pepartment of Computer Science
Royal Holloway, University of London
Egham, Surrey TW20 OEX, UK
vovkGdcs.rhbnc.ac.uk
Abstract
We apply a general algorithm for merging prediction strategies (the
Aggregating Algorithm) to the problem of linear regression with the
square loss; our main assumption is that the response variable is
bounded. It turns out that for this particular problem the Aggregating Algorithm resembles, but is slightly different from, the wellknown ridge estimation procedure. From general results about the
Aggregating Algorithm we deduce a guaranteed bound on the difference between our algorithm's performance and the best, in some
sense, linear regression function's performance. We show that the
AA attains the optimal constant in our bound, whereas the constant attained by the ridge regression procedure in general can be
4 times worse.
1
INTRODUCTION
The usual approach to regression problems is to assume that the data are generated by some stochastic mechanism and make some, typically very restrictive,
assumptions about that stochastic mechanism. In recent years, however, a different
approach to this kind of problems was developed (see, e.g., DeSantis et al. [2], Littlestone and Warmuth [7]): in our context, that approach sets the goal of finding
an on-line algorithm that performs not much worse than the best regression function found off-line; in other words, it replaces the usual statistical analyses by the
competitive analysis of on-line algorithms.
DeSantis et al. [2] performed a competitive analysis of the Bayesian merging scheme
for the log-loss prediction game; later Littlestone and Warmuth [7] and Vovk [10]
introduced an on-line algorithm (called the Weighted Majority Algorithm by the
Competitive On-line Linear Regression
365
former authors) for the simple binary prediction game. These two algorithms (the
Bayesian merging scheme and the Weighted Majority Algorithm) are special cases
of the Aggregating Algorithm (AA) proposed in [9, 11]. The AA is a member of
a wide family of algorithms called "multiplicative weight" or "exponential weight"
algorithms.
Closer to the topic of this paper, Cesa-Bianchi et al. [1) performed a competitive
analysis, under the square loss, of the standard Gradient Descent Algorithm and
Kivinen and Warmuth [6] complemented it by a competitive analysis of a modification of the Gradient Descent, which they call the Exponentiated Gradient Algorithm. The bounds obtained in [1, 6] are of the following type: at every trial
T,
(1)
where LT is the loss (over the first T trials) of the on-line algorithm, LT is the
loss of the best (by trial T) linear regression function, and c is a constant, c >
1; specifically, c = 2 for the Gradient Descent and c = 3 for the Exponentiated
Gradient. These bounds hold under the following assumptions: for the Gradient
Descent, it is assumed that the L2 norm of the weights and of all data items are
bounded by constant 1; for the Exponentiated Gradient, that the Ll norm of the
weights and the Loo norm of all data items are bounded by 1.
In many interesting cases bound (1) is weak. For example, suppose that our comparison class contains a "true" regression function, but its values are corrupted by
an Li.d. noise. Then, under reasonable assumptions about the noise, LT will grow
linearly in T, and inequality (1) will only bound the difference LT - LT by a linear function of T. (Though in other situations bound (1) can be better than our
bound (2), see below. For example, in the case of the Exponentiated Gradient, the
0(1) in (1) depends on the number of parameters n logarithmically whereas our
bound depends on n linearly.)
In this paper we will apply the AA to the problem of linear regression. The AA
has been proven to be optimal in some simple cases [5, 11], so we can also expect
good performance in the problem of linear regression. The following is a typical
result that can be obtained using the AA: Learner has a strategy which ensures
that always
(2)
LT ~ LT + nIn(T + 1) + 1
(n is the number of predictor variables). It is interesting that the assumptions
for the last inequality are weaker than those for both the Gradient Descent and
Exponentiated Gradient: we only assume that the L2 norm of the weights and the
Loo norm of all data items are bounded by constant 1 (these assumptions will be
further relaxed later on). The norms L2 and Loo are not dual, which casts doubt
on the accepted intuition that the weights and data items should be measured by
dual norms (such as Ll-Loo or L 2-L2).
Notice that the logarithmic term nln(T + 1) of (2) is similar to the term ~ In T
occurring in the analysis of the log-loss game and its generalizations, in particular
in Wallace's theory of minimum message length, Rissanen's theory of stochastic
complexity, minimax regret analysis. In the case n = 1 and Xt = 1, Vt, inequality (2)
differs from Freund's [4] Theorem 4 only in the additive constant. In this paper
we will see another manifestation of a phenomenon noticed by Freund [4]: for some
important problems, the adversarial bounds of on-line competitive learning theory
V. Vovk
366
are only a tiny amount worse than the average-case bounds for some stochastic
strategies for Nature.
A weaker variant of inequality (2) can be deduced from Foster's [3] Theorem 1 (if
we additionally assume that the response variable take only two values, -1 or 1):
Foster's result implies
LT
~
LT + 8n In(2n(T + 1)) + 8
(a multiple of 4 arises from replacing Foster's set {O, 1} of possible values of the
response variable by our {-1, 1}j we also replaced Foster's d by 2n: to span our set
of possible weights we need 2n Foster's predictors).
Inequality (2) is also similar to Yamanishi's [12] resultj in that paper, he considers a
more general framework than ours but does not attempt to find optimal constants.
2
ALGORITHM
We consider the following protocol of interaction between Learner and Nature:
FOR t = 1,2, ...
Nature chooses Xt ? m.n
Learner chooses prediction Pt E
Nature chooses Yt E [-Y, Y]
END FOR.
m.
This is a "perfect-information" protocol: either player can see the other player's
moves. The parameters of our protocol are: a fixed positive number n (the dimensionality of our regression problem) and an upper bound Y > 0 on the value Yt
returned by Nature. It is important, however, that our algorithm for playing this
game (on the part of Learner) does not need to know Y.
We will only give a description of our regression algorithmj its derivation from the
general AA will be given in the future full version of this paper. (It is usually a nontrivial task to represent the AA in a computationally efficient form, and the case of
on-line linear regression is not an exception.) Fix n and a > O. The algorithm is as
follows:
A :=alj b:=O
FOR TRlAL t = 1,2, ... :
read new Xt E m.n
A:= A
+XtX~
output prediction Pt := b' A -1 Xt
read new Yt E m.
b:= b+YtXt
END FOR.
In this description, A is an n x n matrix (which is always symmetrical and positive
definite), bE mn , I is the unit n x n matrix, and 0 is the all-O vector.
The naive implementation of this algorithm would require O(n3) arithmetic operations at every trial, but the standard recursive technique allows us to spend only
O(n2 ) arithmetic operations per trial. This is still not as good as for the Gradient
Descent Algorithm and Exponentiated Gradient Algorithm (they require only O(n)
Competitive On-line Linear Regression
367
operations per trial); we seem to have a trade-off between the quality of bounds
on predictive performance and computational efficiency. In the rest of the paper
"AA" will mean the algorithm. described in the previous paragraph (which is the
Aggregating Algorithm applied to a particular uncountable pool of experts with a
particular Gaussian prior).
3
BOUNDS
In this section we state, without proof, results describing the predictive performance
of our algorithm. Our comparison class consists of the linear functions Yt = W? Xt,
where W E m.n ? We will call the possible weights w "experts" (imagine that we
have continuously many experts indexed by W E m.n ; Expert w always recommends
prediction w . Xt to Learner). At every trial t Expert w and Learner suffer loss
(Yt - w . Xt)2 and (Yt - Pt)2, respectively. Our notation for the total loss suffered by
Expert w and Learner over the first T trials will be
T
LT(W) := L(Yt - W? Xt)2
t=1
and
T
LT(Learner) := L(Yt - Pt)2,
t=1
respectively.
For compact pools of experts (which, in our setting, corresponds to the set of
possible weights w being bounded and closed) it is usually possible to derive bounds
(such as (2? where the learner's loss is compared to the best expert's loss. In our
case of non-compact pool, however, we need to give the learner a start on remote
experts. Specifically, instead of comparing Learner's performance to infw LT(W),
we compare it to infw (LT(W) + allwlI 2 ) (thus giving ~arner a start of allwII 2 on
Expert w), where a > 0 is a constant reflecting our prior expectations about the
"complexity" IIwll :=
of successful experts.
-IE:=1 w;
This idea of giving a start to experts allows us to prove stronger results; e.g., the
following elaboration of (2) holds:
(3)
(this inequality still assumes that
IIXtiloo ~ 1 for all t but w is unbounded).
Our notation for the transpose of matrix A will be A'; as usual, vectors are identified
with one-column matrices.
Theorem 1 For any fi:ted n, Leamer has a strategy which ensures that always
V. Vovk
368
II, in addition,
IIxt II 00 $ X, \It,
(4)
The last inequality of this theorem implies inequality (3): it suffices to put X =
Y=a=1.
The term
lndet (1 +; t,x.x:)
in Theorem 1 might be difficult to interpret. Notice that it can be rewritten as
nlnT + lndet
(~I + ~COV(Xl' ... ,Xn)) ,
where cov(Xl , ... , Xn) is the empirical covariance matrix of the predictor variables
(in other words, cov(Xl , ... ,Xn) is the covariance matrix of the random vector
which takes the values Xl, ... ,XT with equal probability ~). We can see that this
term is typically close to n In T.
Using standard transformations, it is easy to deduce from Theorem 1, e.g., the
following results (for simplicity we assume n = 1 and Xt,Yt E [-1,1], 'It):
? if the pool of experts consists of all polynomials of degree d, Learner has a
strategy guaranteeing
? if the pool of experts consists of all splines of degree d with k nodes (chosen
a priori), Learner has a strategy guaranteeing
LT(Learner) S inf
(LT(W)
w
+ Ilw1l 2 ) + (d + k + 1) In(T + 1).
The following theorem shows that the constant n in inequality (4) cannot be improved.
Theorem 2 Fix n (the number 01 attributes) and Y (the upper bound on IlItl). For
any f > 0 there exist a constant C and a stochastic strategy lor Nature such that
IIxtiloo = 1 and Illtl = Y, lor all t, and, lor any stochastic strategy lor Learner,
E (LT(Learner) -
4
inf
w:llwll:SY
LT(W))
~ (n -
f)y21nT - C,
'IT.
COMPARISONS
It is easy to see that the ridge regression procedure sometimes gives results that
are not sensible in our framework where lit E [- Y, Y] and the goal is to compete
Competitive On-line Linear Regression
369
against the best linear regression function. For example, suppose n
and Nature generates outcomes (Xt, Yt), t = 1,2, ... , where
a
?
Xl
?
X2
? ... ,
= 1, Y = 1,
_ { 1, if todd,
Yt -1, if t even.
At trial t = 2,3, ... the ridge regression procedure (more accurately, its natural
modification which truncates its predictions to [-1, 1]) will give prediction Pt = Yt-l
equal to the previous response, and so will suffer a loss of about 4T over T trials.
On the other hand, the AA's prediction will be close to 0, and so the cumulative
loss of the AA over the first T trials will be about T, which is close to the best
expert's loss. We can see that the ridge regression procedure in this situation is
forced to suffer a loss 4 times as big as the AA's loss.
The lower bound stated in Theorem 2 does not imply that our regression algorithm is
better than the ridge regression procedure in our adversarial framework. (Moreover,
the idea of our proof of Theorem 2 is to lower bound the performance of the ridge
regression procedure in the situation where the expected loss of the ridge regression
procedure is optimal.) Theorem 1 asserts that
LT(Leamer) :S
~ (LT( w) + allwl12) + y2
t.
In ( 1 +
t.
~ X~.i)
(5)
when Learner follows the AA. The next theorem shows that the ridge regression
procedure sometimes violates this inequality.
Theorem 3 Let n = 1 (the number 0/ attributes) and Y = 1 (the upper bound
on IYtl); fix a > O. Nature has a strategy such that, when Learner plays the ridge
regression strategy,
LT(Learner) = 4T + 0(1),
(6)
inf (LT(w)
UI
In
as T
5
-4 00
+ allwll2) = T + 0(1),
(1+ ~ t.x~) =
TIn2+ 0(1)
(7)
(8)
(and, there/ore, (5) is violated).
CONCLUSION
A distinctive feature of our approach to linear regression is that our only assumption about the data is that IYt I ~ Y, 'tit; we do not make any assumptions about
stochastic properties of the data-generating mechanism. In some situations (if the
data were generated by a partially known stochastic mechanism) this feature is a
disadvantage, but often it will be an advantage.
This paper was greatly influenced by Vapnik's [8] idea of transductive inference.
The algorithm analyzed in this paper is "transductive", in the sense that it outputs
some prediction Pt for Yt after being given Xt, rather than to output a general rule
for mapping Xt into Ptj in particular, Pt may depend non-linearly on Xt. (It is easy,
however, to extract such a rule from the description of the algorithm once it is
found.)
V. Vovk
370
Acknowledgments
Kostas Skouras and Philip Dawid noticed that our regression algorithm is different
from the ridge regression and that in some situations it behaves very differently.
Manfred Warmuth's advice about relevant literature is also gratefully appreciated.
References
[11 N. Cesa-Bianchi, P. M. Long, and M. K. Warmuth (1996), Worst-case quadratic
loss bounds for on-line prediction of linear functions by gradient descent, IEEE
7rons. Neural Networks 7:604-619.
[21 A. DeSantis, G. Markowsky, and M. N. Wegman (1988), Learning probabilistic
prediction functions, in "Proceedings, 29th Annual IEEE Symposium on Foundations of Computer Science," pp. 110-119, Los Alamitos, CA: IEEE Comput.
Soc.
[31 D. P. Foster (1991), Prediction in the worst case, Ann. Statist. 19:1084-1090.
[4] Y. Freund (1996), Predicting a binary sequence almost as well as the optimal
biased coin, in "Proceedings, 9th Annual ACM Conference on Computational
Learning Theory" , pp. 89-98, New York: Assoc. Comput. Mach.
[5] D. Haussler, J. Kivinen, and M. K. Warmuth (1994), Tight worst-case loss
bounds for predicting with expert advice, University of California at Santa
Cruz, Technical Report UCSC-CRL-94-36, revised December. Short version in
"Computational Learning Theory" (P. Vitanyi, Ed.), Lecture Notes in Computer Science, Vol. 904, pp. 69-83, Berlin: Springer, 1995.
[6] J. Kivinen and M. K. Warmuth (1997), Exponential Gradient versus Gradient
Descent for linear predictors, Inform. Computation 132:1-63.
[7] N. Littlestone and M. K. Warmuth (1994), The Weighted Majority Algorithm,
Inform. Computation 108:212-261.
[8] V. N. Vapnik (1995), The Nature of Statistical Learning Theory, New York:
Springer.
[9] V. Vovk (1990), Aggregating strategies, in "Proceedings, 3rd Annual Workshop
on Computational Learning Theory" (M. Fulk and J. Case, Eds.), pp. 371-383,
San Mateo, CA: Morgan Kaufmann.
[10] V. Vovk (1992), Universal forecasting algorithms, In/orm. Computation
96:245-277.
[11] V. Vovk (1997), A game of prediction with expert advice, to appear in J. Comput. In/orm. Syst. Short version in "Proceedings, 8th Annual ACM Conference
on Computational Learning Theory," pp. 51-60, New York: Assoc. Comput.
Mach., 1995.
[12] K. Yamanishi (1997), A decision-theoretic extension of stochastic complexity
and its applications to learning, submitted to IEEE 7rons. In/orm. Theory.
| 1419 |@word trial:11 version:3 polynomial:1 norm:7 stronger:1 covariance:2 contains:1 ours:1 comparing:1 cruz:1 additive:1 item:4 warmuth:8 short:2 manfred:1 node:1 ron:2 desantis:3 lor:4 unbounded:1 ucsc:1 symposium:1 consists:3 prove:1 paragraph:1 expected:1 wallace:1 bounded:5 notation:2 moreover:1 kind:1 developed:1 tin2:1 finding:1 transformation:1 every:3 assoc:2 uk:2 unit:1 appear:1 positive:2 aggregating:6 todd:1 mach:2 might:1 resembles:1 mateo:1 acknowledgment:1 recursive:1 regret:1 definite:1 differs:1 lndet:2 procedure:9 tw20:1 universal:1 empirical:1 xtx:1 word:2 orm:3 cannot:1 close:3 put:1 context:1 yt:13 simplicity:1 rule:2 haussler:1 pt:7 suppose:2 imagine:1 play:1 logarithmically:1 dawid:1 worst:3 ensures:2 remote:1 trade:1 intuition:1 complexity:3 ui:1 depend:1 tight:1 tit:1 predictive:2 distinctive:1 efficiency:1 learner:19 differently:1 derivation:1 forced:1 london:1 outcome:1 spend:1 cov:3 transductive:2 advantage:1 sequence:1 interaction:1 relevant:1 description:3 asserts:1 los:1 nin:1 yamanishi:2 perfect:1 guaranteeing:2 generating:1 derive:1 ac:1 measured:1 alj:1 soc:1 implies:2 attribute:2 stochastic:9 violates:1 require:2 fix:3 generalization:1 suffices:1 extension:1 hold:2 mapping:1 ptj:1 estimation:1 weighted:3 always:4 gaussian:1 rather:1 greatly:1 adversarial:2 attains:1 sense:2 inference:1 typically:2 dual:2 priori:1 special:1 equal:2 once:1 ted:1 lit:1 future:1 report:1 spline:1 replaced:1 attempt:1 message:1 analyzed:1 closer:1 indexed:1 oex:1 littlestone:3 column:1 disadvantage:1 predictor:4 successful:1 loo:4 corrupted:1 chooses:3 deduced:1 iixt:1 ie:1 probabilistic:1 off:2 pool:5 continuously:1 cesa:2 worse:3 expert:17 li:1 doubt:1 syst:1 depends:2 performed:2 later:2 multiplicative:1 closed:1 competitive:9 start:3 square:2 kaufmann:1 sy:1 weak:1 bayesian:2 accurately:1 fulk:1 submitted:1 influenced:1 inform:2 ed:2 against:1 surrey:1 pp:5 proof:2 dimensionality:1 reflecting:1 attained:1 response:4 improved:1 though:1 hand:1 replacing:1 quality:1 true:1 y2:1 former:1 read:2 ll:2 game:5 manifestation:1 ridge:11 theoretic:1 performs:1 fi:1 behaves:1 he:1 interpret:1 rd:1 gratefully:1 deduce:2 recent:1 inf:3 wellknown:1 inequality:10 binary:2 nln:1 vt:1 morgan:1 minimum:1 relaxed:1 arithmetic:2 ii:2 multiple:1 full:1 technical:1 long:1 elaboration:1 prediction:14 variant:1 regression:29 expectation:1 represent:1 sometimes:2 whereas:2 addition:1 ore:1 grow:1 suffered:1 biased:1 rest:1 member:1 december:1 seem:1 call:2 recommends:1 easy:3 identified:1 idea:3 rhbnc:1 forecasting:1 suffer:3 returned:1 york:3 santa:1 amount:1 statist:1 exist:1 notice:2 per:2 vol:1 rissanen:1 year:1 compete:1 family:1 reasonable:1 almost:1 decision:1 bound:21 guaranteed:1 vitanyi:1 replaces:1 quadratic:1 annual:4 nontrivial:1 n3:1 x2:1 generates:1 span:1 llwll:1 slightly:1 modification:2 computationally:1 turn:1 describing:1 mechanism:4 know:1 end:2 operation:3 rewritten:1 apply:2 egham:1 coin:1 uncountable:1 assumes:1 giving:2 restrictive:1 move:1 noticed:2 alamitos:1 strategy:11 usual:3 gradient:15 berlin:1 majority:3 sensible:1 philip:1 topic:1 considers:1 length:1 iyt:1 difficult:1 truncates:1 stated:1 implementation:1 bianchi:2 upper:3 revised:1 descent:8 wegman:1 situation:5 introduced:1 cast:1 california:1 below:1 usually:2 royal:1 natural:1 predicting:2 kivinen:3 mn:1 minimax:1 scheme:2 imply:1 naive:1 extract:1 prior:2 literature:1 l2:4 freund:3 loss:18 expect:1 lecture:1 interesting:2 proven:1 versus:1 foundation:1 degree:2 foster:6 tiny:1 playing:1 last:2 transpose:1 appreciated:1 exponentiated:6 weaker:2 wide:1 xn:3 cumulative:1 author:1 san:1 compact:2 symmetrical:1 assumed:1 additionally:1 nature:9 ca:2 protocol:3 iiwll:1 main:1 linearly:3 big:1 noise:2 n2:1 advice:3 kostas:1 exponential:2 xl:5 comput:4 theorem:13 xt:14 workshop:1 vapnik:2 merging:3 occurring:1 lt:21 logarithmic:1 partially:1 springer:2 aa:13 corresponds:1 complemented:1 acm:2 goal:2 infw:2 ann:1 crl:1 specifically:2 typical:1 vovk:8 called:2 total:1 accepted:1 player:2 exception:1 holloway:1 arises:1 violated:1 phenomenon:1 |
461 | 142 | 99
Connectionist Learning of Expert Preferences by
Comparison Training
Gerald Tesauro
IBl\f Thomas.1. '''atson Rcsearc11 Centcr
PO Box 704, Yorktown Heights, NY 10598 USA
Abstract
A new training paradigm, caned the "eomparison pa.radigm," is introduced
for tasks in which a. network must learn to choose a prdcrred pattern from a
set of n alternatives, based on examplcs of Imma.n expert prderences. In this
pa.radigm, the inpu t to the network consists of t.wo uf the n alterna tives, and
the trained output is the expert's judgement of which pa.ttern is better. This
para.digm is applied to the lea,rning of hackgammon, a difficult board ga.me in
wllieh the expert selects a move from a. set, of legal mm?es. \Vith compa.rison
training, much higher levels of performance can hc a.chiew~d, with networks
that are much smaller, and with coding sehemes t.hat are much simpler and
easier to understand. Furthermorf', it is possible to set up the network so
tha.t it always produces consisten t rank-orderings .
1.
Introduction
There is now widespread interest, in tlH~ use of conncctillnist networks fllr realworld practical problem solving. The principal areas of applica.tion which
have been studied so far involvc rela tiv('ly low-level signal processing and
pattern recognition t.asks. However, eOllllectionist networks might also he
useful in higher-level tasks whkh a.re curr('nt.l~? tackled hy cxprrt systems
a.nd knowledge engineering approadl(,s [2]. In this pa per, 'vc considcr problem
domains in which tlte expert is givcn a s('t of 71. alt.enulti,?es as input (71. may be
either sma.II or large), and mnst select. t.}l<' m()st dC'sirtlhl(' ()T most prderable
alternative. This type of task occurs rep('atccUy throughout the domains
of politics, business, eeonomics, mcdicine, a.nd many ot hers. \Vhcthcr it is
choosing a fureign-policy option, a. w('apons cont,rador, a course of trea.tment
for a disease, or simply what to have for dinner, prohlems requiring choice
are constan tly being faced and 801\'(\(1 by 111lma.n experts.
How might a. learning system snch as a. cOlllH,ctionist network be set up to
learn to ma.ke such choices from human expert cxampI~s? The immediately
obvious a,pproa,c h is to train the network to produce a numerical output
100
Tesauro
"score" for ea,ch input alterna.tive. To make a, choice, then, une would have
the network score each a.lterna,tive, and select t.he alterna tive with the highest score. Since the lea,rning system learns from examples, it. seems logical to
train the network on a da.ta base of examples in which a, human expert has
entered a numerical score for ea,ch possible choice. Howc\'cr, there are two
major problems with such a,n a,pproa,c h. First, in many domains in which n
is large, it would be tremendously time-consuming for the expert to create
a, data base in which each individual a.lternative has been painstaking evalua,ted, even the vast number of obviously ba,d alterna.tives which are not even
worth considering. (It is importa.nt for the network t,o sec examples of ba,d
alternatives, otherwise it would tend t.o produce high scores for everything.)
l\1:ore importa.ntly, in ma,n y domains human experts do not think in terms of
absolute scoring functions, and it would thus be extremely difficult to create
training data containing a,hsolute scores, because such scoring is alien to the
expert's wa,y of thinking a,h out the problem. Inst.cad, the most natural way
to make training data is simply tu record the expcrt in adinn, i.e., for ea,eh
problem situation, record each of the tlltcrna,th'cs hc htld t.o choose from, a,n d
record which one he actually selected.
.
For these reasons, we advoca.te teaching the network to compare pairs of
alterna.tives, rather than scoring individual aIterna.tivcs. In ot.her words, the
input should be two of the set of n alternatives, and thc output should be a
1 or 0 depending on which of the two alterna.tives is hetter. From a set of
recorded huma,n expert preferences, one can then teach thc network that the
expert's choice is better than all other a.ltcrna tivcs.
One potential concern raised by this approach is tha t, in performance mode
a,fter the network is trained, it might he neccssary t.o mtlke n 2 comparisons to
select the best alterna.tive, whereas only n individual scor('s tlrc needed in the
other approa,ch. However, the network can select the hcst alterna.tive with
only n compa.risons by going through the list of altern a tives in ordcr, a.nd
compa,ring the current alternative wHh thc b('st ait.erntltive secn so fa.r. If
the current alterna.tive is better, it becomcs the ncw 1>('st altcrnative, and if
it is worse, it is rejected. Anot.hcr poten tial COIlcern is tha t tI. network which
only knows how t.o compare might not produce a consist('nt rank-ordering,
i.e., it might sa.y that alternativc a is better t.han h, b is hdtN t.han c, and
c is better than a, and then one do('s not know which alt('rnative to select.
However, we shall see la.ter that it is possiblc to gllartlutc'c (,(lllsist('ncy with a
constrained a.rchitecture which forccs t.he network to c:()mc~ IIp wit.h absolute
numerical scores for individual alterna tives.
In the following, we shall exa.minc the applic~tI. ti()n of t.he comparison training pa.radigm to the ga.me of backgammon, as considC'ra hIe cxperience ha.s
already been obtained in this domain. In previous papC'rs [7,6]' a network
was described which lea.rned to pla.y ba,c kgammon from an expert data, base,
using the so-called "back-proptl,ga.tion" learning rule [5J. In that system, the
network was trained to score individual moyes. In other words, the input
Connectionist Learning of Expert Preferences
consists of a move (drfined by the initi::ll position bdnre the move a.nd the
final position after the move), and the desired output is a. real number indica.ting the strength of the move. Henceforth we shall refer to this training
paradigm as the "rcla.t.ive score" par::ldigm. 'Vhilc this approach produced
considerable success, it had a. number of serious limitations. \Ve sha.ll see
that the compa.rison paradigm solyes one of the most important limita.tions
of t.he previous a.pproach, with the rrsult tha t the overall performance of
Ute network is much bet ter, the llum her of connedions required is greatly
reduced, a.nd. the network's input coding scheme is much simpler and easier
to understa.nd.
2.
Previous backgammon networks
In [7], a network was descrihed which learnrd tl) pla~? fairly good ba.ckga.mmon by ba.ck-propa.gation learning of a. large expert t.raining set, using the
rcla,tive score paradigm describc(l pr(,yiollsly. After tr::lining, the network
was tested both by measuring its prrformance on ::I t('st sct. of positions not
used in training, and hy a,ctual game pl::l)' against hunums ::Ind cOllventional
computer programs. The best net.work was ablr to drfeat. Sun :Microsysterns' Gammontool, Ute best available commercial program, hy a sllbsta.ntial
ma.rgin, but it was still fa.r from human expert-level performance.
The basic conclusion of [7] was that it was possihle to achieve decent levels
of performance by this network learning procedure, but it, was not a.n ea.sy
matter, and required suhst.antial hum::ln intervention. The choice of a. coding
scheme for the inpnt informa.t.ion, for rxamplr, was fonn(l to be an extremely
important issue. The hest coding SclH'IrleS contained a. great deal of doma.inspecific information. The best, encoding of the "raw" board iuforma.tion was
in terms of concepts that 111lman exprrts use to describe loc::II tra.llsitions, such
as "slotting," "stripping," et.c.. Also, a few "pre-colllputrd features" were
required in addition t.o the raw board inform::lt.ion. Thus it ,,?as necessary
to be a domain expert in order to clpsign a suit,a hIe ndwork I.oding sdteme,
a.nd it seemed tha.t the only way t.o discov('r thr hrst cOlling scheme was
by painstaking trial and error. This "'::IS som('\dtat. disappoiut,ing, as it was
hoped tha.t the net\vork lea rIling l)foc('(illre \y()ul<l alit lima t ically produl.e an
expert ba.ckgammon net.work with little or no human ('ffort.
3.
Comparison paradigul network set-up
111 the sta.nd.ard pra.ctice of ha.c:k-propag::lt.inn, a compfHisoIt paradigm network would ha.ve a.n input layer, (lIlC ()r morr layrTs of hidden nnits, and a.n
output layer, wit.h full connectivity hdwcrn adjacell t l::lyers. The input. layer
would represent, two final board posit.ions a and '" and the out.put layer would
ha.ve just a single unit t.o represent which board position was better. The
101
102
Tesauro
teacher signal for the output unit would be a, 1 if hoard position
than b, a,nd a 0 if b was better tha.n a.
(l
was better
The proposed compa,rison pa.r a.digm network "'ould m;ercome the limita,t.ion
of only being able to consider individual moves in isolation, without knO\vledge of what other a.lt.erna,tives arc available. In addit.ion, t.he sophisticated
coding scheme that was developed to encode transition informa tion would not
be needed, since compa,risons could be based solely on the final board sta.tes.
The comparison approa.ch offers grca ter sensitivity in dist.inguishing between
close alternatives, a,nd as stated prc"iously, it corrcsp onds more closely to
the actual form of human expert knowledge.
These advantages a,re formida,b le, hut. t.here are some import.ant. problems
with the a.pproach as curren t.ly descrihed. One technical prohlem is tlta,t the
learning is significantly slower. This is beca,use 271, comparisons per t.ra.ining
position a,re prcsen ted to the network, where 11. '" 20, wh('rC'tls in tllC relative
score approach, only about 3-4 moves per position would he presented. It
was therefore necessary to develop a, number of technical tricks to increase
the speed of the simulator code for t.hi.s specific a.pplica tion (to he described
in a future publication).
A more fundamental problem with the a.pproa("h, however, is the issue of consistency of network comparisons . Two properties arc required for complete
consistency: (1) The comparison between any two positions mllst be unambiguous, i.e., if the network says that. a is bett('r t.han b when a is presented
on the left a,nd b on the right, it !tact hetter say that. a is hetter than b if
a is on the right and b is on the left. One ca.n show tha t t.his requircs t.he
network's output to exaetly invert whenever Ule input. hO(lrd positions are
swapped. (2) The compa,risons must he transitire, as alluded to previously,
i.e., if a is judged better than h, and 1J is judged better than (', the network
lta,d better judge a to be better than c.
Sta.ndard unconstrained networks hayc no gUll ra n tee flf sa t.isfying eit.her of
these properties. After some thought, howe\"('I, one realizc's that the output
inversion symmetry can be enforcC'd by a symmetry rc'ltl tion amongst the
weight.s in the network, and that the transiti"ity ancl rank-OHler consist.ency
can be guaranteed by separability in the ardtitC'dure, as illllstratC'd in Figurc
1. Here we see tlla,t t.his network really consists of t.wo half-networks, one of
which is only conceIllcd with t}w evaluat.ion "fhoard positi.on (I, and the ot.her
of which is concerned only wit.h t.h~ C',"alua t.iol\ of hOH.T() posit.inn b. (Duc to
the indicated symmetry relation, nne neC'ds only store oue half-network in the
simulator code.) Each half-network mrly have on(' or morc' lay('rs of hidden
units, but as long as they a,re not cross-coupled, t.he C'valuat.ion of each of
the two input boa,rd positions is hoilC'd down to a single rC'al number. Since
real numbers always ra,nk-order consist.en tly, t.he ll('t\,'ork's compa,r isons a,re
al ways com;is ten t.
Connectionist Learning of Expert Preferences
final position
(a)
final position (b)
Figure 1: A network design for cOlTlpari~()n trailling wit h gnarallteed consistency of comparisons. \Veight groHps han~ ~YlTlmet.ry rc1ati()ns W 1
W2
and W 3 == - W 4, which ensures I hat the outPllt cxact.Jy in\Trts upon swapping positions in the input arraJ'. Separation or the hidden unils cOllrlenses
the evaluation of each final hoard position into a single real IIIIJTlhcr, thus
ensuring transitivity.
=
An importa.nt a.dded benefit of t.his scheme is that an a.hsoillte hoard evalua.tion fundion is obtained in each half-network. This means t.hat the netwurk, to the extcnt that it.s cvaluation fllnct.ioll is ac-cnrat.e, has an intrinsic
understanding of a. given posit.ion, as opposed to HH'rely heing a hIe to detect features which correspond t.o good moves. As has heen emphasil':cd by
Berliner [1], an intrinsic uudr.rst.a.llding of the position is crucial for play at
the highest levels, a.n<l for use of t,he dOllhling cnl)('. Thnf-:, this a.pproach
can serve as the ba.sis for fut.ure progress, whNeas thr previous approach of
scoring moves was doomed eventually to run into a cirad rud.
4.
Results of comparison training
The training procedure for the comparison paradigm nC't.work was as follows:
Networks ""'ere set up with 289 input. units which enc(l(le a (ic'scription of single final boa.rd position, varying nnmlH'rs (If hidden units, and a single output
unit. The training data was taken from a set (If 40n gamrs ill which the author played both sides. This <ia ta sd con hlins a recording of t.he a.n t,hor's
preferred move for each position, and no ot.her comments. The engaged positions in the da.ta set weIe selected out, (discngflged ra.cing I)()sit.ions were not
studied) and divided int.o five cat<'gories: hearoff, bearin, opponent bearoff,
opponent b~a.rin, a.nd a default ca t.egnry ('overing eVPlything else. In each
103
104
Tesauro
Type of
n.SP net,
CP net
_t~e_st_s~e~t_____(~6_5_1-_1_2-_1~)~(_289-1)_
bcaroff
.82
.83
hearin
.54
.60
opp. bea.roff?
.56
.54
opp. bearin
.60
.66
other
.58
.65
Table 1: Performance of neLs of indicated size on respedin~ test. sets, as measllred by fraction of positions for which TIel agrees wit h lmTllrlll expert choice
of best move. HSP: relative score paradigm, CP: comparison paradigm.
category, 200 positions chosen (I.t. r(1UdOIU were set. (lside to he us('d as testing <lat(l.; tIte remaining da ta. (a bon t. 1000 positions in ('(I eh category except
t.he dcfa.lllt category, for which a.bout 4000 positions were used) was used to
tra.iu networks which spedalil':ed in each category. The learning algorithm
llsed was st.a.nd(l.rd back-propagation with IIWInelltllIn a.nd without weight
decay.
Performa,nce after tr(lining is summaril':ed in Ta.hl('s 1 and 2. Ta.hle 1 gives
the performance of each specialist network on the appropriate set of test
positions. Results for the comparison par(ldigm nebYllrk are shown for networks without hidden units, because it was fouud that the (lddition of hidden
units did not improve the performance. (This is discussed in the following
section.) \Ve contra.st. these results with result.s of tr(lining networks in the
rela.tive score pa,radigm on the same tr(lilling da.t(l sds. "~e see in Ta hIe 1
tha.t for the bearoff and oppon('nt. hearofi' speeialists, there is only a small
cha.nge in perform(lnce under t.he contpa.rison p(lradigm. Fllr the hea.rin and
opponcnt. bcarin specialists, t,herr is an improv('ment. in IH'rformance of a.bout
6 percenta.ge poin ts in each casco For t.his pa.rtkular applica tion, this is a. "ery
substantial improvement in perfoIIuance. How('y<'I, th(~ most, import(lnt finding is for the default ca tegory, which is much l(lrgcr and mor(' difficult than
any of the specialist ea.t.egories. Th(' d('f(lult network's prrformance is the
key factor in determining the syst('m's IIvrr(lll g(lme prrfOrIll(lllCe. 'Vit.h cumpa.rison t.rainillg, we find an improvc'Ineur. in perform(lll('e from 58% t.o 65%.
Given the sil':e and diffkllIt.y of t,his ca t('gory, t his ('(Ill onl)- he desnibcd as a.
huge improvement, in performance, and is all t.h(' mllre rem ark(l hIe when nne
considers th(lt. t,he comp(lrison p(I.radigm net has only 30() w('ights, (IS opposed
to 8000 weight,s for the relative score paradigm net.
Next, a combined game-playing s)-stC'ln was s('t. up llsing t.hr the specialist.
nets for all engaged posit.ions. (The G(llTlmontool evaluat.ion function was
called for ra.cing posit.ions.) Results are given in T(lble 2. Ag(linst, Gammontool itself, the pcrforma.nce und('r thr comparison p(lr(l(ligm improves from
59% to 64%. Against the a.uthor (and tcachE'r), the pC'rforma.nce improves
from a.n estimat,ed 35% (since the nsp net.s (Ire so hig and slow, a.ccnra,te
Connectionist Learning of Expert Preferences
Opponent
Ga,mmon tool
TesauIO
RSP nets
.59 (500 games)
.35 (l00 ga,mes)
CP nets
.64 (2000 games)
.42 (400 games)
Table 2: Game-playing performance of composite network systems against
Gammontool and against the author, as measured by fractioTl of games won,
without counting gammons or backgammons.
statistics could not be obtained) to about 42%.
Qualitatively, one notices a subst.a,ntial overall improvement in the new network's level of pla,y. But what, is most, striking is the nct.work's In)lst case beha,vior. The previous relative-score network ha.d particularly bad worst-case
beha,vior: about once every other ga,me, the network would make an atrocious blunder wltich would seriously jeopardize its dlances of winning that
ga,me [6]. An alarming fradion of these blunders ,,,,ere seemingly random and
could not be logically explained. The new comparison para,digm network's
wOIst-ca.se behavior is vastly improvcd in this rega,rd. The frcquency and
severity of its mistakes are significantly reducE!d, hut more imp(lrtantly, its
mista,kes a.re understandable. (Some of the improvement in this respect may
be dne to the elimination of the noisy teacher signal descrihed in [7].)
5.
Conclusions
\Ve have seen that, in the domain of backgammon, the int.roduct.ion of the
comparison training pa.ra,digm has result.ed in networks whkh perform much
better, with vastly reduced numbers Ilfweights, a.ud with input. coding schemes
tha.t a.r e much simpler and easier t.o understand. It was surprising that snch
high performa.nce could be obtained in "perceptrou" networks, i.E!., networks
withou t hidden units. This reminds us t.hat. one should not. summarily dismiss perceptrons as uninteresting or unworthy of study hecause they arc only
ca.pable of lea.rning linearly separable funct.ions [3]. A su hst.an tial component
of many difficult real-world problems may lie in the liu('arly separable spectrum, and thus it makes sense to try perceptIons at least as a first attempt.
It was also surprising tha.t the use (If hidden units in the comparison- trained
networks does not improvc the pcrformanc('. This is nn('xplained, a.nd is t.he
subject of current resea.rch. It is, however, not without precedent: in at least
one other real-world application [4], it has heen found tha.t networks with
hidden units do not pcrform any bettE'r than netwllrks without hidden units.
?vIore generally, one might conclude t.ha.t, in training a. neural network (or
indeed a.ny learning system) from human expert. E'xamples in a cumplex domain, there should he a. good match hetween Ute na t,ural form of the expert's
knowledge and the method by which the net.work is trained. For domains in
which the expert must seh~ct a preferred alt.crnative from a set of alternatives,
105
106
Tesauro
the expert naturally thinks in terms of comparisons a.mongst the top few a.lterna.tives, and the compa.rison paradigm proposed here takes advantage of
that fact. It would be possihle in principle to train a network using absolute
evaluations, hut the crea.tion of sueh a. training set might he too difficult to
underta.ke on a large scale.
If the above discussion is coned, then the comparisou pa.ra.digm should be
useful in ot,her applications involving expert choice, and in other lea.rning
syst,ems besides connectionist networks. Typically expert systems a.re handcrafted by knowledge engineers, ra.ther than learned from human expert examples; however, there has recently been some interest in sl1pervis(~d lea.rning
approa.ches. It will be interesting to see if the compa.rison paradigm proves to
be useful when supervised lea.rning procedures are applied t,o otller domains
involving expert choice. In using the compa.rison paradigm, it. will be important to ha.ve some way to gua.ra.n t,ee that the syst,em's comparisons will be
unambiguous and t,ra.nsitive. For feed-forward networks, it was shown in this
pa.per how to gua.rantee this using symmetric, separa.t.ed nrtworks; it should
be possible to impose similar constraints Oil otll<'r learning systems to enforce
consistency.
References
[l} H. Berliner, "On the construction
Proc. of IleA T (1979) 53--55.
or c\"aluation
runctions for large domains,"
[2] S. 1. Gallant, "Conncctionist r.xpert systems," Comm. l\eM 31, 152 -- 169
(1988).
P}
M. Minsky and S. Papcrt, Pcrceptrons, MIT Press, Camhridge ~fA (1969).
[,1] N. Qian and T. J. Sejnowski, "Predicting the' secondary slrucl,nre of globular
proteins using neural nctwork models," .1. Mol. Bioi. 202, R(jfi 8R,1 (1988).
[5] D. E. Rumelarl and J. L. ~lcClellanrl (cds.), Parallel J>i.<;frifllltecl Processing:
Explorations in I.Ire Microsirucl.lIl'r
Cogn;t,;on, Vt)ls. I and 2, l\f1T Prrss,
Cambrioge MA (1986).
or
[6} G. Tesauro, "Neural network ddral.s creal.or in hack~annl1(ln match." Univ.
of ll1inoiR, Center for Complex S~' strll1s Technical Heporl. CCSl1-8R-6 (1988).
[71 G. Tesauro and T. J. Sejnowski, "/\ parallel nei.work that ]rarns to play
backgammon," A rtificialTntell;gmrcc, in press (H)89) .
| 142 |@word trial:1 judgement:1 inversion:1 seems:1 llsed:1 nd:15 cha:1 r:3 fonn:1 asks:1 tr:4 liu:1 loc:1 score:15 seriously:1 current:3 com:1 nt:5 cad:1 surprising:2 si:1 must:3 import:2 numerical:3 applica:2 half:4 selected:2 une:1 painstaking:2 alit:1 record:3 lr:1 rch:1 ire:2 preference:5 simpler:3 five:1 height:1 rc:3 consists:3 veight:1 prohlem:1 indeed:1 ra:11 behavior:1 dist:1 simulator:2 ry:1 berliner:2 rem:1 little:1 actual:1 lll:2 considering:1 nre:1 what:3 developed:1 finding:1 ag:1 every:1 ti:3 onds:1 estimat:1 limita:2 ly:2 intervention:1 unit:12 engineering:1 sd:2 mistake:1 encoding:1 ure:1 solely:1 might:7 f1t:1 studied:2 hest:1 practical:1 pla:3 testing:1 cogn:1 procedure:3 area:1 rning:6 significantly:2 thought:1 composite:1 word:2 pre:1 gammon:1 protein:1 beca:1 prc:1 ga:7 close:1 hle:1 judged:2 put:1 center:1 vit:1 l:1 ke:2 wit:5 immediately:1 qian:1 rule:1 his:6 ity:1 construction:1 commercial:1 lima:1 play:2 vork:1 pa:11 trick:1 jfi:1 recognition:1 particularly:1 lay:1 ark:1 worst:1 ensures:1 sun:1 ordering:2 highest:2 withou:1 disease:1 substantial:1 und:1 comm:1 gerald:1 runctions:1 blunder:2 trained:5 solving:1 funct:1 serve:1 upon:1 rin:2 po:1 eit:1 lst:1 cat:1 iol:1 approa:3 train:3 univ:1 describe:1 sejnowski:2 hetter:3 choosing:1 ive:1 cnl:1 say:2 otherwise:1 statistic:1 think:2 itself:1 noisy:1 final:7 seemingly:1 obviously:1 advantage:2 trailling:1 net:12 inn:2 ment:1 performa:2 tu:1 enc:1 entered:1 achieve:1 outpllt:1 rst:1 arly:1 produce:4 ring:1 tions:1 depending:1 develop:1 ac:1 measured:1 ard:1 progress:1 sa:2 c:1 hst:1 judge:1 posit:5 closely:1 kno:1 summarily:1 vc:1 exploration:1 human:8 dure:1 elimination:1 everything:1 globular:1 dne:1 really:1 lma:1 ttern:1 mm:1 pl:1 hut:3 ic:1 great:1 casco:1 vith:1 sma:1 major:1 doma:1 proc:1 agrees:1 create:2 ere:2 tool:1 mit:1 always:2 rather:1 ck:1 cr:1 dinner:1 minc:1 bet:1 varying:1 publication:1 encode:1 lman:1 improvement:4 separa:1 rank:3 backgammon:5 logically:1 alien:1 greatly:1 ibl:1 tremendously:1 hcr:1 sense:1 detect:1 inst:1 nn:1 alarming:1 typically:1 her:6 hidden:10 relation:1 going:1 selects:1 iu:1 overall:2 issue:2 ill:2 vior:2 scription:1 raised:1 constrained:1 fairly:1 once:1 ted:2 possihle:2 contra:1 thinking:1 imp:1 future:1 lrd:1 connectionist:5 serious:1 few:2 sta:3 ve:6 individual:6 minsky:1 suit:1 curr:1 attempt:1 pra:1 interest:2 huge:1 evaluation:2 pc:1 swapping:1 necessary:2 re:7 desired:1 gull:1 ncy:1 tial:2 hetween:1 measuring:1 whh:1 bon:1 trea:1 uninteresting:1 too:1 stripping:1 para:2 teacher:2 combined:1 st:6 fundamental:1 sensitivity:1 ches:1 na:1 connectivity:1 vastly:2 recorded:1 containing:1 choose:2 iip:1 opposed:2 henceforth:1 worse:1 expert:31 syst:3 potential:1 coding:6 sec:1 scor:1 matter:1 tra:2 int:2 tion:9 tiv:1 try:1 option:1 sct:1 parallel:2 ery:1 roff:1 rison:8 sy:1 correspond:1 ant:1 subst:1 raw:2 produced:1 mc:1 worth:1 comp:1 inform:1 whenever:1 ed:5 snch:2 against:4 unils:1 lnt:1 hack:1 obvious:1 naturally:1 con:1 wh:1 logical:1 rega:1 knowledge:4 improves:2 sophisticated:1 ea:5 actually:1 back:2 feed:1 higher:2 ta:7 supervised:1 box:1 rejected:1 just:1 d:1 dismiss:1 duc:1 su:1 ency:1 propagation:1 widespread:1 mode:1 indicated:2 hor:1 oil:1 usa:1 requiring:1 concept:1 symmetric:1 deal:1 ind:1 reminds:1 ll:3 game:7 transitivity:1 unambiguous:2 yorktown:1 won:1 complete:1 cp:3 recently:1 ltl:1 handcrafted:1 discussed:1 he:24 hie:5 mor:1 doomed:1 refer:1 compa:11 rd:4 unconstrained:1 consistency:4 teaching:1 had:1 poin:1 ute:3 han:4 base:3 gammontool:3 tesauro:7 store:1 pproach:3 rep:1 success:1 tee:1 vt:1 scoring:4 seen:1 impose:1 paradigm:12 ud:1 signal:3 ii:2 llsing:1 full:1 ural:1 ing:1 technical:3 match:2 offer:1 long:1 cross:1 seh:1 divided:1 discov:1 jy:1 inpu:1 ensuring:1 beha:2 involving:2 basic:1 rformance:1 represent:2 invert:1 ion:14 lea:8 whereas:1 addition:1 ore:1 tact:1 else:1 howe:1 crucial:1 ot:5 alterna:10 swapped:1 w2:1 comment:1 recording:1 tend:1 subject:1 ee:1 counting:1 ter:3 curren:1 decent:1 concerned:1 isolation:1 ioll:1 reduce:1 ould:1 politics:1 rud:1 ul:1 wo:2 ntial:2 useful:3 generally:1 se:1 ten:1 category:4 bett:1 reduced:2 egories:1 notice:1 oding:1 per:4 shall:3 key:1 kes:1 hoard:3 ncw:1 vast:1 fraction:1 nce:4 tlh:1 realworld:1 tly:2 ork:1 run:1 striking:1 throughout:1 separation:1 layer:4 hi:1 ct:1 guaranteed:1 played:1 tackled:1 bea:1 exa:1 hig:1 strength:1 constraint:1 anot:1 hy:3 ining:1 speed:1 extremely:2 separable:2 uf:1 gory:1 smaller:1 llum:1 em:3 separability:1 hl:1 explained:1 pr:1 taken:1 legal:1 ln:3 alluded:1 previously:1 boa:2 lta:1 hh:1 needed:2 know:2 eventually:1 ge:1 available:2 opponent:3 appropriate:1 enforce:1 fundion:1 alternative:7 specialist:4 ho:1 slower:1 hat:4 thomas:1 top:1 remaining:1 lat:1 ting:1 prof:1 move:11 already:1 occurs:1 hum:1 fa:3 sha:1 amongst:1 hsp:1 addit:1 me:5 hea:1 considers:1 reason:1 code:2 besides:1 cont:1 nc:1 difficult:5 onl:1 teach:1 stated:1 ba:7 design:1 understandable:1 policy:1 lil:1 perform:3 gallant:1 arc:3 t:1 tment:1 situation:1 severity:1 dc:1 nei:1 introduced:1 tive:8 pair:1 required:4 thr:3 bout:2 huma:1 learned:1 applic:1 ther:1 able:1 pattern:2 perception:1 program:2 ia:1 improv:1 business:1 natural:1 eh:2 rely:1 predicting:1 hr:1 oue:1 scheme:6 tives:8 improve:1 pproa:3 rela:2 coupled:1 lining:3 nsp:1 faced:1 understanding:1 determining:1 relative:4 precedent:1 evalua:2 par:2 ically:1 hoh:1 limitation:1 interesting:1 sil:1 nels:1 whkh:2 principle:1 playing:2 cd:2 nne:2 course:1 side:1 understand:2 nge:1 absolute:3 benefit:1 raining:1 default:2 transition:1 world:2 seemed:1 author:2 qualitatively:1 forward:1 ights:1 far:1 preferred:2 l00:1 opp:2 conclude:1 consuming:1 inpnt:1 spectrum:1 propa:1 constan:1 table:2 learn:2 ca:6 symmetry:3 mol:1 hc:2 complex:1 som:1 domain:11 da:4 thc:3 sp:1 flf:1 did:1 stc:1 linearly:1 ait:1 ule:1 tl:2 en:1 board:6 ny:2 slow:1 n:1 position:23 winning:1 lie:1 learns:1 down:1 tlte:1 specific:1 bad:1 lme:1 ilea:1 list:1 decay:1 alt:3 concern:1 sit:1 consist:3 intrinsic:2 ih:1 nec:1 te:3 hoped:1 positi:1 nk:1 crea:1 easier:3 lt:4 simply:2 rsp:1 contained:1 ch:4 informa:2 tha:12 ma:4 bioi:1 fut:1 considerable:1 gation:1 except:1 principal:1 engineer:1 called:2 secondary:1 engaged:2 e:2 la:1 perceptrons:1 select:5 fter:1 tested:1 |
462 | 1,420 | How to Dynamically Merge Markov
Decision Processes
Satinder Singh
Department of Computer Science
University of Colorado
Boulder, CO 80309-0430
baveja@cs.colorado.edu
David Cohn
Adaptive Systems Group
Harlequin, Inc.
Menlo Park, CA 94025
cohn@harlequin.com
Abstract
We are frequently called upon to perform multiple tasks that compete for our attention and resource. Often we know the optimal
solution to each task in isolation; in this paper, we describe how
this knowledge can be exploited to efficiently find good solutions
for doing the tasks in parallel. We formulate this problem as that of
dynamically merging multiple Markov decision processes (MDPs)
into a composite MDP, and present a new theoretically-sound dynamic programming algorithm for finding an optimal policy for the
composite MDP. We analyze various aspects of our algorithm and
illustrate its use on a simple merging problem.
Every day, we are faced with the problem of doing mUltiple tasks in parallel, each
of which competes for our attention and resource. If we are running a job shop,
we must decide which machines to allocate to which jobs, and in what order, so
that no jobs miss their deadlines. If we are a mail delivery robot, we must find the
intended recipients of the mail while simultaneously avoiding fixed obstacles (such
as walls) and mobile obstacles (such as people), and still manage to keep ourselves
sufficiently charged up.
Frequently we know how to perform each task in isolation; this paper considers how
we can take the information we have about the individual tasks and combine it to
efficiently find an optimal solution for doing the entire set of tasks in parallel. More
importantly, we describe a theoretically-sound algorithm for doing this merging
dynamically; new tasks (such as a new job arrival at a job shop) can be assimilated
online into the solution being found for the ongoing set of simultaneous tasks.
1058
1
S. Singh and D. Cohn
The Merging Framework
Many decision-making tasks in control and operations research are naturally formulated as Markov decision processes (MDPs) (e.g., Bertsekas & Tsitsikiis, 1996). Here
we define MDPs and then formulate what it means to have multiple simultanous
MDPs.
1.1
Markov decision processes (MDPs)
An MDP is defined via its state set 8, action set A, transition probability matrices
P, and payoff matrices R. On executing action a in state s the probability of
transiting to state s' is denoted pa(ss') and the expected payoff associated with
that transition is denoted Ra (ss'). We assume throughout that the payoffs are
non-negative for all transitions. A policy assigns an action to each state of the
MDP. The value of a state under a policy is the expected value of the discounted
sum of payoffs obtained when the policy is followed on starting in that state. The
objective is to find an optimal policy, one that maximizes the value of every state.
The optimal value of state s, V* (s), is its value under the optimal policy.
The optimal value function is the solution to the Bellman optimality equations: for
all s E 8 , V(s) = maxaEA(E s pa(ss') [Ra (ss') +/V(s'))), where the discount factor
o ~ / < 1 makes future payoffs less valuable than more immediate payoffs (e.g.,
Bertsekas & Tsitsiklis, 1996). It is known that the optimal policy 7r* can be determined from V* as follows: 7r*(s) = argmaxaE A(E s pa(ss')[Ra(ss') +/V*(s'))).
Therefore solving an MDP is tantamount to computing its optimal value function.
l
l
1.2
Solving MDPs via Value Iteration
Given a model (8, A, P, R) of an MDP value iteration (e.g., Bertsekas & Tsitsikiis,
1996) can be used to determine the optimal value function. Starting with an initial
guess, Vo, iterate for all s Vk+1(S) = maxaEA(E sIES pa(ss')[Ra(ss') + /Vk(S'))). It
is known that maxsES 1Vk+1 (s) - V*(s)1 ~ / maxsES IVk(S) - V*(s)1 and therefore
Vk converges to V* as k goes to infinity. Note that a Q-value (Watkins, 1989) based
version of value iteration and our algorithm presented below is also easily defined.
1.3
Multiple Simultaneous MDPs
The notion of an optimal policy is well defined for a single task represented as
an MDP. If, however, we have multiple tasks to do in parallel, each with its own
state, action, transition probability, and payoff spaces, optimal behavior is not
automatically defined. We will assume that payoffs sum across the MDPs, which
means we want to select actions for each MDP at every time step so as to maximize
the expected discounted value of this summed payoff over time. If actions can be
chosen independently for each MDP, then the solution to this "composite" MDP
is obvious - do what's optimal for each MDP. More typically, choosing an action
for one MDP constrains what actions can be chosen for the others. In a job shop
for example, actions correspond to assignment of resources, and the same physical
resource may not be assigned to more than one job simultaneously.
Formally, we can define a composite MDP as a set of N MDPs {Mi}f. We will use
superscripts to distinguish the component MDPs, e.g., 8i , Ai, pi, and Ri are the
state, action, transition probability and payoff parameters of MDP Mi. The state
space of the composite MDP, 8, is the cross product of the state spaces of the component MDPs, i.e., 8 = 8 1 X 8 2 X ... X 8 N . The constraints on actions implies that
1059
How to Dynamically Merge Markov Decision Processes
the action set of the composite MDP, A, is some proper subset of the cross product
of the N component action spaces. The transition probabilities and the payoffs of
the composite MDP are factorial because the following decompositions hold: for
all s, s' E S and a E A, pa(ss') = nf:lpai (SiSi') and Ra(ss') = l:~l ~i (SiSi').
Singh (1997) has previously studied such factorial MDPs but only for the case of a
fixed set of components.
The optimal value function of a composite MDP is well defined, and satisfies the
following Bellman equation: for all s E S,
N
V(s)
=
~a:L (nf:lpa'(sisi')[LRa\sisi')+'YV(s')]).
~ES
(1)
i=l
Note that the Bellman equation for a composite MDP assumes an identical discount
factor across component MDPs and is not defined otherwise.
1.4
The Dynamic Merging Problem
Given a composite MDP, and the optimal solution (e.g. the optimal value function)
for each of its component MDPs, we would like to efficiently compute the optimal
solution for the composite MDP. More generally, we would like to compute the
optimal composite policy given only bounds on the value functions of the component
MDPs (the motivation for this more general version will become clear in the next
section). To the best of our knowledge, the dynamic merging question has not been
studied before.
Note that the traditional treatment of problems such as job-shop scheduling would
formulate them as nonstationary MDPs (however, see Zhang and Dietterich, 1995
for another learning approach). This normally requires augmenting the state space
to include a "time" component which indexes all possible state spaces that could
arise (e.g., Bertsekas, 1995). This is inefficient, and potentially infeasible unless we
know in advance all combinations of possible tasks we will be required to solve. One
contribution of this paper is the observation that this type of nonstationary problem
can be reformulated as one of dynamically merging (individually) stationary MDPs.
1.4.1
The naive greedy policy is suboptimal
Given bounds on the value functions of the component MDPs, one heuristic composite policy is that of selecting actions according to a one-step greedy rule:
N
7I"(s) = argmax(l: nf:,lpa i (Si si')[l:(Rai (si, ai ) + 'YXi(Si'))]),
a
8'
i=l
where Xi is the upper or lower bound of the value function, or the mean of the
bounds. It is fairly easy however, to demonstrate that these policies are substantially
suboptimal in many common situations (see Section 3).
2
Dynamic Merging Algorithm
Consider merging N MDPs; job-shop scheduling presents a special case of merging
a new single MDP with an old composite MDP consisting of several factor MDPs.
One obvious approach to finding the optimal composite policy would be to directly
perform value iteration in the composite state and action space. A more efficient
approach would make use of the solutions (bounds on optimal value functions) of
the existing components; below we describe an algorithm for doing this.
S. SinghandD. Cohn
1060
Our algorithm will assume that we know the optimal values, or more generally,
upper and lower bounds to the optimal values of the states in each component
MDP. We use the symbols Land U for the lower and upper bounds; if the optimal
value function for the ith factor MDP is available then Li = U i = V?,i.l
Our algorithm uses the bounds for the component MDPs to compute bounds on
the values of composite states as needed and then incrementally updates and narrows these initial bounds using a form of value iteration that allows pruning of
actions that are not competitive, that is, actions whose bounded values are strictly
dominated by the bounded value of some other action.
Initial State: The initial composite state So is composed from the start state of
all the factor MOPs. In practice (e.g. in job-shop scheduling) the initial composite
state is composed of the start state of the new job and whatever the current state
of the set of old jobs is. Our algorithm exploits the initial state by only updating
states that can occur from the initial state under competitive actions.
Initial Value Step: When we need the value of a composite state S for the first
time. we compute upper and lower bounds to its optimal value as follows: L(s) =
max!1 Li(Si), and U(s) = E~1 Ui(S).
Initial Update Step: We dynamically allocate upper and lower bound storage
space for composite states as we first update them. We also create the initial set of
competitive actions for S when we first update its value as A(s) = A. As successive
backups narrow the upper and lower bounds of successor states, some actions will
no longer be competitive, and will be eliminated from further consideration.
Modified Value Iteration Algorithm:
At step t if the state to be updated is St:
Lt+l(St)
Ut+l(St)
At+l (St)
=
(L pa(sts')[Ra(st. s') + -yLt(s')])
s
max (L pa(sts')[Ra(st, s') + -yUt(s')])
Ua E At(st) AND L pa(sts')[Ra(st, s') + -yUt(s')]
s'
;::: argmax L pb(sts')[Rb(st, s') + -yLt(s')]
max
aEAt{st}
aEAt(st}
J
s'
bEAt(St)
St+l
{
8'
So if s~ is terminal for all Si E s
s' E S such that 3a E A t+1 (St), pa(StS')
> 0 otherwise
The algorithm terminates when only one competitive action remains for each state,
or when the range of all competitive actions for any state are bounded by an indifference parameter ?.
To elaborate, the upper and lower bounds on the value of a composite state are
backed up using a form of Equation 1. The set of actions that are considered
competitive in that state are culled by eliminating any action whose bounded values
is strictly dominated by the bounded value of some other action in At(st). The
next state to be updated is chosen randomly from all the states that have non-zero
1 Recall that unsuperscripted quantities refer to the composite MDP while superscripted
quantities refer to component MDPs. Also, A is the set of actions that are available to the
composite MDP after taking into account the constraints on picking actions simultaneously
for the factor MDPs.
How to Dynamically Merge Markov Decision Processes
1061
pro babili ty of occuring from any action in At+! (St) or, if St is the terminal state of
all component MDPs, then StH is the start state again.
A significant advantage of using these bounds is that we can prune actions whose
upper bounds are worse than the best lower bound. Only states resulting from
remaining competitive actions are backed up. When only one competitive action
remains, the optimal policy for that state is known, regardless of whether its upper
and lower bounds have converged.
Another important aspect of our algorithm is that it focuses the backups on states
that are reachable on currently competitive actions from the start state. The combined effect of only updating states that are reachable from the start state and
further only those that are reachable under currently competitive actions can lead
to significant computational savings. This is particularly critical in scheduling,
where jobs proceed in a more or less feedforward fashion and the composite start
state when a new job comes in can eliminate a large portion of the composite state
space. Ideas based on Kaelbling's (1990) interval-estimation algorithm and Moore
& Atkeson's (1993) prioritized sweeping algorithm could also be combined into our
algorithm.
The algorithm has a number of desirable "anytime" characteristics: if we have to
pick an action in state So before the algorithm has converged (while multiple competitive actions remain), we pick the action with the highest lower bound. If a new
MDP arrives before the algorithm converges, it can be accommodated dynamically
using whatever lower and upper bounds exist at the time it arrives.
2.1
Theoretical Analysis
In this section we analyze various aspects of our algorithm.
UpperBound Calculation: For any composite state, the sum of the optimal
values of the component states is an upper bound to the optimal value of the
composite state, i.e., V*(s = SI, S2, .. . , SN) ~ 2:~1 V*,i(Si).
If there were no constraints among the actions of the factor MDPs then V* (s) would
equal L~l V*,i(Si) because of the additive payoffs across MDPs. The presence of
constraints implies that the sum is an upper bound. Because V*,i(S') ~ Ut(Si) the
result follows.
LowerBound Calculation: For any composite state, the maximum of the optimal
values of the component states is a lower bound to the optimal value of the composite
states, i.e., V*(s = SI, S2, . .. ,SN) ~ max~1 V*,i(Si).
To see this for an arbitrary composite state s, let the MDP that has the largest component optimal value for state s always choose its component-optimal action first
and then assign actions to the other MDPs so as to respect the action constraints
encoded in set A. This guarantees at least the value promised by that MDP because
the payoffs are all non-negative. Because V*,i(Si) ~ Lt(Si) the result follows.
Pruning of Actions: For any composite state, if the upper bound for any composite action, a, is lower than the lower bound for some other composite action,
then action a cannot be optimal - action a can then safely be discarded from the
max in value iteration. Once discarded from the competitive set, an action never
needs to be reconsidered.
Our algorithm maintains the upper and lower bound status of U and L as it updates
them. The result follows.
S. Singh and D. Cohn
1062
Convergence: Given enough time our algorithm converges to the optimal policy
and optimal value function for the set of composite states reachable from the start
state under the optimal policy.
If every state were updated infinitely often, value iteration converges to the optimal
solution for the composite problem independent of the intial guess Vo. The difference
between standard value iteration and our algorithm is that we discard actions and
do not update states not on the path from the start state under the continually
pruned competitive actions. The actions we discard in a state are guaranteed not
to be optimal and therefore cannot have any effect on the value of that state. Also
states that are reachable only under discarded actions are automatically irrelevant
to performing optimally from the start state.
3
An Example: Avoiding Predators and Eating Food
We illustrate the use of the merging algorithm on a simple avoid-predator-andeat-food problem, depicted in Figure 1a. The component MDPs are the avoidpredator task and eat-food task; the composite MDP must solve these problems
simultaneously. In isolation, the tasks avoid-predator and eat-food are fairly easy
to learn. The state space of each task is of size n\ 625 states in the case illustrated.
Using value iteration, the optimal solutions to both component tasks can be learned
in approximately 1000 backups. Directly solving the composite problem requires
n 6 states (15625 in our case), and requires roughly 1 million backups to converge.
Figure 1b compares the performance of several solutions to the avoid-predatorand-eat-food task. The opt-predator and opt-food curves shows the performance
of value iteration on the two component tasks in isolation; both converge qUickly
to their optima. While it requires no further backups, the greedy algorithm of
Section 1.4.1 falls short of optimal performance. Our merging algorithm, when
initialized with solutions for the component tasks (5000 backups each) converges
quickly to the optimal solution. Value iteration directly on the composite state space
also finds the optimal solutions, but requires 4-5 times as many backups. Note that
value iteration in composite state space also updated states on trajectories (as in
Barto etal.'s, 1995 RTDP algorithm) through the state space just as in our merging
algorithm, only without the benefit of the value function bounds and the pruning
of non-competitive actions.
4
Conclusion
The ability to perform multiple decision-making tasks simultaneously, and even
to incorporate new tasks dynamically into ongoing previous tasks, is of obvious
interest to both cognitive science and engineering. Using the framework of MDPs
for individual decision-making tasks, we have reformulated the above problem as
that of dynamically merging MDPs. We have presented a modified value iteration
algorithm for dynamically merging MDPs, proved its convergence, and illustrated
its use on a simple merging task.
As future work we intend to apply our merging algorithm to a real-world jobshop scheduling problem, extend the algorithm into the framework of semi-Markov
decision processes, and explore the performance of the algorithm in the case where
a model of the MDPs is not available.
How to Dynamically Merge Markov Decision Processes
a)
b)
0.80
1063
I
f
0.70
*'~
f
P
Q.
A
& 0.60
j
opt-predator
0.40
0.0
,
I --
- - - - : : - :- . . . . , .- : - - -- - - ' - . , . . . , . . , . --
500000.0
1000000.0
-
---'
1500000.0
Number 01 Backups
Figure 1: a) Our agent (A) roams an n by n grid. It gets a payoff of 0.5 for every time
step it avoids predator (P), and earns a payoff of 1.0 for every piece of food (f) it finds.
The agent moves two steps for every step P makes, and P always moves directly toward
A. When food is found , it reappears at a random location on the next time step. On every
time step, A has a 10% chance of ignoring its policy and making a random move. b) The
mean payoff of different learning strategies vs. number of backups. The bottom two lines
show that when trained on either task in isolation, a learner reaches the optimal payoff for
that task in fewer than 5000 backups. The greedy approach makes no further backups, but
performs well below optimal. The optimal composite solution, trained ab initio, requires
requires nearly 1 million backups. Our algorithm begins with the 5000-backup solutions
for the individual tasks, and converges to the optimum 4-5 times more quickly than the
ab initio solution.
Acknowledgements
Satinder Singh was supported by NSF grant IIS-9711753.
References
Barto, A. G., Bradtke, S. J., & Singh, S. (1995) . Learning to act using real-time dynamic
programming. Artificial Intelligence, 72, 81-138.
Bertsekas, D. P. (1995). Dynamic Programming and Optimal Control. Belmont, MA:
Athena Scientific.
Bertsekas, D. P. & Tsitsiklis, J. N. (1996). Neuro-Dynamic Programming. Belmont, MA:
Athena Scientific.
Kaelbling, L. P. (1990) . Learning in Embedded Systems. PhD thesis, Stanford University,
Department of Computer Science, Stanford, CA. Technical Report TR-90-04.
Moore, A. W . & Atkeson, C. G. (1993). Prioritized sweeping: Reinforcement learning with
less data and less real time. Machine Learning, 19(1).
Singh , S. (1997). Reinforcement learning in factorial environments. submitted.
Watkins, C. J. C. H. (1989). Learning from Delayed Rewards. PhD thesis, Cambridge
Univ., Cambridge, England.
Zhang, W. & Dietterich, T . G. (1995). High-performance job-shop scheduling with a time
delay TD(lambda) network. In NIPSystems 8. MIT Press.
| 1420 |@word version:2 eliminating:1 decomposition:1 pick:2 tr:1 initial:10 selecting:1 existing:1 current:1 com:1 si:14 must:3 belmont:2 additive:1 update:6 v:1 stationary:1 greedy:4 fewer:1 guess:2 intelligence:1 reappears:1 ith:1 short:1 location:1 successive:1 zhang:2 become:1 combine:1 theoretically:2 ra:8 expected:3 roughly:1 behavior:1 frequently:2 terminal:2 bellman:3 discounted:2 automatically:2 food:8 td:1 ua:1 begin:1 competes:1 bounded:5 maximizes:1 what:4 substantially:1 rtdp:1 sisi:4 finding:2 guarantee:1 safely:1 every:8 nf:3 act:1 control:2 normally:1 whatever:2 grant:1 bertsekas:6 continually:1 before:3 engineering:1 path:1 merge:4 approximately:1 studied:2 dynamically:12 co:1 range:1 lowerbound:1 practice:1 composite:41 get:1 cannot:2 scheduling:6 storage:1 charged:1 backed:2 go:1 attention:2 starting:2 independently:1 regardless:1 formulate:3 assigns:1 rule:1 importantly:1 notion:1 updated:4 colorado:2 programming:4 us:1 pa:9 particularly:1 updating:2 bottom:1 highest:1 valuable:1 environment:1 ui:1 constrains:1 reward:1 dynamic:7 trained:2 singh:7 solving:3 upon:1 learner:1 easily:1 various:2 represented:1 univ:1 describe:3 artificial:1 choosing:1 whose:3 heuristic:1 encoded:1 solve:2 reconsidered:1 stanford:2 s:10 otherwise:2 ability:1 superscript:1 online:1 advantage:1 product:2 convergence:2 optimum:2 executing:1 converges:6 yxi:1 illustrate:2 augmenting:1 job:15 c:1 implies:2 come:1 successor:1 assign:1 wall:1 opt:3 strictly:2 hold:1 initio:2 sufficiently:1 considered:1 mop:1 estimation:1 currently:2 individually:1 largest:1 create:1 mit:1 always:2 modified:2 avoid:3 mobile:1 eating:1 barto:2 focus:1 vk:4 ylt:2 entire:1 typically:1 eliminate:1 among:1 denoted:2 summed:1 fairly:2 special:1 equal:1 once:1 saving:1 never:1 eliminated:1 identical:1 park:1 argmaxae:1 nearly:1 future:2 others:1 report:1 randomly:1 composed:2 simultaneously:5 individual:3 delayed:1 intended:1 ourselves:1 argmax:2 consisting:1 ab:2 interest:1 arrives:2 unless:1 old:2 accommodated:1 initialized:1 theoretical:1 tsitsikiis:2 obstacle:2 assignment:1 kaelbling:2 lpai:1 subset:1 intial:1 delay:1 optimally:1 combined:2 st:17 picking:1 quickly:3 yut:2 again:1 thesis:2 manage:1 choose:1 worse:1 cognitive:1 lambda:1 inefficient:1 li:2 account:1 upperbound:1 inc:1 sts:5 piece:1 doing:5 analyze:2 portion:1 yv:1 competitive:15 start:9 parallel:4 maintains:1 predator:6 contribution:1 characteristic:1 efficiently:3 correspond:1 trajectory:1 converged:2 submitted:1 simultaneous:2 reach:1 ty:1 obvious:3 naturally:1 associated:1 mi:2 proved:1 treatment:1 recall:1 knowledge:2 ut:2 anytime:1 day:1 just:1 cohn:5 incrementally:1 scientific:2 mdp:31 dietterich:2 effect:2 assigned:1 moore:2 illustrated:2 occuring:1 demonstrate:1 vo:2 performs:1 bradtke:1 pro:1 consideration:1 common:1 physical:1 million:2 extend:1 refer:2 significant:2 cambridge:2 ai:2 grid:1 etal:1 baveja:1 reachable:5 robot:1 longer:1 own:1 irrelevant:1 discard:2 lra:1 exploited:1 prune:1 determine:1 maximize:1 converge:2 semi:1 ii:1 multiple:8 sound:2 desirable:1 technical:1 england:1 calculation:2 cross:2 deadline:1 neuro:1 iteration:14 want:1 interval:1 nonstationary:2 presence:1 feedforward:1 easy:2 enough:1 iterate:1 isolation:5 jobshop:1 earns:1 suboptimal:2 idea:1 whether:1 allocate:2 reformulated:2 proceed:1 action:52 generally:2 clear:1 factorial:3 discount:2 exist:1 nsf:1 rb:1 group:1 pb:1 promised:1 superscripted:1 sum:4 compete:1 throughout:1 decide:1 delivery:1 decision:11 bound:27 followed:1 distinguish:1 guaranteed:1 occur:1 lpa:2 infinity:1 constraint:5 ri:1 dominated:2 aspect:3 optimality:1 pruned:1 performing:1 eat:3 department:2 transiting:1 according:1 rai:1 combination:1 across:3 terminates:1 remain:1 sth:1 making:4 boulder:1 resource:4 equation:4 previously:1 remains:2 needed:1 know:4 available:3 operation:1 apply:1 recipient:1 assumes:1 running:1 include:1 remaining:1 exploit:1 objective:1 intend:1 question:1 quantity:2 move:3 strategy:1 traditional:1 athena:2 mail:2 considers:1 toward:1 index:1 culled:1 potentially:1 negative:2 proper:1 policy:17 perform:4 upper:14 observation:1 markov:8 discarded:3 beat:1 immediate:1 payoff:17 situation:1 sweeping:2 arbitrary:1 david:1 required:1 learned:1 narrow:2 below:3 maxaea:2 max:7 critical:1 shop:7 mdps:32 naive:1 sn:2 faced:1 acknowledgement:1 tantamount:1 embedded:1 agent:2 pi:1 land:1 supported:1 infeasible:1 tsitsiklis:2 fall:1 taking:1 benefit:1 curve:1 transition:6 world:1 avoids:1 adaptive:1 reinforcement:2 atkeson:2 pruning:3 status:1 keep:1 satinder:2 xi:1 learn:1 assimilated:1 ca:2 menlo:1 ignoring:1 motivation:1 backup:13 arise:1 arrival:1 s2:2 elaborate:1 fashion:1 harlequin:2 watkins:2 symbol:1 merging:17 phd:2 depicted:1 lt:2 explore:1 infinitely:1 indifference:1 satisfies:1 chance:1 ma:2 formulated:1 prioritized:2 determined:1 miss:1 called:1 e:1 select:1 formally:1 people:1 ongoing:2 incorporate:1 avoiding:2 |
463 | 1,421 | Bayesian model of surface perception
William T. Freeman
MERL, Mitsubishi Electric Res. Lab .
201 Broadway
Cambridge, MA 02139
Paul A. Viola
Artificial Intelligence Lab
Massachusetts Institute of Technology
Cambridge, MA 02139
freeman~erl.com
viola~ai.mit.edu
Abstract
Image intensity variations can result from several different object
surface effects, including shading from 3-dimensional relief of the
object, or paint on the surface itself. An essential problem in vision ,
which people solve naturally, is to attribute the proper physical
cause, e.g. surface relief or paint, to an observed image. We addressed this problem with an approach combining psychophysical
and Bayesian computational methods.
We assessed human performance on a set of test images, and found
that people made fairly consistent judgements of surface properties.
Our computational model assigned simple prior probabilities to
different relief or paint explanations for an image, and solved for
the most probable interpretation in a Bayesian framework. The
ratings of the test images by our algorithm compared surprisingly
well with the mean ratings of our subjects.
1
Introduction
When people study a picture, they can judge whether it depicts a shaded , 3dimensional surface, or simply a flat surface with markings or paint on it. The
two images shown in Figure 1 illustrate this distinction [1]. To many observers
Figure 1a appears to be a raised plateau lit from the left. Figure 1b is simply a
re-arrangement of the local features of la, yet it does not give an impression of
shape or depth. There is no simple correct answer for this problem; either of these
images could be explained as marks on paper, or as illuminated shapes. Nevertheless people tend to make particular judgements of shape or reflectance . We seek an
algorithm to arrive at those same judgements.
There are many reasons to study this problem. Disentangling shape and reflectance
is a prototypical underdetermined vision problem, which biological vision systems
routinely solve. Insights into this problem may apply to other vision problems
W T. Freeman and P. A. Viola
788
as well. A machine that could interpret images as people do would have many
applications , such as the interactive editing and manipulation of images. Finally,
there is a large body of computer vision work on "shape from shading"-inferring
the 3-dimensional shape of a shaded object [4]. Virtually every algorithm assumes
that all image intensity changes are caused by shading; these algorithms fail for any
image with reflectance changes. To bring this body of work into practical use , we
need to be able to disambiguate shading from reflectance changes.
There has been very little work on this problem. Sinha and Adelson [9] examined
a world of painted polyhedra, and used consistancy constraints to identify regions
of shape and reflectance changes. Their consistancy constraints involved specific
assumptions which need not always hold and may be better described in a probabilistic framework . In addition , we seek a solution for more general, greyscale
lmages.
Our approach combines psychophysics and computational modeling. First we will
review the physics of image formation and describe the under-constrained surface
perception problem. We then describe an experiment to measure the interpretations
of surface shading and reflectance among different individuals. We will see that the
judgements are fairly consistent across individuals and can be averaged to define
"ground truth" for a set of test images. Our approach to modeling the human
judgements is Bayesian . We begin by formulating prior probabilities for shapes and
reflectance images , in the spirit of recent work on the statistical modeling of images
[5 , 8, 11]. Using these priors, the algorithm then determines whether an image is
more likely to have been generated by a 3D shape or as a pattern of reflectance.
We compare our algorithm's performance to that of the human subjects.
(a)
(c )
(b)
(d)
Figure 1: Images (a) and (b), designed by Adelson [1] , are nearly the same
everywhere, yet give different percepts of shading and reflectance . (a) looks like
a plateau, lit from the left ; (b) looks like marks on paper. Illustrating the underconstrained nature of perception , both images can be explained either by reflectance
changes on paper (they are), or, under appropriate lighting conditions, by the
shapes (c) and (d) , respectively (vertical scale exaggerated).
Bayesian Model of Surface Perception
2
789
Physics of Imaging
One simple model for the generation of an image from a three dimensional shape is
the Lambertian model:
I(x, y)
= R(x, y)
(i. n(x, y)) ,
(1)
where I (x, y) is an image indexed by pixel location , n( x, y) is the surface normal at
every point on the surface conveniently indexed by the pixel to which that surface
patch projects, i is a unit vector that points in the direction of the light source,
and R( x, y) is the reflectance at every point on the surface 1 . A patch of surface
is brighter if the light shines onto it directly and darker if the light shines on it
obliquely. A patch can also be dark simply because it is painted with a darker
pigment. The shape of the object is probably more easily described as a depth map
z(x, y) from which n(x, y) is computed.
The classical "shape from shading" task attempts to compute z from I given knowledge of i and assuming R is everywhere constant. Notice that the problem is "illposed"; while I( x, y) does constrain n( x, y) it is not sufficient to uniquely determine
the surface normal at each pixel. Some assumption about global properties of z is
necessary to condition the problem. If R is allowed to vary, the problem becomes
even more under-constrained. For example, R = I and n( x, y) = j is a valid solution
for every image. This is the "all reflectance" hypothesis, where the inferred surface
is flat and? all of the image variation is due to reflectance. Interestingly there is also
an "all shape" solution for every image where R =. 1 and I(x, y) = i? n(x, y) (see
Figure 1 for examples of such shapes).
Since the relationship between z and I is non-linear, "shape from shading" cannot
be solved directly and requires a time consuming search procedure. For our computational experiments we seek a rendering model for shapes which simplifies the
mathematics, yet maintains the essential ambiguities of the problem. We use the approximations of linear shading [6]. This involves two sets of approximations. First,
that the rendered image I (x, y) is some function, G (
only of the surface
slope at any point:
g; , g; ),
(2)
The second approximation is that the rendering function G itself is a linear function
of the surface slopes,
oz oz
G( ox' Oy) ~ kl
oz
oz
+ k2 ox + k3 oy .
(3)
Under linear shading, finding a shape which explains a given image is a trivial
integration along the direction of the assumed light source. Despite this simplicity,
images rendered under linear shading appear fairly realistically shaded [6].
3
Psychophysics
We used a survey to assess subjects' image judgements. We made a set of 60 test
images, using Canvas and Photoshop programs to generate and manipulate the
images. Our goal was to create a set of images with varying degrees of shadedness.
We sought to assess to what extent each subject saw each image as created by
1
Note: we assume orthographic projection, a distant light source, and no shadowing.
W T. Freeman and P. A. Vwla
790
shading changes or reflectance changes. Each of our 18 naive observers was given a
4 page survey showing the images in a different random order.
To explain the problem of image interpretation quickly to naive subjects , we used
a concrete story (Adelson's Theater Set Shop analogy [2] is a related didactic example). The survey instructions were as follows :
Pretend that each of the following pictures is a photograph of work
made by either a painter or a sculptor.
The painter could use paint , markers, air brushes, computer, etc. ,
to make any kind of mark on a fiat canvas. The paint had no
3-dimensionality; everything was perfectly fiat.
The sculptor could make 3-dimensional obJects, but could make no
markings on them . She could mold, sculpt, and scrape her sculptures, but could not draw or paint. All the objects were made out
of a uniform plaster material and were made visible by lighting and
shading effects.
The subjects used a 5-point rating scale to indicate whether each image was made
by the painter (P) or sculptor (S): S, S? , ?, P?, P.
3.1
Survey Results
We examined a non-parametric comparison of the image ratings, the rank order
correlation (the linear correlation of image rankings in order of shapeness by each
observer) [7]. Over all possible pairings of subjects, the rank order correlations
ranged from 0.3 to 0.9, averaging 0.65. All of these correlations were statistically
significant , most at the 0.0001 level. We concluded that for our set of test images ,
people do give a very similar set of interpretations of shading and reflectance.
We assigned a numerical value to each of the 5 survey responses (S=2; S?=l; ?=O;
P?=-l; P=-2) and found the average numerical "shadedness" score for each image.
Figure 2 shows a histogram of the survey responses for each image , ordered in
decreasing order of shadedness. The two images of Figure 1 had average scores of 1.7
and -1.6 , respectively, confirming the impressions of shading and reflectance . There
was good consensus for the rankings of the most paint-like and most sculpturelike images; the middle images showed a higher score variance . The rankings by
each individual showed a strong correlation with the rankings by the average of the
remaining subjects, ranging from 0.6 to 0.9 . Figure 4 shows the histogram of those
correlations. The ordering of the images by the average of the subjects ' responses
provides a "ground truth" with which to compare the rankings of our algorithm.
Figure 3, left , shows a randomly chosen subset of the sorted images , in decreasing
order of assessed sculptureness .
4
Algorithm
We will assume that people are choosing the most probable interpretation of the
observed image. We will adopt a Bayesian approach and calculate the most probable
interpretation for each image under a particular set of prior probabilities for images
and shapes. To parallel the choices we gave our subjects , we will choose between
interpretations that account for the image entirely by shape changes, or entirely
by reflectance changes. Thus, our images are either a rendered shape , multiplied
by a uniform reflectance image , or a flat shape , multiplied by some non-uniform
reflectance image.
Bayesian Model of Surface Perception
791
intensity: score frequency for each image
l!!S
o
(,)
cnp
10
20
30
40
50
60
image number (sorted by shapeness)
Figure 2: Histogram of survey responses. Intensity shows the number of responses
of each score (vertical scale) for each image (horizontal, sorted in increasing order
of shape ness ).
To find the most probable interpretation, given an image, we need to assign prior
probabilities to shape and reflectance configurations. There has been recent interest in characterizing the probabilities of images by the expected distributions of
subband coefficient values [5, 8, 11]. The statistical distribution of bandpass linear
filter outputs, for natural images, is highly kurtotic; the output is usually small ,
but in rare cases it takes on very large values. This non-gaussian behavior is not a
property of the filter operation, because filtered "random" images appear gaussian.
Rather it is a property of the structure of natural images. An exponential distribution, P(c) ex: e- 1cl , where c is the filter coefficient value, is a reasonable model.
These priors have been used in texture synthesis, noise removal, and receptive field
modeling. Here, we apply them to the task of scene interpretation.
We explored using a very simple image prior:
P(I) ex: exp (-
L
r,y
oI(x,y)2 + OI(x , y)2)
ox
oy
(4)
Here we treat the image derivative as an image subband corresponding to a very
simple filter. We applied this image prior to both reflectance images, I(x , y), as
well as range images , z(x, y).
For any given picture, we seek to decide whether a shape or a reflectance explanation
is more probable. The proper Bayesian approach would be to integrate the prior
probabilities of all shapes which could explain the image in order to arrive at the
total probability of a shape explanation. (The reflectance explanation, R is unique ;
the image itself). We employed a computationally simpler procedure, a very rough
approximation to the proper calculation: we evaluated the prior probability, P(S)
of the single, most probable shape explanation, S, for the image. Using the ratio
test of a binary hypothesis, we formed a shapeness index, J, by the ratio of the
probabilities for the shape and reflectance explanations, J = ~i!~. The index J
was used to rank the test images by shapeness.
We need to find the most probable shape explanation. The overall log likelihood of
a shape, z, given an image is, using the linear shading approximations of Eq. (3):
log P(z, kl' k2, k31I)
+ log P(z) + c
Lx,y(I - kl + k2* + k3~~)2 + Lx,y
log P(Ilz, kl' k2, k3)
Jg~2 +
g;2
+ c,
(5)
where c is a normalization constant. We use a multi-scale gradient descent algorithm
that simultaneously determines the optimal shape and illumination parameters for
an image (similar to that used by [10]). The optimization procedure has three
stages starting with a quarter resolution version of I, and moving to the half and
W. T. Freeman and P. A. Vwla
792
Figure 3: 28 of the 60 test images, arranged in decreasing order of subjects'
shapeness ratings. Left: Subjects' rankings. Right: Algorithm 's rankings .
then full resolution. The solution found at the low resolution is interpolated up to
the next level and is used as a starting point for the next step in the optimization.
In our experiments images are 128x128 pixels. The optimization procedure takes
4000 descent steps at each resolution level.
5
Results
Surprisingly, the simple prior probability of Eq. (4) accounts for much of the ratings
of shading or paint by our human subjects. Figure 3 compares the rankings (shown
in raster scan order) of a subset of the test images for our algorithm and the average
of our subjects. The overall agreement is good . Figure 4 compares two measures: (1)
the correlations (dark bars) of the subjects' individual ratings to the mean subject
rating with (2) the correlation of our algorithm's ratings to the mean subject rating.
SUbjects show correlations between 0.6 and 0.9; our Bayesian algorithm showed a
correlation of 0.64. Treating the mean subjects' ratings as the right answer, our
algorithm did worse than most subjects but not as badly as some subjects.
Figure 1 illustrates how our algorithm chooses an interpretation for an image. If a
simple shape explains an image, such as the shape explanation (c) for image (a),
the shape gradient penalties will be small, assigning a high prior probability to that
shape. If a complicated shape (d) is required to explain a simple image (b), the
Bayesian Model of Surface Perception
793
low prior probability of the shape and the high prior probability of the reflectance
image will favor a "paint" explanation.
We noted that many of the shapes inferred from paint-like images showed long
ridges coincidently aligned with the assumed light direction. The assumption of
generic light direction can be applied in a Bayesian framework [3] to penalize such
coincidental alignments. We speculate that such a term would further penalize
those unlikely shape interpretations and may improve algorithm performance.
Rank order corralaUon with mean Imago rating
0.2
Figure 4: Correlation of individual subjects' image ratings compared with the
mean rating (bars) compared with correlation of algorithm's rating with the mean
rating (dashed line).
Acknow ledgements
We thank E. Adelson, D. Brainard, and J. Tenenbaum for helpful discussions.
References
[1] E. H. Adelson , 1995. personal communication.
[2] E. H. Adelson and A. P. Pentland. The perception of shading and reflectance. In
B. Blum, editor, Channels in the Visual Nervous System: Neurophysiology, Psychophysics, and Models, pages 195-207. Freund Publishing, London, 1991.
[3] W. T. Freeman. The generic viewpoint assumption in a framework for visual perception. Nature, 368(6471):542-545, April 7 1994.
[4] B. K. P. Horn and M. J. Brooks. Shape from shading. MIT Press, Cambridge, MA,
1989.
[5] B. A . Olshausen and D. J. Field. Emergence of simple-cell receptive field properties
by learning a sparse code for ?natural images. Nature, 381:607-609, 1996.
[6] A. P. Pentland. Linear shape from shading. Inti. J. Compo Vis., 1(4):153-162, 1990.
[7] W. H. Press, S. A . Teukolsky, W. T. Vetterling, and B. P. Flannery. Numerical Recipes
in C. Cambridge Univ. Press, 1992.
[8] E. P. Simoncelli and E. H. Adelson. Noise removal via Bayesian wavelet coring. In
3rd Annual Inti. Con/. on Image Processing, Laussanne, Switzerland, 1996. IEEE.
[9] P. Sinha and E. H. Adelson . Recovering reflectance and illumination in a world of
painted polyhedra. In Proc. 4th Intl. Con/. Computer Vision, pages 156-163. IEEE,
1993.
[10] D. Terzopoulos. Multilevel computational processes for visual surface reconstruction.
Comp o Vis., Graphics, Image Proc., 24:52-96 , 1983.
[11] S. C. Zhu and D. Mumford. Learning generic prior models for visual computation.
Submitted to IEEE Trans. PAMI, 1997.
| 1421 |@word neurophysiology:1 illustrating:1 version:1 middle:1 judgement:6 instruction:1 seek:4 mitsubishi:1 shading:20 configuration:1 score:5 interestingly:1 com:1 yet:3 assigning:1 distant:1 visible:1 numerical:3 confirming:1 shape:42 designed:1 treating:1 intelligence:1 half:1 nervous:1 compo:1 filtered:1 provides:1 location:1 lx:2 simpler:1 x128:1 along:1 pairing:1 combine:1 expected:1 behavior:1 multi:1 photoshop:1 freeman:6 decreasing:3 little:1 increasing:1 becomes:1 begin:1 project:1 what:1 kind:1 coincidental:1 finding:1 every:5 interactive:1 k2:4 unit:1 appear:2 local:1 treat:1 despite:1 painted:3 pami:1 examined:2 shaded:3 range:1 statistically:1 averaged:1 practical:1 unique:1 horn:1 orthographic:1 illposed:1 procedure:4 projection:1 onto:1 cannot:1 map:1 starting:2 survey:7 resolution:4 simplicity:1 insight:1 theater:1 erl:1 variation:2 hypothesis:2 agreement:1 observed:2 solved:2 calculate:1 region:1 ordering:1 personal:1 easily:1 routinely:1 univ:1 describe:2 london:1 artificial:1 formation:1 choosing:1 sculpt:1 solve:2 favor:1 emergence:1 itself:3 reconstruction:1 shine:2 aligned:1 combining:1 oz:4 realistically:1 obliquely:1 recipe:1 intl:1 object:6 brainard:1 illustrate:1 eq:2 strong:1 recovering:1 involves:1 judge:1 indicate:1 direction:4 switzerland:1 correct:1 attribute:1 filter:4 human:4 material:1 everything:1 explains:2 multilevel:1 assign:1 probable:7 biological:1 underdetermined:1 hold:1 ground:2 normal:2 exp:1 k3:3 vary:1 sought:1 adopt:1 proc:2 shadowing:1 saw:1 create:1 mit:2 rough:1 always:1 gaussian:2 rather:1 varying:1 she:1 laussanne:1 polyhedron:2 rank:4 likelihood:1 helpful:1 vetterling:1 unlikely:1 her:1 pixel:4 overall:2 among:1 raised:1 constrained:2 fairly:3 psychophysics:3 integration:1 ness:1 field:3 lit:2 adelson:8 look:2 nearly:1 randomly:1 simultaneously:1 individual:5 relief:3 william:1 attempt:1 interest:1 highly:1 alignment:1 light:7 necessary:1 indexed:2 re:2 sinha:2 merl:1 modeling:4 kurtotic:1 subset:2 rare:1 uniform:3 graphic:1 answer:2 chooses:1 probabilistic:1 physic:2 synthesis:1 quickly:1 concrete:1 ambiguity:1 choose:1 worse:1 derivative:1 account:2 speculate:1 coefficient:2 caused:1 ranking:8 vi:2 observer:3 lab:2 maintains:1 parallel:1 complicated:1 slope:2 ass:2 formed:1 painter:3 air:1 oi:2 variance:1 percept:1 identify:1 bayesian:12 lighting:2 comp:1 submitted:1 plateau:2 explain:3 raster:1 frequency:1 involved:1 naturally:1 con:2 massachusetts:1 knowledge:1 dimensionality:1 fiat:2 appears:1 higher:1 response:5 april:1 editing:1 arranged:1 evaluated:1 ox:3 stage:1 correlation:12 canvas:2 horizontal:1 marker:1 olshausen:1 effect:2 ranged:1 assigned:2 uniquely:1 noted:1 impression:2 ridge:1 bring:1 image:84 ranging:1 quarter:1 physical:1 interpretation:11 interpret:1 significant:1 cambridge:4 ai:1 rd:1 mathematics:1 had:2 jg:1 moving:1 surface:22 etc:1 recent:2 showed:4 exaggerated:1 manipulation:1 binary:1 employed:1 determine:1 dashed:1 full:1 simoncelli:1 mold:1 calculation:1 long:1 manipulate:1 vision:6 histogram:3 normalization:1 cell:1 penalize:2 addition:1 addressed:1 source:3 concluded:1 probably:1 subject:22 tend:1 virtually:1 spirit:1 rendering:2 gave:1 brighter:1 perfectly:1 simplifies:1 whether:4 penalty:1 cause:1 dark:2 tenenbaum:1 ilz:1 generate:1 notice:1 ledgements:1 didactic:1 nevertheless:1 blum:1 imaging:1 everywhere:2 arrive:2 reasonable:1 decide:1 patch:3 draw:1 illuminated:1 entirely:2 annual:1 badly:1 constraint:2 constrain:1 scene:1 flat:3 interpolated:1 formulating:1 rendered:3 marking:2 across:1 explained:2 inti:2 computationally:1 fail:1 operation:1 multiplied:2 apply:2 lambertian:1 appropriate:1 generic:3 assumes:1 remaining:1 publishing:1 shadedness:3 pretend:1 reflectance:27 subband:2 classical:1 psychophysical:1 paint:11 arrangement:1 mumford:1 parametric:1 receptive:2 gradient:2 thank:1 sculpture:1 extent:1 consensus:1 trivial:1 reason:1 assuming:1 code:1 index:2 relationship:1 ratio:2 disentangling:1 greyscale:1 broadway:1 acknow:1 proper:3 vertical:2 descent:2 pentland:2 viola:3 communication:1 intensity:4 inferred:2 rating:16 required:1 kl:4 distinction:1 brook:1 trans:1 able:1 bar:2 usually:1 perception:8 pattern:1 program:1 including:1 explanation:9 natural:3 zhu:1 shop:1 improve:1 technology:1 picture:3 created:1 naive:2 prior:15 review:1 removal:2 freund:1 prototypical:1 generation:1 oy:3 analogy:1 integrate:1 degree:1 coring:1 sufficient:1 consistent:2 editor:1 story:1 viewpoint:1 surprisingly:2 terzopoulos:1 institute:1 characterizing:1 sparse:1 depth:2 world:2 valid:1 made:6 global:1 assumed:2 consuming:1 search:1 disambiguate:1 nature:3 channel:1 cl:1 electric:1 did:1 noise:2 paul:1 allowed:1 scrape:1 body:2 depicts:1 darker:2 inferring:1 bandpass:1 exponential:1 vwla:2 wavelet:1 specific:1 showing:1 explored:1 essential:2 underconstrained:1 texture:1 illumination:2 illustrates:1 imago:1 flannery:1 photograph:1 simply:3 likely:1 visual:4 conveniently:1 ordered:1 brush:1 truth:2 determines:2 teukolsky:1 ma:3 goal:1 sorted:3 change:9 averaging:1 total:1 la:1 people:7 mark:3 scan:1 assessed:2 ex:2 |
464 | 1,422 | An application of Reversible-J ump
MCMC to multivariate spherical Gaussian
mixtures
Alan D. Marrs
Signal & Information Processing Dept.
Defence Evaluation & Research Agency
Gt. Malvern, UK WR14 3PS
marrs@signal.dra.hmg.gb
Abstract
Applications of Gaussian mixture models occur frequently in the
fields of statistics and artificial neural networks. One of the key
issues arising from any mixture model application is how to estimate the optimum number of mixture components. This paper
extends the Reversible-Jump Markov Chain Monte Carlo (MCMC)
algorithm to the case of multivariate spherical Gaussian mixtures
using a hierarchical prior model. Using this method the number
of mixture components is no longer fixed but becomes a parameter of the model which we shall estimate. The Reversible-Jump
MCMC algorithm is capable of moving between parameter subspaces which correspond to models with different numbers of mixture components. As a result a sample from the full joint distribution of all unknown model parameters is generated. The technique
is then demonstrated on a simulated example and a well known
vowel dataset.
1
Introduction
Applications of Gaussian mixture models regularly appear in the neural networks
literature. One of their most common roles in the field of neural networks, is in
the placement of centres in a radial basis function network. In this case the basis
functions are used to model the distribution of input data (Xi == [Xl, X2, .?? , Xd]T,
(i = l,n)), and the problem is one of mixture density estimation.
A D. Marrs
578
k
p(Xi ) =
L 7Ijp(X I9j),
i
(1)
j=1
where k is the number of mixture components, 7rj the weight or mixing proportion for component j and 8 j the component parameters (mean & variance in this
case). The mixture components represent the basis functions of the neural network
and their parameters (centres & widths) may be estimated using the expectationmaximisation (EM) algorithm.
One of the key issues arising in the use of mixture models is how to estimate the
number of components. This is a model selection problem: the problem of choosing
the 'correct' number of components for a mixture model. This may be thought of
as one of comparing two (or more) mixture models with different components, and
choosing the model that is 'best' based upon some criterion. For example, we might
compare a two component model to one with a single component.
(2)
This may appear to be a case of testing of nested hypotheses. However, it has
been noted [5] that the standard frequentist hypothesis testing theory (generalised
likelihood ratio test) does not apply to this problem because the desired regularity
conditions do not hold. In addition, if the models being tested have 2 and 3 components respectively, they are not strictly nested. For example, we could equate any
pair of components in the three component model to the components in the two
component model, yet how do we choose which component to 'leave out'?
2
Bayesian approach to Gaussian mixture models
A full Bayesian analysis treats the number of mixture components as one of the
parameters of the model for which we wish to find the conditional distribution. In
this case we would represent the joint distribution as a hierarchical model where we
may introduce prior distributions for the model parameters, ie.
p(k, 7r, Z, 9, X) = p(k)p(7rlk)p(zl7r, k)p(9Iz, 7r, k)p(XI9, z, 7r, k),
(3)
where7r = (7rj)J=1, 9 = (9 j )J=1 and z = (Zi)f::l are allocation variables introduced
by treating mixture estimation as a hidden data problem with Zi allocating the ith
observation to a particular component. A simplified version of this model can be
derived by imposing further conditional independencies, leading to the following
expression for the joint distribution
p(k, 7r, Z, 9, X)
= p(k)p(7rlk)p(zl7r, k)p(9Ik)p(XI9, z).
(4)
In addition, we add an extra layer to the hierarchy representing priors on the model
parameters giving the final form for the joint distribution
peA, 6, T}, k, 7r, Z, 9, X) = p(A)p(6)p(T})p(kIA)p(7rlk, 6)p(zl7r, k) x
p(9Ik, T})p(XI9, z).
(5)
Until recently a full Bayesian analysis has been mathematically intractable. Model
comparison was carried out by conducting an extensive search over all possible
Reversible-Jump MCMC for Multivariate Spherical Gaussian Mixtures
579
model orders comparing Bayes factors for all possible pairs of models. What we
really desire is a method which will estimate the model order along with the other
model parameters. Two such methods based upon Markov Chain Monte Carlo
(MCMC) techniques are reversible-jump MCMC [2] and jump-diffusion [3].
In the following sections, we extend the reversible-jump MCMC technique to multivariate spherical Gaussian mixture models. Results are then shown for a simulated
example and an example using the Peterson-Barney vowel data.
3
Reversible-jump MCMC algorithm
Following [4) we define the priors for our hierarchical model and derive a set of
5 move types for the reversible jump MCMC sampling scheme. To simplify some
of the MCMC steps we choose a prior model where the prior on the weights is
is that they are
Dirichlet and the prior model for IJ.j = [JLji' .. .,JLjclV and
drawn independently with normal and gamma priors,
U;2
(6)
where for the purposes of this study we follow[4] and define the hyper-parameters
thus: 6 = 1.0; 'TJ is set to be the mean of the data; A is the diagonal precision
matrix for the prior on IJ.j with components aj which are taken to be liT] where
Tj is the data range in dimension j; a = 2.0 and (3 is some small multiple of liT;'
The moves then consist of: I: updating the weights; II: updating the parameters
(IJ., u); III: updating the allocation; IV: updating the hyper-parameters; V: splitting one component into two, or combining two into one.
The first 4 moves are relatively simple to define, since the conjugate nature of the
priors leads to relatively simple forms for the full conditional distribution of the
desired parameter. Thus the first 4 moves are Gibbs sampling moves and the full
conditional distributions for the weights 1rj, means Jij, variances Uj and allocation
variables Zi are given by:
(7)
where
nk
is the number of observations allocated to component k;
d
p(ltjl .. ?) =
-2
-
II P(JLjml .. ?) : p(JLj..'!-.. ) '" N( njXimUj
-2
m=1
(njuj
+ am'f/m
-2
-1
,(njuj + am) ),
+ am)
(8)
where we recognise that IJ.j is an d dimensional vector with components JL;m (m =
1, d), 'f/m are the components of the Itj prior mean and am represent the diagonal
components of A.
-2
p(u j
\... )
1
== r(lI + nj - 1, '2
n
~
L-
(9)
i=I :Zi;=l
and
(10)
580
A. D. Marrs
The final move involves splitting/combining model components. The main criteria
which need to be met when designing these moves are that they are irreducible,
aperiodic, form a reversible pair and satisfy detailed balance [1]. The MCMC step
for this move takes the form of a Metropolis-Hastings step where a move from state
y to state y' is proposed, with 1r(Y) the target probability distribution and qm(Y, Y')
the proposal distribution for the move m. The resulting move is then accepted with
probability am
1r(Y')qm(y/,y)}
(11)
am - mtn , 1r()
Y qm (y, Y') .
_ . {I
In the case of a move from state Y to a state y' which lies in a higher dimensional
space, the move may be implemented by drawing a vector of continuous random
variables u, independent of y. The new state y' is then set using an invertible deterministic function of x and u. It can be shown [2] that the acceptance probability
is then given by
. {
1r(y')Tm{y')
8y' }
(12)
am=mm 1'1r(y)Tm (y)q{u)1 8 (y,u)1 ,
where Tm(Y) is the probability of choosing move type m when in state y, and q(u)
is the density function of u.
The initial application of the reversible jump MCMC technique to normal mixtures
[4J was limited to the univariate case. This yielded relatively simple expressions for
the split/combine moves, and, most importantly, the determinant ofthe Jacobian of
the tra~formation from a model with k components to one with k + 1 components
was simple to derive. In the more general case of multivariate normal models care
must be taken in prescribing move transformations. A complicated transformation
will lead to problems when the !Jacobian I for a d-dimensional model is required.
For multivariate spherical Gaussian models, we randomly choose a model component from the current k component model. The decision is then made to split or
combine with one of its neighbours with probability P'k and PCIr respectively (where
PCk = 1- Pile)' If the choice is to combine the component, we label the chosen component Zl, and choose Z2 to be a neighbouring component i with probability Q( l/T;
where Tj is the distance from the component Zl. The new component resulting from
the combination of Zl and Z2 is labelled Zc and its parameters are calculated from:
(13)
If the decision is to split, the chosen component is labelled Zc and it is used to define
two new model components Zl and Z2 with weights and parameters conforming to
(13). In making this transformation there are 2 + d degrees of freedom, so we need
to generate 2 + d random numbers to enable the specification of the new component
parameters. The random numbers are denoted u}, U2 = [U211 ... , u2dlT and U3.
All are drawn from Beta{2,2) distributions while the components of U2 each have
probability 0.5 of being negative. The split transformation is then defined by:
Reversible-Jwnp MCMC for Multivariate Spherical Gaussian Mixtures
581
2
7r ZI
(14)
, UZ2
= (1 - U3 )U2Zc .
7r Z2
Once the new components have been defined it is necessary to evaluate the probability of choosing to combine component ZI with component Z2 in this new model.
Having proposed the split/combine move all that remains is to calculate the
Metropolis-Hastings acceptance probability (t, where (t = min(I, R) for the split
move and (t = min(I, 1/R) for the combine move. Where in the case of a split move
from a model with k components to one with k+ 1 components, or a combine move
from k + 1 to k, R is given by:
n~
p(X, le,e) n~
n
._l:?ij-?h;-I:?,,-.c p~~I,:~e,:;2
R=
p(X,le,e)
X
o-l+nl 0-I+n2
11' '"
1r "2
'II"!c 1+nl +n2 B(6,k6) x
n~;::1 J(~;) exp ( -~am ((/LZl m -11m)2 + (/LZ2 m -
11m? - (JLzcm - 11m)) )
X
&(c7(';~2) (a-I) exp ( -f3(u;..2 +u~2 -u;:2)) X
n;=1 g2,2(U2,)) X
c7d+1
p:c::;;oc (g2,2(Ut)gl,1 (U3)
'II"
?c ?c
(2?I-uI)uI)(d+ 1 )/2 J(I- u 3)u3) ,
(15)
where g2,20 denotes a Beta(2,2) density function. The first line on the R.H.S is
due to the ratio of likelihoods for those observations assigned to the components in
question, the subsequent three lines are due to the prior ratios, the fifth line is due to
the the proposal ratio and the last line due to the IJacobian I of the transformation.
The term Palloe represents a combination of the probability of obtaining the current
allocation of data to the components in question and the probability of choosing to
combine components Zl and Z2.
4
Results
To assess this approach to the estimation of multivariate spherical Gaussian mixture models, we firstly consider a toy problem where 1000 bivariate samples were
generated from a known 20 component mixture model. This is followed by an analysis of the Peterson-Barney vowel data set comprising 780 samples of the measured
amplitUde of four formant frequencies for 10 utterances. For this mixture estimation example, we ignore the class labels and consider the straight forward density
estimation problem.
4.1
Simulated data
The resulting reversible-jump MCMC chain of model order can be seen in figure
1, along with the resulting histogram (after rejecting the first 2000 MCMC sampies). The histogram shows that the maximum a posteriori value for model order
is 17. The MAP estimate of model parameters was obtained by averaging all the
17 component model samples, the estimated model is shown in figure 2 alongside
the original generating model. The results are rather encouraging given the large
number of model components and the relatively small number of samples.
A. D. Marrs
582
200
100
19
20
{k}
Iteration
Figure 1: Reversible-jump MCMC chain and histogram of model order for simulated
data.
]0
20
.
',a,.1
.'
. .......~~ \.::0:''.~. '?".'
.'
~.....?
-. \
.,:,
??
.'
.
'. \0"
.....
10.?.~???? ? ? ? ?
~'&.""
.. w:
" :. .
-20
itt
.
.....
....
S,.'
.;~:!
: . :,~'
.
,.'~,
.~.:'
-30
Cenerating Modal
10
-20
-10
.~.
/I
/1.\
.,.
?
~
-10
: ...
. . ~.
.~~
, '.l' .
??
-]0
"
,.' ?.
.. .yt.'
. -.
~~:!
-20 ',~ ?
?
. ie'';':', ,
,T...,.
~
: I.
-]0
:.'0
.~.
~.
tI.-'
;
.
,,;,~.,
.1....... \...
_
..
/I. ?
,:~:.
~
..
. ..
.... ..
"'~" ~""
~:':;.,., ~
? ?
. ' . tI? ?:_
~.
~
' .,.
.
.,
.,'
rl
? f#.
20
.."~'~:"""''''~'
. ~"
,~, ' 'Ie" ,:":
".!'J" ? , ..
?'
?'
- 10
]0
??
:' :,~'
":~.:~
MAP E.tilllllta Modal
: I.
0
10
20
]0
-]0
-20
-10
0
10
20
]0
Figure 2: Example of model estimation for simulated data.
4.2
Vowel data
The reversible-jump MCMC chain of model order for the Peterson-Barney vowel
data example is shown in figure 3, alongside the resulting MAP model estimate.
For ease of visualisation, the estimated model and data samples have been projected onto the first two principal components of the data. Again, the results are
encouraging.
5
Conclusion
One of the key problems when using Gaussian mixture models is estimation of the
optimum number of components to include in the model. In this paper we extend
the reversible-jump MCMC technique for estimating the parameters of Gaussian
mixtures with an unknown number of components to the multivariate spherical
Gaussian case. The technique is then demonstrated on a simulated data example
and an example using a well known dataset.
The attraction of this approach is that the number of mixture components is not
fixed at the outset but becomes a parameter of the model. The reversible-jump
MCMC approach is then capable of moving between parameter subspaces which
Reversible-Jwnp MCMC for Multivariate Spherical Gaussian Mixtures
583
.
Figure 3: Reversible-jump MCMC chain of model order and MAP estimate of model
(projected onto first two principal components) for vowel data.
correspond to models with different numbers of mixture components. As a result
a sample of the full joint distribution is generated from which the posterior distribution for the number of model components can be derived. This information may
then either be used to construct a Bayesian classifier or to define the centres in a
radial basis function networ k.
References
[1] W.R. Gilks, S. Richardson, and D.J. Spiegelhalter Eds. Markov Chain Monte
Carlo in Practice. Chapman and Hall, 1995.
[2] P.J. Green. Reversible jump MCMC computation and Bayesian model determination. Boimetrika, 82:711-732, 1995.
[3] D.B. Phillips and A.F.M. Smith. Bayesian model comparison via jump diffusions. In W.R. Gilks, S. Richardson, and D.J. Spiegelhalter, editors, Markov
Chain Monte Carlo in Practice. Chapman and Hall, 1995.
[4] S. Richardson and P. J. Green. On Bayesian analysis of mixtures with an unknown number of components. J. Royal Stat. Soc. Series B, 59(4), 1997.
[5] D.M. Titterington, A.F.M. Smith, and U.E. Makov. Statistical Analysis of Finite
Mixture Distributions. Wiley, 1985.
?British Crown Copyright 1998 /DERA.
Published with the permission of the controller of Her Britannic Majesty's
Stationary Office.
| 1422 |@word determinant:1 version:1 proportion:1 barney:3 initial:1 series:1 current:2 comparing:2 z2:6 yet:1 must:1 conforming:1 subsequent:1 treating:1 stationary:1 ith:1 smith:2 firstly:1 along:2 beta:2 ik:2 combine:8 introduce:1 frequently:1 spherical:9 encouraging:2 becomes:2 estimating:1 what:1 titterington:1 transformation:5 nj:1 ti:2 xd:1 qm:3 classifier:1 uk:1 zl:5 appear:2 generalised:1 treat:1 might:1 ease:1 limited:1 range:1 gilks:2 testing:2 practice:2 thought:1 outset:1 radial:2 onto:2 selection:1 deterministic:1 demonstrated:2 map:4 yt:1 independently:1 splitting:2 attraction:1 importantly:1 hierarchy:1 target:1 neighbouring:1 designing:1 hypothesis:2 jlj:1 updating:4 xi9:3 role:1 calculate:1 agency:1 ui:2 networ:1 upon:2 basis:4 joint:5 monte:4 artificial:1 hyper:2 choosing:5 formation:1 drawing:1 statistic:1 formant:1 richardson:3 final:2 jij:1 combining:2 mixing:1 regularity:1 p:1 optimum:2 generating:1 leave:1 derive:2 stat:1 measured:1 ij:5 soc:1 implemented:1 involves:1 met:1 aperiodic:1 correct:1 pea:1 enable:1 really:1 mathematically:1 strictly:1 hold:1 mm:1 hall:2 normal:3 exp:2 u3:4 purpose:1 estimation:7 label:2 gaussian:14 defence:1 rather:1 office:1 derived:2 majesty:1 likelihood:2 am:8 posteriori:1 prescribing:1 hidden:1 visualisation:1 her:1 comprising:1 issue:2 denoted:1 k6:1 field:2 once:1 f3:1 having:1 construct:1 sampling:2 chapman:2 represents:1 lit:2 simplify:1 irreducible:1 randomly:1 neighbour:1 gamma:1 vowel:6 freedom:1 acceptance:2 evaluation:1 mixture:31 nl:2 copyright:1 tj:3 chain:8 allocating:1 capable:2 necessary:1 iv:1 desired:2 wr14:1 density:4 ie:3 invertible:1 itj:1 ijp:1 again:1 choose:4 dra:1 leading:1 li:1 toy:1 makov:1 tra:1 satisfy:1 ump:1 bayes:1 complicated:1 ass:1 variance:2 conducting:1 equate:1 correspond:2 ofthe:1 bayesian:7 rejecting:1 carlo:4 straight:1 published:1 ed:1 c7:1 frequency:1 dataset:2 ut:1 amplitude:1 hmg:1 higher:1 follow:1 modal:2 until:1 hastings:2 reversible:19 aj:1 assigned:1 width:1 noted:1 oc:1 criterion:2 pck:1 crown:1 recently:1 common:1 rl:1 jl:1 extend:2 imposing:1 gibbs:1 phillips:1 centre:3 moving:2 specification:1 longer:1 gt:1 add:1 multivariate:10 posterior:1 seen:1 care:1 signal:2 ii:4 full:6 multiple:1 rj:3 alan:1 determination:1 dept:1 controller:1 histogram:3 represent:3 iteration:1 proposal:2 addition:2 lzl:1 allocated:1 extra:1 regularly:1 iii:1 split:7 zi:6 tm:3 expression:2 gb:1 detailed:1 generate:1 estimated:3 arising:2 shall:1 iz:1 key:3 independency:1 four:1 drawn:2 diffusion:2 extends:1 recognise:1 decision:2 layer:1 followed:1 yielded:1 occur:1 placement:1 mtn:1 x2:1 min:2 relatively:4 combination:2 conjugate:1 em:1 metropolis:2 making:1 taken:2 remains:1 apply:1 hierarchical:3 frequentist:1 permission:1 dera:1 original:1 denotes:1 dirichlet:1 include:1 giving:1 uj:1 move:21 question:2 marrs:5 diagonal:2 subspace:2 distance:1 simulated:6 ratio:4 balance:1 negative:1 unknown:3 observation:3 markov:4 finite:1 introduced:1 pair:3 required:1 extensive:1 alongside:2 royal:1 green:2 representing:1 scheme:1 spiegelhalter:2 carried:1 utterance:1 prior:12 literature:1 allocation:4 degree:1 editor:1 pile:1 kia:1 gl:1 last:1 zc:2 peterson:3 fifth:1 dimension:1 calculated:1 forward:1 made:1 jump:17 projected:2 simplified:1 ignore:1 xi:2 search:1 continuous:1 expectationmaximisation:1 nature:1 itt:1 obtaining:1 main:1 n2:2 malvern:1 wiley:1 precision:1 wish:1 xl:1 lie:1 jacobian:2 british:1 rlk:3 bivariate:1 intractable:1 consist:1 nk:1 univariate:1 desire:1 g2:3 u2:3 nested:2 conditional:4 labelled:2 averaging:1 principal:2 accepted:1 evaluate:1 mcmc:22 tested:1 |
465 | 1,423 | Gradients for retinotectal mapping
Geoffrey J. Goodhill
Georgetown Institute for Cognitive and Computational Sciences
Georgetown University Medical Center
3970 Reservoir Road
Washington IX: 20007
geoff@giccs.georgetown.edu
Abstract
The initial activity-independent formation of a topographic map
in the retinotectal system has long been thought to rely on the
matching of molecular cues expressed in gradients in the retina
and the tectum. However, direct experimental evidence for the
existence of such gradients has only emerged since 1995. The new
data has provoked the discussion of a new set of models in the experimentalliterature. Here, the capabilities of these models are analyzed, and the gradient shapes they predict in vivo are derived.
1 Introduction
During the early development of the visual system in for instance rats, fish and
chickens, retinal axons grow across the surface of the optic tectum and establish
connections so as to form an ordered map. Although later neural activity refines
the map, it is not required to set up the initial topography (for reviews see Udin
& Fawcett (1988); Goodhill (1992?. A long-standing idea is that the initial topography is formed by matching gradients of receptor expression in the retina with
gradients of ligand expression in the tectum (Sperry, 1963). Particular versions of
this idea have been formalized in theoretical models such as those of Prestige &
Willshaw (1975), Willshaw & von der Malsburg (1979), Whitelaw & Cowan (1981),
and Gierer (1983;1987). However, these models were developed in the absence
of any direct experimental evidence for the existence of the necessary gradients.
Since 1995, major breakthroughs have occurred in this regard in the experimental
literature. These center around the Eph (Erythropoetin-producing hepatocellular)
subfamily of receptor tyrosine kinases. Eph receptors and their ligands have been
shown to be expressed in gradients in the developing retina and tectum respectively, and to playa role in guiding axons to appropriate positions. These exciting
new developments have led experimentalists to discuss theoretical models differ-
Gradients/or Retinotectal Mapping
153
ent from those previously proposed (e.g. Tessier-Lavigne (1995); Tessier-Lavigne
& Goodman (1996); Nakamoto et aI, (1996)). However, the mathematical consequences of these new models, for instance the precise gradient shapes they require,
have not been analyzed. In this paper, it is shown that only certain combinations
of gradients produce appropriate maps in these models, and that the validity of
these models is therefore experimentally testable.
2 Recent experimental data
Receptor tyrosine kinases are a diverse class of membrane-spanning proteins. The
Eph subfamily is the largest, with over a dozen members. Since 1990, many of the
genes encoding Eph receptors and their ligands have been shown to be expressed
in the developing brain (reviewed in Friedman & O'Leary, 1996). Ephrins, the
ligands for Eph receptors, are all membrane anchored. This is unlike the majority
of receptor tyrosine kinase ligands, which are usually soluble. The ephrins can be
separated into two distinct groups A and B, based on the type of membrane anchor.
These two groups bind to distinct sets of Eph receptors, which are thus also called
A and B, though receptor-ligand interaction is promiscuous within each subgroup.
Since many research groups discovered members of the Eph family independently,
each member originally had several names. However a new standardized notation
was recently introduced (Eph Nomenclature Committee, 1997), which is used in
this paper.
With regard to the mapping from the nasal-temporal axis of the retina to the
anterior-posterior axis of the tectum (figure 1), recent studies have shown the following (see Friedman & O'Leary (1996) and Tessier-Lavigne & Goodman (1996)
for reviews).
? EphA3 is expressed in an increasing nasal to temporal gradient in the
retina (Cheng et aI, 1995).
? EphA4 is expressed uniformly in the retina (Holash & Pasquale, 1995).
? Ephrin-A2, a ligand of both EphA3 and EphA4, is expressed in an increasing rostral to caudal gradient in the tectum (Cheng et aI, 1995).
? Ephrin-A5, another ligand of EphA3 and EphA4, is also expressed in an
increasing rostral to caudal gradient in the tectum, but at very low levels
in the rostral half of the tectum (Drescher et aI, 1995).
All of these interactions are repulsive. With regard to mapping along the complementary dimensions, EphB2 is expressed in a high ventral to low dorsal gradient
in the retina, while its ligand ephrin-B1 is expressed in a high dorsal to low ventral
gradient in the tectum (Braisted et aI, 1997). Members of the Eph family are also
beginning to be implicated in the formation of topographic projections between
many other pairs of structures in the brain (Renping Zhou, personal communication). For instance, EphA5 has been found in an increasing lateral to medial gradient in the hippocampus, and ephrin-A2 in an increasing dorsal to ventral gradient
in the septum, consistent with a role in establishing the topography of the map
between hippocampus and septum (Gao et aI, 1996).
The current paper focusses just on the paradigm case of the nasal-temporal to
anterior-posterior axis of the retinotectal mapping. Actual gradient shapes in this
system have not yet been quantified. The analysis below will assume that certain
gradients are linear, and derive the consequences for the other gradients.
G. J. Goodhill
154
TECTUM
RETINA
c
N
L(y)
R(x)
k:==::y
"'-----.......;;;~x
Figure 1: This shows the mapping that is normally set up from the retina to the
tectum. Distance along the nasal-temporal axis of the retina is referred to as x and
receptor concentration as R( x). Distance along the rostral-caudal axis of the tectum
is referred to as y and ligand concentration as L(y).
3
Mathematical models
Let R be the concentration of a receptor expressed on a growth cone or axon, and
L the concentration of a ligand present in the tectum. Refer to position along the
nasal-temporal axis of the retina as x, and position along the rostral-caudal axis of
the tectum as y, so that R = R(x) and L = L(y) (see figure 1). Gierer (1983; 1987)
discusses how topographic information could be signaled by interactions between
ligands and receptors. A particular type of interaction, proposed by Nakamoto et
al (1996), is that the concentration of a "topographic signal", the signal that tells
axons where to stop, is related to the concentration of receptor and ligand by the
law of mass action:
G(x, y) = kR(x)L(y)
(1)
where G(x, y) is the concentration of topographic signal produced within an axon
originating from position x in the retina when it is at position y in the tectum,
and k is a constant. In the general case of multiple receptors and ligands, with
promiscuous interactions between them, this equation becomes
G(x, y) =
L: kijRi(X)Lj(Y)
(2)
i,j
Whether each receptor-ligand interaction is attractive or repulsive is taken care of
by the sign of the relevant k ij ?
Two possibilities for how G(x, y) might produce a stop (or branch) signal in the
growth cone (or axon) are that this occurs when (1) a "set point" is reached (discussed in, for example, Tessier-Lavigne & Goodman (1996); Nakamoto et al (1996?
,i.e. G (x, y) = c where c is a constant, or (2) attraction (or repulsion) reaches a local
maximum (or minimum), i.e. &G~~,y) = 0 (Gierer, 1983; 1987). For a smooth, uni-
Gradients for Retinotectal Mapping
155
form mapping, one of these conditions must hold along a line y ex: x. For simplicity
assume the constant of proportionality is unity.
3.1
Set point rule
For one gradient in the retina and one gradient in the tectum (i.e. equation 1), this
requires that the ligand gradient be inversely proportional to the receptor gradient:
c
L(x) = R(x)
If R(x) is linear (c.f. the gradient of EphA3 in the retina), the ligand concentration
is required to go to infinity at one end of the tectum (see figure 2). One way round
this is to assume R(x) does not go to zero at x = 0: the experimental data is not
precise enough to decide on this point. However, the addition of a second receptor
gradient gives
c
L(x) = k1R1 (x) + k2R2(X)
If R1 (x) is linear and R 2(x) is flat (c.f. the gradient of EphA4 in the retina), then
L (y) is no longer required to go to infinity (see figure 2). For two receptor and two
ligand gradients many combinations of gradient shapes are possible. As a special
case, consider R1 (x) linear, R 2(x) flat, and L 1(y) linear (c.f. the gradient of Elfl in
the tectum). Then L2 is required to have the shape
L ( ) = ay2
2 Y
dy
+ by
+e
where a, b, d, e are constants. This shape depends on the values of the constants,
which depend on the relative strengths of binding between the different receptor
and ligand combinations. An interesting case is where R1 binds only to L1 and R2
binds only .to L 2 , i.e. there is no promiscuity. In this case we have
L 2(y) ex: y2
(see figure 2). This function somewhat resembles the shape of the gradient that
has been reported for ephrin-AS in the tectum. However, this model requires one
gradient to be attractive, whereas both are repulsive.
3.2
Local optimum rule
For one retinal and one tectal gradient we have the requirement
R(x) aL(y) = 0
ay
This is not generally true along the line y = x, therefore there is no map. The same
problem arises with two receptor gradients, whatever their shapes. For two receptor and two ligand gradients many combinations of gradient shapes are possible.
(Gierer (1983; 1987) investigated this case, but for a more complicated reaction law
for generating the topographic signal than mass action.) For the special case introduced above, L 2 (y) is required to have the shape
L2(y) = ay + blog(dy
+ e) + f
where a, b, d, e, and f are constants as before. Considering the case of no promiscuity, we again obtain
L 2(y) ex: y2
i.e. the same shape for L2 (y) as that specified by the set point rule.
G. 1. Goodhill
156
A
L
B
L
c
Figure 2: Three combinations of gradient shapes that are sufficient to produce a
smooth mapping with the mass action rule. In the left column the horizontal axis
is position in the retina while the vertical axis is the concentration of receptor. In
the right column the horizontal axis is position in the tectum while the vertical axis
is the concentration of ligand. Models A and B work with the set point but not the
local optimum rule, while model C works with both rules. For models B and C,
one gradient is negative and the other positive.
Gradients for Retinotectal Mapping
157
4 Discussion
For both rules, there is a set of gradient shapes for the mass-action model that is
consistent with the experimental data, except for the fact that they require one gradient in the tectum to be attractive. Both ephrin-A2 and ephrin-A5 have repulsive
effects on their receptors expressed in the retina, which is clearly a problem for
these models. The local optimum rule is more restrictive than the set point rule,
since it requires at least two ligand gradients in the tectum. However, unlike the set
point rule, it supplies directional information (in terms of an appropriate gradient
for the topographic signal) when the axon is not at the optimal location.
In conclusion, models based on the mass action assumption in conjunction with either a "set point" or "local optimum" rule can be true only if the relevant gradients
satisfy the quantitative relationships described above. A different theoretical approach, which analyzes gradients in terms of their ability to guide axons over the
maximum possible distance, also makes predictions about gradient shapes in the
retinotectal system (Goodhill & Baier, 1998). Advances in experimental technique
should enable a more quantitative analysis of the gradients in situ to be performed
shortly, allowing these predictions to be tested. In addition, analysis of particular
Eph and ephrin knockout mice (for instance ephrin-A5 (Yates et aI, 1997? is now
being performed, which should shed light on the role of these gradients in normal
map development.
Bibliography
Braisted, J.E., McLaughlin, T., Wang, H.U., Friedman, G.C, Anderson, D.J. &
O'Leary, D.D.M. (1997). Graded and lamina-specific distributions of ligands
of EphB receptor tyrosine kinases in the developing retinotectal system. Developmental Biology, 19114-28.
Cheng, H.J., Nakamoto, M., Bergemann, A.D & Flanagan, J.G. (1995). Complementary gradients in expression and binding of Elf-1 and Mek4 in development of
the topographic retinotectal projection map. Cell, 82,371-381.
Drescher, U., Kremoser, C, Handwerker, C, Loschinger, J., Noda, M. & Bonhoeffer,
F. (1995). In-vitro guidance of retinal ganglion-cell axons by RAGS, a 25 KDa
tectal protein related to ligands for Eph receptor tyrosine kinases. Cell, 82, 359370.
Eph Nomenclature Committee (1997). Unified nomenclature for Eph family receptors and their ligands, the ephrins. Cell, 90, 403-404.
Friedman, G.C & O'Leary, D.D.M. (1996). Eph receptor tyrosine kinases and their
ligands in neural development. Curro Opin. Neurobiol., 6, 127-133.
Gierer, A. (1983). Model for the retinotectal projection. Proc. Roy. Soc. Lond. B, 218,
77-93.
Gierer, A. (1987). Directional cues for growing axons forming the retinotectal projection. Development, 101,479-489.
Gao, P.-P., Zhang, J.-H., Yokoyama, M., Racey, R, Dreyfus, CF., Black, LR & Zhou,
R. (1996). Regulation of topographic projection in the brain: Elf-1 in the hippocampalseptal system. Proc. Nat. Acad. Sci. USA, 93, 11161-11166.
Goodhill, G.J. (1992). Correlations, Competition and Optimality: Modelling
the Development of Topography and Ocular Dominance. Cognitive Science Research Paper CSRP 226, University of Sussex. Available from
www.giccs.georgetown.edu/ "'geoff
158
G. 1. Goodhill
Goodhill, G.J. & Baier, H. (1998). Axon guidance: stretching gradients to the limit.
Neural Computation, in press.
Holash, J.A. & Pasquale, E.B. (1995). Polarized expression of the receptor proteintyrosine kinase CekS in the developing avian visual system. Developmental Biology, 172, 683-693.
Nakamoto, M., Cheng H.J., Friedman, G.C, Mclaughlin, T., Hansen, M.J., Yoon,
CH., O'Leary, D.D.M. & Flanagan, J.G. (1996). Topographically specific effects
of ELF-Ion retinal axon guidance in-vitro and retinal axon mapping in-vivo.
Cell, 86, 755-766.
Prestige, M.C & Willshaw, D.J. (1975). On a role for competition in the formation
of patterned neural connexions. Proc. R. Soc. Lond. B, 190, 77-98.
Sperry, RW. (1963). Chemoaffinity in the orderly growth of nerve fiber patterns
and connections. Proc. Nat. Acad. Sci., U.S.A., 50, 703-710.
Tessier-Lavigne, M. (1995). Eph receptor tyrosine kinases, axon repulsion, and the
development of topographic maps. Cell, 82, 345-348.
Tessier-Lavigne, M. and Goodman, CS. (1996). The molecular biology of axon
guidance. Science, 274, 1123-1133.
Udin, S.B. & Fawcett, J.W. (1988). Formation of topographic maps. Ann. Rev. Neurosci., 11,289-327.
Whitelaw, V.A. & Cowan, J.D. (1981). Specificity and plasticity of retinotectal connections: a computational model. Jou. Neurosci., 1, 1369-1387.
Willshaw, D.J. & Malsburg, C von der (1979). A marker induction mechanism for
the establishment of ordered neural mappings: its application to the retinotectal problem. Phil. Trans. Roy. Soc. B, 287, 203-243.
Yates, P.A., McLaughlin, T., Friedman, G.C, Frisen, J., Barbacid, M. & O'Leary,
D.D.M. (1997). Retinal axon guidance defects in mice lacking ephrin-A5 (ALl/RAGS). Soc. Neurosci. Abstracts, 23, 324.
| 1423 |@word version:1 hippocampus:2 proportionality:1 initial:3 reaction:1 current:1 anterior:2 yet:1 must:1 refines:1 plasticity:1 shape:14 opin:1 medial:1 cue:2 half:1 beginning:1 lr:1 location:1 zhang:1 mathematical:2 along:7 direct:2 supply:1 rostral:5 growing:1 brain:3 actual:1 considering:1 increasing:5 becomes:1 notation:1 mass:5 neurobiol:1 developed:1 unified:1 temporal:5 quantitative:2 growth:3 shed:1 willshaw:4 whatever:1 normally:1 medical:1 producing:1 knockout:1 before:1 positive:1 bind:3 local:5 limit:1 consequence:2 acad:2 receptor:29 encoding:1 establishing:1 might:1 black:1 resembles:1 quantified:1 patterned:1 flanagan:2 thought:1 matching:2 projection:5 road:1 specificity:1 protein:2 www:1 map:10 center:2 phil:1 go:3 independently:1 formalized:1 simplicity:1 rule:11 attraction:1 tectum:22 roy:2 giccs:2 role:4 yoon:1 wang:1 developmental:2 personal:1 tyrosine:7 depend:1 topographically:1 geoff:2 fiber:1 separated:1 distinct:2 tell:1 formation:4 emerged:1 ability:1 topographic:11 nakamoto:5 interaction:6 relevant:2 promiscuous:2 tessier:6 competition:2 ent:1 optimum:4 r1:3 requirement:1 produce:3 generating:1 lamina:1 derive:1 avian:1 ij:1 soc:4 c:1 differ:1 enable:1 require:2 promiscuity:2 hold:1 around:1 normal:1 mapping:12 predict:1 major:1 ventral:3 early:1 a2:3 proc:4 hansen:1 largest:1 caudal:4 clearly:1 establishment:1 zhou:2 conjunction:1 derived:1 focus:1 modelling:1 repulsion:2 lj:1 originating:1 development:8 breakthrough:1 special:2 washington:1 biology:3 elf:3 retina:17 friedman:6 a5:4 possibility:1 situ:1 analyzed:2 light:1 csrp:1 necessary:1 signaled:1 guidance:5 theoretical:3 instance:4 column:2 ay2:1 udin:2 reported:1 standing:1 mouse:2 leary:6 von:2 again:1 soluble:1 provoked:1 cognitive:2 retinal:6 satisfy:1 depends:1 later:1 performed:2 reached:1 capability:1 complicated:1 vivo:2 formed:1 stretching:1 directional:2 drescher:2 produced:1 reach:1 ocular:1 stop:2 nerve:1 originally:1 prestige:2 though:1 anderson:1 just:1 correlation:1 horizontal:2 marker:1 usa:1 effect:2 validity:1 name:1 y2:2 true:2 attractive:3 round:1 during:1 sussex:1 rat:1 ay:2 l1:1 dreyfus:1 recently:1 vitro:2 discussed:1 occurred:1 refer:1 ai:7 had:1 longer:1 surface:1 playa:1 pasquale:2 posterior:2 recent:2 certain:2 blog:1 der:2 minimum:1 analyzes:1 care:1 somewhat:1 paradigm:1 signal:6 tectal:2 branch:1 multiple:1 smooth:2 long:2 molecular:2 prediction:2 experimentalists:1 fawcett:2 chicken:1 cell:6 ion:1 addition:2 whereas:1 grow:1 goodman:4 unlike:2 cowan:2 member:4 enough:1 connexion:1 idea:2 mclaughlin:3 whether:1 expression:4 septum:2 nomenclature:3 action:5 generally:1 nasal:5 rw:1 fish:1 sign:1 diverse:1 jou:1 yates:2 group:3 dominance:1 defect:1 cone:2 sperry:2 bergemann:1 family:3 decide:1 polarized:1 dy:2 curro:1 cheng:4 activity:2 strength:1 optic:1 infinity:2 flat:2 bibliography:1 lond:2 optimality:1 developing:4 eph:15 combination:5 membrane:3 across:1 unity:1 rev:1 taken:1 equation:2 previously:1 discus:2 committee:2 mechanism:1 kda:1 end:1 repulsive:4 available:1 appropriate:3 shortly:1 existence:2 standardized:1 cf:1 lavigne:6 malsburg:2 restrictive:1 testable:1 establish:1 graded:1 occurs:1 concentration:10 gradient:52 distance:3 lateral:1 sci:2 majority:1 baier:2 spanning:1 induction:1 relationship:1 regulation:1 negative:1 kinase:8 allowing:1 vertical:2 noda:1 communication:1 precise:2 discovered:1 introduced:2 pair:1 required:5 specified:1 connection:3 rag:2 subgroup:1 trans:1 usually:1 goodhill:8 below:1 pattern:1 rely:1 inversely:1 axis:11 review:2 literature:1 l2:3 georgetown:4 relative:1 law:2 lacking:1 topography:4 interesting:1 proportional:1 geoffrey:1 sufficient:1 consistent:2 exciting:1 implicated:1 guide:1 institute:1 regard:3 dimension:1 uni:1 gene:1 orderly:1 anchor:1 b1:1 anchored:1 reviewed:1 retinotectal:13 investigated:1 neurosci:3 complementary:2 reservoir:1 referred:2 axon:16 position:7 guiding:1 ix:1 dozen:1 specific:2 r2:1 evidence:2 kr:1 nat:2 bonhoeffer:1 led:1 subfamily:2 gao:2 ganglion:1 visual:2 forming:1 expressed:11 ordered:2 binding:2 ligand:26 ch:1 ann:1 whitelaw:2 absence:1 experimentally:1 except:1 uniformly:1 called:1 experimental:7 arises:1 dorsal:3 tested:1 ex:3 |
466 | 1,424 | Learning nonlinear overcomplete
representations for efficient coding
Michael S. Lewicki
Terrence J. Sejnowski
lewicki~salk.edu
terry~salk.edu
Howard Hughes Medical Institute
Computational Neurobiology Lab
The Salk Institute
10010 N. Torrey Pines Rd.
La Jolla, CA 92037
Abstract
We derive a learning algorithm for inferring an overcomplete basis
by viewing it as probabilistic model of the observed data. Overcomplete bases allow for better approximation of the underlying
statistical density. Using a Laplacian prior on the basis coefficients
removes redundancy and leads to representations that are sparse
and are a nonlinear function of the data. This can be viewed as
a generalization of the technique of independent component analysis and provides a method for blind source separation of fewer
mixtures than sources. We demonstrate the utility of overcomplete representations on natural speech and show that compared
to the traditional Fourier basis the inferred representations potentially have much greater coding efficiency.
A traditional way to represent real-values signals is with Fourier or wavelet bases.
A disadvantage of these bases, however, is that they are not specialized for any
particular dataset. Principal component analysis (PCA) provides one means for
finding an basis that is adapted for a dataset, but the basis vectors are restricted
to be orthogonal. An extension of PCA called independent component analysis
(Jutten and Herault, 1991; Comon et al., 1991; Bell and Sejnowski, 1995) allows
the learning of non-orthogonal bases. All of these bases are complete in the sense
that they span the input space, but they are limited in terms of how well they can
approximate the dataset's statistical density.
Representations that are overcomplete, i. e. more basis vectors than input variables,
can provide a better representation, because the basis vectors can be specialized for
Learning Nonlinear Overcomplete Representations for Efficient Coding
557
a larger variety of features present in the entire ensemble of data. A criticism of
overcomplete representations is that they are redundant, i.e. a given data point may
have many possible representations, but this redundancy is removed by the prior
probability of the basis coefficients which specifies the probability of the alternative
representations.
Most of the overcomplete bases used in the literature are fixed in the sense that
they are not adapted to the structure in the data. Recently Olshausen and Field
(1996) presented an algorithm that allows an overcomplete basis to be learned. This
algorithm relied on an approximation to the desired probabilistic objective that had
several drawbacks, including tendency to breakdown in the case of low noise levels
and when learning bases with higher degrees of overcompleteness. In this paper, we
present an improved approximation to the desired probabilistic objective and show
that this leads to a simple and robust algorithm for learning optimal overcomplete
bases.
1
Inferring the representation
The data,
X 1 :L
'
are modeled with an overcomplete linear basis plus additive noise:
x = AS+i
(1)
where A is an L x M matrix, whose columns are the basis vectors, where M ~ L .
We assume Gaussian additive noise so that 10gP(xIA, s) ()( -A(X - As)2/2, where
A = 1/(12 defines the precision of the noise.
The redundancy in the overcomplete representation is removed by defining a density
for the basis coefficients, P(s), which specifies the probability of the alternative
representations. The most probable representation, 5, is found by maximizing the
posterior distribution
s = maxP(sIA,x) =
8
maxP(s)P(xIA,s)
(2)
8
P(s) influences how the data are fit in the presence of noise and determines the
uniqueness of the representation. In this model, the data is a linear function of
s, but s is not, in general, a linear function of the data. IT the basis function is
complete (A is invertible) then, assuming broad priors and low noise, the most
probable internal state can be computed simply by inverting A. In the case of an
overcomplete basis, however, A can not be inverted. Figure 1 shows how different
priors induce different representations.
Unlike the Gaussian prior, the optimal representation under the Laplacian prior
cannot be obtained by a simple linear operation. One approach for optimizing sis
to use the gradient of the log posterior in an optimization algorithm. An alternative
method for finding the most probable internal state is to view the problem as the
linear program: min 1 T s such that As = x. This can be generalized to handle both
positive and negative s and solved efficiently and exactly with interior point linear
programming methods (Chen et al., 1996) .
M. S. Lewicki and T. 1. Sejnowski
558
a
b
if
L2
G)
:::l
~ 0.1
10'05~\
L1
8
\"
o
~-
20
-~ -
-
~-
- 80
100
120
Figure 1: Different priors induce different representations. (a) The 2D data distribution
has three main axes which form an overcomplete representation. The graphs marked "L2"
and "L1" show the optimal scaled basis vectors for the data point x under the Gaussian
and Laplacian prior, respectively. Assuming zero noise, a Gaussian for P{s) is equivalent
to finding the exact fitting s with minimum L2 norm, which is given by the pseudoinverse
s = A+x. A Laplacian prior (P{Sm) ex: exp[-OlsmlJ) yields the exact fit with minimum L1
norm, which is a nonlinear operation which essentially selects a subset of the basis vectors
to represent the data (Chen et al., 1996). The resulting representation is sparse. (b) A
64-sample segment of speech was fit to a 2x overcomplete Fourier representation (128 basis
vectors). The plot shows rank order distribution of the coefficients of s under a Gaussian
prior (dashed); and a Laplacian prior (solid). Far more significantly positive coefficients
are required under the Gaussian prior than under the Laplacian prior.
2
Learning
The learning objective is to adapt A to maximize the probability of the data which
is computed by marginalizing over the internal states
P(xIA) =
J
(3)
ds P(s)P(xIA, s)
general, this integral cannot be evaluated analytically but can be approximated
with a Gaussian integral around s, yielding
log P(xIA)
~
const.
+ log pes) - ~ (x -
As)2 -
~ log det H
(4)
where H is the Hessian of the log posterior at S, given by )'ATA - VVlogP(s).
To avoid a singularity under the Laplacian prior, we use the approximation
(logP(sm?)' ~ -8tanh(,8sm) which gives the Hessian full rank and positive determinant. For large ,8 this approximates the true Laplacian prior. A learning rule
can be obtained by differentiating log P(xIA) with respect to A.
In the following discussion, we will present the derivations of the three terms in (4)
and simplifying assumptions that lead to the following simple form of the learning
rule
(5)
Learning Nonlinear Ollercomplete Representations for Efficient Cod(ng
2.1
559
Deriving V log pes)
This term specifies how to change A so as to make the probability of the representation s more probable. IT we assume a Laplacian prior, this component changes A
to make the representation more sparse.
pes)
= rIm P(Sm).
In order to obtain 8sm/8aij, we need to describe
If the basis is complete (and we assume low noise), then we
may simply invert A to obtain s = A -IX. When A is overcomplete, however, there
We assume
s as a function of A.
is no simple expression, but we may still make an approximation.
Under priors, the most probable solution, s, will yield at most L non-zero elements.
In effect, this selects a complete basis from A. Let A represent this reduced basis
under s. We then have s = A -1(X- ?) where s is equal to s with M - L zero-valued
elements removed. A-I obtained by removing the columns of A corresponding to
the M - L zero-valued elements of s. This allows us to use results obtained for the
case when A is invertible. Following MacKay (1996) we obtain
(6)
Rewriting in matrix notation we have
A~ -Tv~T
zs
810gP(s) _
8A
- -
(7)
We can use this to obtain an expression in terms of the original variables. We
simply invert the mapping s ~ s to obtain Z f- z and W T f- A-T (row-wise) with
Zm = 0 and row m ofWT = 0 if 8m = O. We then have
8 log P(s)
8A
2.2
=-
WT
zs
T
(8)
Deriving Vex - As)2
The second term specifies how to change A so as to minimize the data misfit.
Letting ek = [x - AS]k and using the results and notation from above we have:
8
A~ 2
-8
.. "2 L-ek
a"
k
~
~
k
I
8s,
alJ
(9)
-aklWliSj
(10)
= AeiSj + AL-ek L-ak'~
= AeiSj + ALek L
k
= AeiSj - AeiSj
I
=0
(11)
Thus no gradient component arises from the error term.
2.3
Deriving V log det H
The third term in the learning rule specifies how to change the weights so as to
minimize the width of the posterior distribution P(xIA) and thus increase the
overall probability of the data. An element of H is defined by Hmn = Cmn + bmn
M. S. Lewicki and T. J. Sejnowski
560
where Cmn = Ek Aakmakn and bmn = [- V'V' log P(s)]mn. This gives
8logdetH _ ""H- 1 [8e mn
- ~
8a ~??
mn nm 8a U..
+ 8bmn ]
8a??
~
(12)
First considering 8Cmn/8aij, we can obtain
L H~~ ~~~ = L
mn
~3
Using the fact that H~;
m:f.j
H~;.\aim + L Hj~.\aim + Hj/ 2Aaij
(13)
m:f.j
= Hj~ due to the symmetry of the Hessian, we have
(14)
Next we derive 8bmn/8aij. We have that V'V'logP(s) is diagonal, because we
assume pes) =
P(sm). Letting 2Ym = H~!n8bmm/8sm and using the result
under the reduced representation (6) we can obtain
nm
(15)
2.4
Stabilizing and simplifying the learning rule
Putting the terms together yields a problematic expression due to the matrix inverses. This can be alleviated by multiplying the gradient by an appropriate positive
definite matrix, which rescales the gradient components but preserves a direction
valid for optimization. Noting that ATW T = I we have
(16)
H'\ is large (low noise) then the Hessian is dominated by AATA and we have
(17)
The vector y hides a computation involving the inverse Hessian. IT the basis vectors
in A are randomly distributed, then as the dimensionality of A increases the basis
vectors become approximately orthogonal and consequently the Hessian becomes
approximately diagonal. It can be shown that if log pes) and its derivatives are
smooth, Ym vanishes for large A. Combining the remaining terms yields equation
(5). Note that this rule contains no matrix inverses and the vector z involves only
the derivative of the log prior.
In the case where A is square, this form of the rule is similar to the natural gradient
independent component analysis (ICA) learning rule (Amari et al., 1996). The
difference in the more general case where A is rectangular is that s must maximize
the posterior distribution P(slx, A) which cannot be done simply with the filter
matrix as in standard ICA algorithms.
Learning Nonlinear Overcomplete Representations/or Efficient Coding
3
561
Examples
More sources than inputs. In these 2D examples, the bases were initialized
to random, normalized vectors. The coefficients were solved using BPMPD and
publicly available interior point linear programming package (Meszaros, 1997) which
gives the most probable solution under the Laplacian prior assuming zero noise.
The algorithm was run for 30 iterations using equation (5) with a stepsize of 0.001
and a batchsize of 200. Convergence was rapid, typically requiring less than 20
iterations. In all cases, the direction of the learned vectors matched those of the
true generating distribution; the magnitude was estimated less precisely, possibly
due to the approximation oflogP(xIA). This can be viewed as a source separation
problem, but true separation will be limited due to the projection of the sources
down to a smaller subspace which necessarily loses information .
. ' '.
~
Figure 2: Examples illustrating the fitting of 2D distributions with overcomplete bases.
The first example is equivalent to 3 sources mixed into 2 channels; the second to 4 sources
mixed into 2 channels. The data in both examples were generated from the true basis A
using x = As with the elements of s distributed according to an exponential distribution
with unit mean. Identical results were obtained by drawing s from a Laplacian prior
(positive and negative coefficients). The overcomplete bases allow the model to capture
the true underlying statistical structure in the 2D data space.
Overcomplete representations of speech. Speech data were obtained from the
TIMIT database, using a single speaker was speaking ten different example sentences with no preprocessing. The basis was initialized to an overcomplete Fourier
basis. A conjugate gradient routine was used to obtain the most probable basis
coefficients. The stepsize was gradually reduced over 10000 iterations. Figure 3
shows that the learned basis is quite different from the Fourier representation. The
power spectrum for the learned basis vectors can be multimodal and/or broadband.
The learned basis achieves greater coding efficiency: 2.19 ? 0.59 bits per sample
compared to 3.86 ? 0.28 bits per sample for a 2x overcomplete Fourier basis.
4
Summary
Learning overcomplete representations allows a basis to better approximate the underlying statistical density of the data and consequently the learned representations
have better encoding and denoising properties than generic bases. Unlike the case
for complete representations and the standard ICA algorithm, the transformation
562
M S. Lewicki and T. 1. Sejnowski
.....
...........
~
~
~
~
JvV
~
~
~ fIN ~ ~
~ ~ ."... f\;vJ'v
+
~
..
... + ..... ..... .......
~
.....
-JVV
H
~
~~J~~
~ ~ ~~~
A-
L
-L
J-.-L
JL~~JLi~~L~
~J~--L~
Figure 3: An example of fitting a 2x overcomplete representation to segments of from
natural speech. Each segment consisted of 64 samples, sampled at a frequency of 8000 Hz
(8 msecs). The plot shows a random sample of 30 of the 128 basis vectors (each scaled to
full range). The right graph shows the corresponding power spectral densities (0 to 4000
Hz).
from the data to the internal representation is non-linear. The probabilistic formulation of the basis inference problem offers the advantages that assumptions about
the prior distribution on the basis coefficients are made explicit and that different
models can be compared objectively using log P(xIA).
References
Amari, S., Cichocki, A., and Yang, H. H. (1996). A new learning algorithm for blind
signal separation. In Advances in Neural and Information Processing Systems,
volume 8, pages 757-763, San Mateo. Morgan Kaufmann.
Bell, A. J. and Sejnowski, T. J. (1995). An information maximization approach to
blind separation and blind deconvolution. Neural Computation, 7(6}:1129-1159.
Chen, S., Donoho, D. L., and Saunders, M. A. (1996). Atomic decomposition by
basis pursuit. Technical report, Dept. Stat., Stanford Univ., Stanford, CA.
Comon, P., Jutten, C., and Herault, J. (1991). Blind separation of sources .2.
problems statement. Signal Processing, 24(1}:11-20.
Jutten, C. and Herault, J. (1991). Blind separation of sources .1. an adaptive
algorithm based on neuromimetic architecture. Signal Processing, 24(1):1-10.
MacKay, D. J. C. (1996). Maximum likelihood and covariant algorithms for independent component analysis. University of Cambridge, Cavendish Laboratory.
Available at ftp: / /wol. ra. phy. cam. ac. uk/pub/mackay / ica. ps. gz.
Meszaros, C. (1997). BPMPD: An interior point linear programming solver. Code
available at ftp: / /ftp.netlib. org/ opt/bpmpd. tar. gz.
Olshausen, B. A. and Field, D. J. (1996). Emergence of simple-cell receptive-field
properties by learning a sparse code for natural images. Nature, 381:607-609.
| 1424 |@word determinant:1 illustrating:1 norm:2 decomposition:1 simplifying:2 solid:1 phy:1 contains:1 pub:1 si:1 must:1 additive:2 remove:1 plot:2 fewer:1 provides:2 org:1 become:1 fitting:3 ra:1 ica:4 rapid:1 considering:1 solver:1 becomes:1 underlying:3 notation:2 matched:1 z:2 finding:3 transformation:1 exactly:1 scaled:2 uk:1 unit:1 medical:1 positive:5 encoding:1 ak:1 approximately:2 plus:1 mateo:1 limited:2 slx:1 range:1 atomic:1 hughes:1 definite:1 bell:2 significantly:1 alleviated:1 projection:1 induce:2 cannot:3 interior:3 influence:1 equivalent:2 maximizing:1 rectangular:1 stabilizing:1 rule:7 deriving:3 handle:1 jli:1 cavendish:1 exact:2 programming:3 element:5 approximated:1 atw:1 breakdown:1 database:1 observed:1 solved:2 capture:1 removed:3 vanishes:1 cam:1 segment:3 efficiency:2 basis:35 multimodal:1 derivation:1 univ:1 cod:1 describe:1 sejnowski:6 saunders:1 whose:1 quite:1 larger:1 valued:2 stanford:2 drawing:1 amari:2 maxp:2 objectively:1 torrey:1 gp:2 emergence:1 advantage:1 zm:1 combining:1 convergence:1 p:1 generating:1 ftp:3 derive:2 ac:1 stat:1 alj:1 involves:1 direction:2 drawback:1 filter:1 viewing:1 wol:1 generalization:1 opt:1 probable:7 singularity:1 extension:1 batchsize:1 around:1 exp:1 mapping:1 pine:1 achieves:1 uniqueness:1 tanh:1 overcompleteness:1 gaussian:7 aim:2 alek:1 avoid:1 hj:3 tar:1 ax:1 rank:2 likelihood:1 criticism:1 sense:2 inference:1 entire:1 typically:1 selects:2 overall:1 herault:3 mackay:3 field:3 equal:1 ng:1 identical:1 broad:1 report:1 randomly:1 preserve:1 mixture:1 yielding:1 integral:2 orthogonal:3 initialized:2 desired:2 overcomplete:24 column:2 disadvantage:1 logp:2 maximization:1 subset:1 density:5 probabilistic:4 terrence:1 invertible:2 michael:1 ym:2 together:1 nm:2 possibly:1 ek:4 derivative:2 coding:5 coefficient:9 rescales:1 blind:6 view:1 lab:1 relied:1 timit:1 minimize:2 square:1 publicly:1 kaufmann:1 efficiently:1 ensemble:1 yield:4 misfit:1 multiplying:1 frequency:1 sampled:1 dataset:3 dimensionality:1 rim:1 routine:1 higher:1 improved:1 formulation:1 evaluated:1 done:1 d:1 nonlinear:6 jutten:3 defines:1 olshausen:2 effect:1 normalized:1 true:5 requiring:1 consisted:1 analytically:1 laboratory:1 width:1 speaker:1 generalized:1 complete:5 demonstrate:1 l1:3 image:1 wise:1 recently:1 specialized:2 volume:1 jl:1 approximates:1 bmn:4 cambridge:1 rd:1 had:1 base:12 posterior:5 hide:1 optimizing:1 jolla:1 inverted:1 morgan:1 minimum:2 greater:2 maximize:2 redundant:1 signal:4 dashed:1 full:2 smooth:1 technical:1 adapt:1 offer:1 hmn:1 laplacian:11 involving:1 essentially:1 iteration:3 represent:3 invert:2 cell:1 source:9 unlike:2 hz:2 presence:1 noting:1 yang:1 variety:1 fit:3 architecture:1 det:2 expression:3 pca:2 utility:1 speech:5 hessian:6 speaking:1 ten:1 reduced:3 specifies:5 problematic:1 estimated:1 per:2 redundancy:3 putting:1 rewriting:1 graph:2 run:1 inverse:3 package:1 separation:7 vex:1 bit:2 adapted:2 precisely:1 dominated:1 fourier:6 span:1 min:1 tv:1 according:1 conjugate:1 smaller:1 comon:2 restricted:1 gradually:1 equation:2 letting:2 jvv:2 neuromimetic:1 available:3 operation:2 pursuit:1 appropriate:1 generic:1 spectral:1 stepsize:2 alternative:3 original:1 remaining:1 const:1 objective:3 receptive:1 traditional:2 diagonal:2 gradient:6 subspace:1 assuming:3 sia:1 code:2 modeled:1 potentially:1 statement:1 negative:2 sm:7 howard:1 fin:1 defining:1 neurobiology:1 inferred:1 inverting:1 required:1 sentence:1 learned:6 program:1 including:1 terry:1 power:2 natural:4 mn:4 gz:2 cichocki:1 prior:21 literature:1 l2:3 marginalizing:1 mixed:2 degree:1 ata:1 row:2 summary:1 aij:3 allow:2 institute:2 differentiating:1 sparse:4 distributed:2 xia:9 valid:1 made:1 adaptive:1 preprocessing:1 san:1 far:1 approximate:2 pseudoinverse:1 spectrum:1 channel:2 nature:1 robust:1 ca:2 symmetry:1 necessarily:1 vj:1 cmn:3 main:1 noise:10 broadband:1 salk:3 precision:1 inferring:2 msec:1 explicit:1 exponential:1 pe:5 third:1 wavelet:1 ix:1 removing:1 down:1 deconvolution:1 magnitude:1 chen:3 simply:4 lewicki:5 covariant:1 loses:1 determines:1 netlib:1 viewed:2 marked:1 consequently:2 donoho:1 change:4 wt:1 denoising:1 principal:1 called:1 tendency:1 la:1 internal:4 arises:1 dept:1 ex:1 |
467 | 1,425 | ?
Experiences with Bayesian Learning In
a
Real World Application
Peter Sykacek, Georg Dorffner
Austrian Research Institute for Artificial Intelligence
Schottengasse 3, A-10ID Vienna Austria
peter, georg@ai.univie.ac.at
Peter Rappelsberger
Institute for Neurophysiology at the University Vienna
Wahringer StraBe 17, A-lOgO Wien
Peter.Rappelsberger@univie.ac.at
Josef Zeitlhofer
Department of Neurology at the AKH Vienna
Wahringer Giirtel 18-20, A-lOgO Wien
Josef.Zeitlhofer@univie.ac.at
Abstract
This paper reports about an application of Bayes' inferred neural network classifiers in the field of automatic sleep staging. The
reason for using Bayesian learning for this task is two-fold. First,
Bayesian inference is known to embody regularization automatically. Second, a side effect of Bayesian learning leads to larger
variance of network outputs in regions without training data. This
results in well known moderation effects, which can be used to
detect outliers. In a 5 fold cross-validation experiment the full
Bayesian solution found with R. Neals hybrid Monte Carlo algorithm, was not better than a single maximum a-posteriori (MAP)
solution found with D.J. MacKay's evidence approximation. In a
second experiment we studied the properties of both solutions in
rejecting classification of movement artefacts.
Experiences with Bayesian Learning in a Real World Application
1
965
Introduction
Sleep staging is usually based on rules defined by Rechtschaffen and Kales (see [8]).
Rechtschaffen and Kales rules define 4 sleep stages, stage one to four, as well as rapid
eye movement (REM) and wakefulness. In [1] J. Bentrup and S. Ray report that
every year nearly one million US citizens consulted their physicians concerning their
sleep. Since sleep staging is a tedious task (one all night recording on average takes
abou t 3 hours to score manually), much effort was spent in designing automatic
sleep stagers.
Sleep staging is a classification problem which was solved using classical statistical
t.echniques or techniques emerged from the field of artificial intelligence (AI) . Among
classical techniques especially the k nearest neighbor technique was used. In [1]
J. Bentrup and S. Ray report that the classical technique outperformed their AI
approaches. Among techniques from the field of AI, researchers used inductive
learning to build tree based classifiers (e.g. ID3, C4.5) as reported by M. Kubat et.
a1. in [4]. Neural networks have also been used to build a classifier from training
examples. Among those who used multi layer perceptron networks to build the
classifier, the work of R. Schaltenbrand et. a1. seems most interesting. In [to] they
use a separate network to refuse classification of too distant input vectors. The
performance usually reported is in the range of 75 to 85 percent.
\Vhich enhancements to these approaches can be made to get a. reliable system
wit.h hopefully better performance? According to S. Roberts et . al. in [9], outlier
detection is important to get reliable results in a critical (e.g. medical) environment.
To get reliable results one must refuse classification of dubious inputs. Those inputs
are marked separately for further inspection by a human expert. To be able to
detect such dubious inputs, we use Bayesian inference to calculate a distribution
over the neural network weights. This approach automatically incorporates the
calculation of confidence for each network estimate. Bayesian inference has the
further advantage that regularization is part of the learning algorithm. Additional
methods like weight decay penalty a.nd cross validation for decay parameter tuning
are no longer needed . Bayesian inference for neural networks was among others
investigated by D.J. MacKay (see [5]), Thodberg (see [11]) and Buntine and Weigend
(~ee [3]).
The a.im of this paper is to study how Bayesian inference leads to probabilities for
classes, which together with doubt levels allow to refuse classification of outliers.
As we are interested in evaluating the resulting performance, we use a comparative
method on the same data set and use a significance test, such that the effect of the
method can easily be evaluated.
2
Methods
In this section we give a short description of the inference techniques used to perform
the experiments. We have used two approaches using neural networks as classifiers
and an instance based approach in order to make the performance estimates comparable to other methods .
2.1
Architecture for polychotomous classification
For polychotomous classification problems usually a l-of-c target coding scheme is
used. Usually it is sufficient to use a network architecture with one hidden layer. In
[2] pp. 237-240, C. Bishop gives a general motivation for the softmax data model,
P. Sykacek, G. Dorffner; P. Rappelsberger and 1. Zeitlhofer
966
which should be used if one wants the network outputs to be probabilities for classes.
If we assume that the class conditional densities, p(? I Ck), of the hidden unit activation vector, ?, are from the general family of exponential distributions, then using
t.he transformation in (1), allows to interpret the network outputs as probabilities
for classes. This transformation is known as normalized exponential or softmax
activation function.
p(Ck 1?) =
exp(ak)
(1)
Lkl exp(ak l )
In ! 1) t.he value ak is the value at output node k before applying softmax activation. Softmax transformation of the activations in the output layer is used for both
network approaches used in this paper.
2.2
Bayesian Inference
In [6] D.J. MacKay uses Bayesian inference and marginalization to get moderated
probabilities for classes in regions where the network is uncertain about the class
label. In conjunction with doubt levels this allows to suppress a classification of
such patterns. A closer investigation of this approach showed that marginalization
leads to moderated probabilities, but the degree of moderation heavily depends on
the direction in which we move away from the region with sufficient training data.
Therefore one has to be careful about whether the moderation effect should be used
for outliers detection.
A Bayesian solution for neural networks is a posterior distribution over weight space
calculated via Bayes' theorem using a prior over weights.
(2)
In (2), w is the weight vector of the network and V represents the training data.
Two different possibilities are known to calculate the posterior in (2). In [5] D.J .
MacKay derives an analytical expression assuming a Gaussian distribution. In [7]
R. Neal uses a hybrid Monte Carlo method to sample from the posterior. For one
input pattern, the posterior over weight space will lead to a distribution of network
outputs.
For a classification problem, following MacKay [6], the network estimate is calculated by marginalization over the output distribution.
P(C 1
I~, V)
=.J
P(C 1
=
I~, w)p(w I V)dw
Jy(~,
w)p(w I V)dw
(3)
In general, the distribution over output activations will have small variance in regions well represented in the training data and large variance everywhere else. The
reason for that is the influence of the likelihood term p(V I w), which forces the
network mapping to lie close to the desired one in regions with training data, but
which has no influence on the network mapping in regions without training data.
At least for for generalized linear models applied to regression, this property is
quantifiable. In [12] C. Williams et.al. showed that the error bar is proportional to
the inverse input data density p(~)-l. A similar relation is also plausible for the
output activation in classification problems.
Experiences with Bayesian Learning in a Real World Application
967
Due to the nonlinearity of the softmax transformation, marginalization will moderate probabilities for classes. Moderation will be larger in regions with large variance
of the output activation. Compared to a decision made with the most probable
weight, the network guess for the class label will be less certain. This moderation
effect allows to reject classification of outlying patterns.
Since upper integral can not be solved analytically for classification problems, there
are t.wo possibilities to solve it. In [6] D.J. MacKay uses an approximation. Using
hybrid Monte Carlo sampling as an implementation of Bayesian inference (see R.
Neal in [7]), there is no need to perform upper integration analytically. The hybrid
Monte Carlo algorithm samples from the posterior and upper integral is calculated
as a finite sum.
1 L
P(C 1 I~, 1)) ~ L LY(~' Wi)
(4)
i=l
Assuming, that the posterior over weights is represented exactly by the sampled
weights, there is no need to limit the number of hidden units, if a correct (scaled)
prior is used. Consequently in the experiments the network size was chosen to be
large. We used 25 hidden units. Implementation details of the hybrid Monte Carlo
algorithm may be found in [7].
2.3
The Competitor
The classifier, used to give performance estimates to compare to, is built as a two
layer perceptron network with softmax transformation applied to the outputs. As
an error function we use the cross entropy error including a consistent weight decay
penalty, as it is e.g. proposed by C. Bishop in [2], pp. 338. The decay parameters
are estimated with D.J. MacKay's evidence approximation ( see [5] for details).
Note that the restriction of D.J. MacKay's implementation of Bayesian learning,
which has no solution to arrive at moderated probabilities in l-of-c classification
problems, do not apply here since we use only one MAP value. The key problem
with this approach is the Gaussian approximation of the posterior over weights,
which is used to derive the most probable decay parameters. This approximation is
certainly only valid if the number of network parameters is small compared to the
number of training samples. One consequence is, that the size of the network has
to be restricted . Our model uses 6 hidden units.
To make the performance of the Bayes inferred classifier also comparable to other
methods, we decided to include performance estimates of a k nearest neighbor
algorithm. This algorithm is easy to implement and from [1] we have some evidence
that its performance is good.
3
Experiments and Results
In this sect.ion we discuss the results of a sleep staging experiment based on the
t.echniques described in the "Methods" section.
3.1
Data
All experiments are performed with spectral features calculated from a database of 5
different healthy subjects. All recordings were scored according to the Rechtschaffen
& Kales rules. The data pool consisted from data calculated for all electrodes
968
P. Sykacek, G. Doif.fner, P. Rappelsberger and J. Zeitlhofer
available, which were horizontal eye movement, vertical eye movement and 18 EEG
f'lectrodes placed with respect to the international 10-20 system.
The data were transformed into the frequency domain. We used power density
values as well as coherency between different electrodes, which is a correlation
coefficient expressed as a function of frequency as input features. All data were
transformed to zero mean and unit variance. From the resulting feature space we
selected 10 features, which were used as inputs for classification. Feature selection
was done with a suboptimal search algorithm which used the performance of a k
nearest neighbor classifier for evaluation. We used more than 2300 samples during
t.raining and about 580 for testing.
3.2
Analysis of Both Classifiers
The analysis of both classifiers described in the "Methods" section should reveal
whether besides good classification performance the Bayes' inferred classifier is also
,apable of refusing outlying test patterns. Increasing the doubt level should lead to
better results of the classifier trained by Bayesian Inference if the test data contains
out.lying patterns. We performed two experiments. During the first experiment
Wf' calculated results from a 5 fold cross validation, where training is done with 4
subjects and tests are performed with one independent test person. In a second
j,f'St. we examine the differences of both algorithms on patterns which are definitely
outliers. We used the same classifiers as in the first experiment. Test patterns for
t his experiment were classified movement artefacts, which should not be classified
as one of the sleep stages.
The classifier used in conjunction with Bayesian inference was a 2-layer neural
net.work with 10 inputs, 25 hidden units with sigmoid activation and five output
units with softmax activation. The large number of hidden units is motivated by
the results reported from R. Neal in [7]. R. Neal studied the properties of neural
networks in a Bayesian framework when using Gaussian priors over weights. He
concluded that there is no need for limiting the complexity of the network when
using a correct Bayesian approach. The standard deviation of the Gaussian prior
1S scaled by the number of hidden units.
For the comparative approach we used a neural network with 10 inputs, 6 hidden
units and 5 outputs with softmax activation. Optimization was done via the BFGS
algorit.hm (see C. Bishop in [2]) with automatic weight decay parameter tuning
(D.J. MacKay's evidence approximation). As described in the methods section,
the smaller network used here is motivated by the Gaussian approximation of the
posterior over weights, which is used in the expression for the most probable decay
parameters.
The third result is a result achieved with a k nearest neighbor classifier with k set
to three.
All results are summaried in table 1. Each column summarizes the results achieved
with one of the algorithms and a certain doubt level during the cross validation run.
As the k nearest neighbor classifier gives only coarse probability estimates, we give
only the performance estimate when all test patterns are classified.
An examination of table 1 shows that the differences between the MAP-solution
and the Bayesian solution are extremely small. Consequently, using a t-test, the
O-hypothesis could not be rejected at any reasonable significance level. On the other
hand compared to the Bayesian solution, the performance of the k nearest neighbor
classifier is significantly lower (the significance level is 0.001).
Experiences with Bayesian Learning in a Real World Application
969
Table 1: Classification Performance
Doubt Cases
Mean Perf.
Std. Dev.
Doubt Cases
Mean Perf.
Std. Dev.
Doubt Cases
Mean Perf.
Std. Dev.
MAP
0
5%
100/0
78.6% 80.4% 81.6%
9.1%
9.4%
9.4%
Bayes
0
5%
10%
78.4% 80.2% 82.2%
9.0%
8.6%
9.4%
k nearest neighbor
0
5%
10%
74.6%
8.4%
-
15%
83.2'70
9.1%
15%
83.6%
9.1%
15%
-
Table 2: Rejection of Movement Periods
Method
recognized outliers
No.
0
1
2
0
1
MAP
%
0%
7.7%
15.4%
0%
7.7%
No.
1
6
5
5
3
Bayes
%
7.1%
46.f%
38.5%
38.5%
23.1%
The last experiment revealed that both training algorithms lead to comparable performance estimates, when clean data is used. When using the classifier in practice
there is no guarantee that the data are clean. One common problem of all night
recordings are the so called movement periods, which are periods with muscle activity due to movements of the sleeping subject. During a second experiment we tried
t.o assess the robustness of both neural classifiers against such inputs. During this
experiment we used a fixed doubt level, for which approximately 5% of the clean
t.est. data from the last experiment were rejected. With this doubt level we classified
13 movement periods, which should not be assigned to any of the other stages. The
number of correctly refused outlying patterns are shown in table 2. Analysis of the
results with a t-test showed a significant higher rate of removed outliers for the full
Bayesian approach. Nevertheless as the number of misclassified outliers is large,
one has to be careful in using this side-effect of Bayesian inference.
4
Conclusion
Using Bayesian Inference for neural network training is an approach which leads to
better classification results compared with simpler training procedures. Comparing
wit.h the "one MAP" solution, we observed significantly larger reliability in detecting
dubious patterns. The large amount of remaining misclassified patterns, which were
obviously outlying, shows that we should not rely blindly on the moderating effect
of marginalization. Despite the large amount of time which is required to calculate
t.he solution, Bayesian inference has relevance for practical applications. On one
hand the Bayesian solution shows good performance. But the main reason is the
a.bility to encode a validity region of the model into the solution. Compared to all
methods which do not aim at a predictive distribution, this is a clear advantage for
Bayesian inference.
P. Sykacek, G. Doif.fner, P. Rappelsberger and 1. Zeitlhofer
970
Acknowledgements
We want to acknowledge the work of R. Neal from the Departments of Statistics
and Computer Science at the University of Toronto, who made his implementation
of hybrid Monte-Carlo sampling for Bayesian inference available electronically. His
software was used to calculate the full Bayes' inferred classification results. We also
want to express gratitude to S. Roberts from Imperial College London, one of the
partners in the ANNDEE project. His work and his consequence in insisting on
confidence measures for network decisions had a large positive impact on our work.
This work was sponsored by the Austrian Federal Ministry of Science and Transport.
It was done in the framework of the BIOMED 1 concerted action ANNDEE, financed
by the European Commission, DG. XII.
References
[1] J.A. Bentrup and S.R. Ray. An examination of inductive learning algorithms for
the classification of sleep signals. Technical Report UIUCDCS-R-93-1792, Dept of
Computer Science, University of Illinois, Urbana-Champaign, 1993.
[3] C. M. Bishop. Neural Networks for Pattern Recognition. Clarendon Press, Oxford,
1995.
[3] W. L. Buntine and A. S. Weigend. Bayesian back-propagation. Complex Systems,
5:603-643, 1991.
[4] M. Kubat, G. Pfurtscheller, and D. Flotzinger. Discrimination and classification using
bot.h binary and continuous variables. Biological Cybernetics, 70:443-448, 1994.
[5] D. J. C. MacKay. Bayesian interpolation. Neural Computation, 4:415-447, 1992.
[6] D. J. C. MacKay. The evidence framework applied to classification networks. Neural
Computation, 4:720-736, 1992.
[7] R. M. Neal. Bayesian Learning for Neural Networks. Springer, New York, 1996.
[8] A. Rechtschaffen and A. Kales. A manual of standardized terminology, techniques
and scoring system for sleep stages of human subjects. NIH Publication No. 204, US
Government Printing Office, Washington, DC., 1968.
[9] S. Roberts, L. Tarassenko, J. Pardey, and D. Siegwart. A confidence measure for
artificial neural networks. In International Conference Neural Networks and Expert
Systems in Medicine and Healthcare, pages 23-30, Plymouth, UK, 1994.
[10] N. Schaltenbrand, R. Lengelle, and J.P. Macher. Neural network model: application
to automatic analysis of human sleep. Computers and Biomedical Research, 26:157171, 1993.
~ll]
H. H. Thodberg. A review of bayesian neural networks with an application to near
infrared spectroscopy. IEEE Transactions on Neural Networks, 7(1):56-72, January
1996.
[12] C. K. I. Williams, C. Quazaz, C. M. Bishop, and H. Zhu. On the relationship between
bayesian error bars and the input data density. In Fourth International Conference
on Artificial Neural Networks, Churchill Col/ege, University of Cambridge, UK. lEE
Conference Publication No. 409, pages 160-165, 1995.
| 1425 |@word neurophysiology:1 seems:1 nd:1 tedious:1 tried:1 abou:1 contains:1 score:1 comparing:1 activation:10 must:1 distant:1 sponsored:1 discrimination:1 intelligence:2 selected:1 guess:1 inspection:1 short:1 detecting:1 coarse:1 node:1 toronto:1 simpler:1 five:1 stager:1 ray:3 concerted:1 rapid:1 embody:1 examine:1 bility:1 multi:1 rem:1 automatically:2 increasing:1 project:1 kubat:2 churchill:1 transformation:5 guarantee:1 vhich:1 every:1 exactly:1 classifier:19 scaled:2 uk:2 healthcare:1 unit:10 medical:1 ly:1 before:1 positive:1 limit:1 consequence:2 despite:1 ak:3 id:1 oxford:1 interpolation:1 approximately:1 logo:2 studied:2 range:1 decided:1 practical:1 testing:1 practice:1 implement:1 procedure:1 reject:1 significantly:2 confidence:3 flotzinger:1 get:4 close:1 selection:1 applying:1 influence:2 restriction:1 map:6 kale:4 williams:2 wit:2 rule:3 his:5 dw:2 moderated:3 limiting:1 target:1 heavily:1 us:4 designing:1 hypothesis:1 recognition:1 std:3 infrared:1 tarassenko:1 database:1 observed:1 solved:2 moderating:1 calculate:4 algorit:1 region:8 sect:1 movement:9 removed:1 environment:1 complexity:1 trained:1 predictive:1 easily:1 represented:2 london:1 monte:6 artificial:4 emerged:1 larger:3 plausible:1 solve:1 statistic:1 id3:1 obviously:1 advantage:2 analytical:1 net:1 wakefulness:1 financed:1 description:1 quantifiable:1 enhancement:1 electrode:2 comparative:2 spent:1 derive:1 ac:3 nearest:7 artefact:2 direction:1 correct:2 human:3 government:1 investigation:1 probable:3 biological:1 im:1 lying:1 lkl:1 exp:2 mapping:2 outperformed:1 label:2 healthy:1 federal:1 gaussian:5 aim:1 ck:2 publication:2 conjunction:2 office:1 encode:1 likelihood:1 detect:2 wf:1 posteriori:1 inference:16 hidden:9 relation:1 transformed:2 misclassified:2 interested:1 josef:2 biomed:1 classification:21 among:4 softmax:8 mackay:11 integration:1 field:3 washington:1 sampling:2 manually:1 represents:1 nearly:1 report:4 others:1 dg:1 detection:2 possibility:2 evaluation:1 certainly:1 staging:5 citizen:1 integral:2 closer:1 plymouth:1 experience:4 tree:1 desired:1 siegwart:1 uncertain:1 instance:1 column:1 dev:3 deviation:1 too:1 buntine:2 reported:3 commission:1 person:1 density:4 international:3 st:1 definitely:1 lee:1 physician:1 rechtschaffen:4 pool:1 together:1 polychotomous:2 pardey:1 expert:2 doubt:9 bfgs:1 wien:2 coding:1 coefficient:1 depends:1 performed:3 bayes:7 neals:1 ass:1 variance:5 who:2 bayesian:35 rejecting:1 carlo:6 refusing:1 researcher:1 cybernetics:1 classified:4 manual:1 competitor:1 against:1 pp:2 frequency:2 sampled:1 austria:1 back:1 clarendon:1 higher:1 evaluated:1 done:4 rejected:2 sykacek:4 stage:5 biomedical:1 correlation:1 hand:2 horizontal:1 night:2 transport:1 hopefully:1 propagation:1 reveal:1 effect:7 validity:1 normalized:1 consisted:1 inductive:2 regularization:2 analytically:2 assigned:1 neal:6 ll:1 during:5 generalized:1 percent:1 nih:1 sigmoid:1 common:1 dubious:3 million:1 he:4 interpret:1 significant:1 cambridge:1 ai:4 automatic:4 tuning:2 nonlinearity:1 illinois:1 had:1 reliability:1 longer:1 posterior:8 showed:3 moderate:1 certain:2 binary:1 ege:1 muscle:1 scoring:1 ministry:1 additional:1 recognized:1 period:4 signal:1 full:3 champaign:1 technical:1 calculation:1 cross:5 concerning:1 jy:1 a1:2 impact:1 regression:1 austrian:2 blindly:1 achieved:2 ion:1 sleeping:1 want:3 separately:1 else:1 concluded:1 recording:3 subject:4 incorporates:1 ee:1 near:1 revealed:1 easy:1 marginalization:5 architecture:2 suboptimal:1 whether:2 expression:2 motivated:2 effort:1 penalty:2 wo:1 peter:4 dorffner:2 york:1 schottengasse:1 action:1 clear:1 amount:2 coherency:1 bot:1 estimated:1 correctly:1 xii:1 georg:2 express:1 key:1 four:1 terminology:1 nevertheless:1 imperial:1 clean:3 year:1 sum:1 weigend:2 run:1 inverse:1 everywhere:1 fourth:1 arrive:1 family:1 reasonable:1 decision:2 summarizes:1 comparable:3 layer:5 fold:3 sleep:12 activity:1 software:1 extremely:1 department:2 according:2 smaller:1 wi:1 outlier:8 restricted:1 discus:1 needed:1 available:2 apply:1 away:1 spectral:1 robustness:1 standardized:1 remaining:1 include:1 vienna:3 medicine:1 especially:1 build:3 classical:3 move:1 separate:1 partner:1 reason:3 assuming:2 echniques:2 besides:1 relationship:1 robert:3 suppress:1 implementation:4 perform:2 upper:3 vertical:1 urbana:1 finite:1 acknowledge:1 anndee:2 january:1 dc:1 consulted:1 inferred:4 gratitude:1 required:1 c4:1 hour:1 able:1 bar:2 usually:4 univie:3 pattern:12 refuse:3 built:1 reliable:3 including:1 power:1 critical:1 moderation:5 hybrid:6 force:1 examination:2 rely:1 zhu:1 scheme:1 eye:3 perf:3 hm:1 prior:4 review:1 acknowledgement:1 interesting:1 proportional:1 validation:4 degree:1 sufficient:2 consistent:1 thodberg:2 placed:1 last:2 electronically:1 side:2 allow:1 perceptron:2 institute:2 neighbor:7 calculated:6 raining:1 world:4 evaluating:1 valid:1 made:3 outlying:4 transaction:1 neurology:1 search:1 continuous:1 table:5 eeg:1 spectroscopy:1 investigated:1 european:1 complex:1 domain:1 significance:3 main:1 motivation:1 scored:1 pfurtscheller:1 exponential:2 col:1 lie:1 third:1 printing:1 theorem:1 bishop:5 decay:7 evidence:5 derives:1 rejection:1 entropy:1 expressed:1 springer:1 refused:1 insisting:1 uiucdcs:1 conditional:1 marked:1 consequently:2 careful:2 called:1 est:1 college:1 relevance:1 dept:1 |
468 | 1,426 | RCC Cannot Compute Certain FSA,
Even with Arbitrary Transfer Functions
Mark Ring
RWCP Theoretical Foundation GMD Laboratory
GMD - German National Research Center for Information Technology
Schloss Birlinghoven
D-53 754 Sankt Augustin, Germany
email: Mark .Ring@GMD.de
Abstract
Existing proofs demonstrating the computational limitations of Recurrent Cascade Correlation and similar networks (Fahlman, 1991;
Bachrach, 1988; Mozer, 1988) explicitly limit their results to units
having sigmoidal or hard-threshold transfer functions (Giles et aI.,
1995; and Kremer, 1996). The proof given here shows that for
any finite, discrete transfer function used by the units of an RCC
network, there are finite-state automata (FSA) that the network
cannot model, no matter how many units are used. The proof also
applies to continuous transfer functions with a finite number of
fixed-points, such as sigmoid and radial-basis functions.
1
Introduction
The Recurrent Cascade Correlation (RCC) network was proposed by Fahlman
(1991) to offer a fast and efficient alternative to fully connected recurrent networks.
The network is arranged such that each unit has only a single recurrent connection:
the connection that goes from itself to itself. Networks with the same structure
have been proposed by Mozer (Mozer, 1988) and Bachrach (Bachrach, 1988). This
structure is intended to allow simplified training of recurrent networks in the hopes
of making them computationally feasible. However, this increase in efficiency comes
at the cost of computational power: the networks' computational capabilities are
limited regardless of the power of their activation functions. The remaining input to
each unit consists of the input to the network as a whole together with the outputs
from all units lower in the RCC network. Since it is the structure of the network
and not the learning algorithm that is of interest here, only the structure will be
described in detail.
M. Ring
620
Figure 1: This finite-state automaton was shown by Giles et al. (1995) to be unrepresentable by an Ree network whose units have hard-threshold or sigmoidal
transfer functions. The arcs are labeled with transition labels of the FSA which
are given as input to the Ree network. The nodes are labeled with the output
values that the network is required to generate. The node with an inner circle is an
accepting or halting state .
Figure 2: This finite-state automaton is one of those shown by Kremer (1996) not to
be representable by an Ree network whose units have a hard-threshold or sigmoidal
transfer function . This FSA computes the parity of the inputs seen so far .
The functionality of a network of N
the following way:
Ree
units, Uo, .. ., UN-l can be described in
1?
(1)
/o([(t), Vo(t /x(i(t), Vx(t - 1), Vx-1(t), Vx-2(t), .. . , Vo(t?,
(2)
where Vx(t) is the output value of Ux at time step t, and l(t) is the input to the
network at time step t. The value of each unit is determined from: (1) the network
input at the current time step, (2) its own value at the previous time step, and (3)
the output values of the units lower in the network at the current time step . Since
learning is not being considered here, the weights are assumed to be constant.
2
Existing Proofs
The proof of Giles, et al (1995) showed that an Ree network whose units had a
hard-threshold or sigmoidal transfer function cannot produce outputs that oscillate
with a period greater than two when the network input is constant. (An oscillation
has a period of x if it repeats itself every x steps.) Thus, the FSA shown in Figure 1
cannot be modeled by such an Ree network, since its output (shown as node labels)
oscillates at a period greater than two given constant input. Kremer (1996) refined
the class of FSA representable by an Ree network showing that, if the input to
the net oscillates with period p, then the output can only oscillate with a period of
w, where w is one of p's factors (or of 2p's factors if p is odd). An unrepresentable
example, therefore, is the parity FSA shown in Figure 2, whose output has a period
of four given the following input (of period two): 0,1,0,1, ....
Both proofs, that by Giles et al. and that by Kremer, are explicitly designed with
RCC Cannot Compute Certain FSA, Even with Arbitrary Transfer Functions
621
*0,1
Figure 3: This finite-state automaton cannot be modeled with any RCC network
whose units are capable of representing only k discrete outputs. The values within
the circles are the state names and the output expected from the network. The arcs
describe transitions from state to state, and their values represent the input given
to the network when the transition is made. The dashed lines indicate an arbitrary
number of further states between state 3 and state k which are connected in the
same manner as states 1,2, and 3. (All states are halting states.)
hard-threshold and sigmoidal transfer functions in mind, and can say nothing about
other transfer functions. In other words, these proofs do not demonstrate the limitations of the RCC-type network structure, but about the use of threshold units
within this structure. The following proof is the first that actually demonstrates
the limitations of the single-recurrent-link network structure.
3
Details of the Proof
This section proves that RCC networks are incapable even in principle of modeling
certain kinds of FSA, regardless of the sophistication of each unit's transfer function,
provided only that the transfer function be discrete and finite, meaning only that
the units of the RCC network are capable of generating a fixed number, k, of distinct
output values. (Since all functions implemented on a discrete computer fall into this
category, this assumption is minor. Furthermore, as will be discussed in Section 4,
the outputs of most interesting continuous transfer functions reduce to only a small
number of distinct values.) This generalized RCC network is proven here to be
incapable of modeling the finite-state automaton shown in Figure 3.
MRing
622
For ease of exposition, let us call any FSA of the form shown in Figure 3 an RFk+l
for Ring FSA with k + 1 states. I Further, call a unit whose output can be any of
k distinct values and whose input includes its own previous output, a DRU k for
Discrete Recurrent Unit. These units are a generalization ofthe units used by RCC
networks in that the specific transfer function is left unspecified. By proving the
network is limited when its units are DRUbs proves the limitations of the network's
structure regardless of the transfer function used.
Clearly, a DRUk+1 with a sufficiently sophisticated transfer function could by itself
model an RFk+1 by simply allocating one of its k + 1 output values for each of
the k + 1 states. At each step it would receive as input the last state of the FSA
and the next transition and could therefore compute the next state. By restricting
the units in the least conceivable manner, i.e., by reducing the number of distinct
output values to k, the RCC network becomes incapable of modeling any RFk+1
regardless of how many DRUk's the network contains. This will now be proven.
The proof is inductive and begins with the first unit in the network, which, after
being given certain sequences of inputs, becomes incapable of distinguishing among
any states of the FSA. The second step, the inductive step, proves that no finite
number of such units can 'assist a unit hi~her in the ReC network in making a
distinction between any states of the RFk+ .
Lemma 1 No DR Uk whose input is the current transition of an RFk+1 can reliably
distinguish among any states of the RP+I. More specifically, at least one of the
DR Uk,s k output values can be generated in all of the RP+I 's k + 1 states.
Proof: Let us name the DRUbs k distinct output values VO, VI, ... , Vk-I. The
mapping function implemented by the DRU k can be expressed as follows:
(V X , i)
=}
VY,
which indicates that when the unit's last output was V X and its current input is i,
then its next output is VY.
Since an RFk is cyclical, the arithmetic in the following will also be cyclical (i.e.,
modular):
where 0
~
xtfJy
=
x8y
-
{ x+y
x+y-k
x-y
{ x+k-y
if x + y < k
if x + y ~ k
if x 2: y
if x < y
x < k and 0 ~ y < k.
Since it is impossible for the DRU k to represent each of the RFk+I,s k +1 states with
a distinct output value, at least two of these states must be represented ambiguously
by the same value. That is, there are two RFk+l states a and b and one DRU k
value V a / b such that V a / b can be generated by the unit both when the FSA is in
state a and when it is in state b. Furthermore, this value will be generated by the
unit given an appropriate sequence of inputs. (Otherwise the value is unreachable,
serves no purpose, and can be discarded, reducing the unit to a DRU k- I .)
Once the DRU k has generated V a / b , it cannot in the next step distinguish whether
the FSA's current state is a or b. Since the FSA could be in either state a or b, the
next state after a b transition could be either a or b tfJ 1. That is:
(va/b, b) =} Va/bEl'll,
(3)
IThanks to Mike Mozer for suggesting this catchy name.
RCC Cannot Compute Certain FSA, Even with Arbitrary Transfer Functions
623
where a e b ~ be a and k > 1. This new output value Va/b$l can therefore be
generated when the FSA is in either state a or state b EB 1. By repeatedly replacing
b with b EB 1 in Equation 3, all states from b to a e 1 can be shown to share output
values with state a, i.e., V a / b , Va/b$l, V a / b$2, ... , va/ae2, v a / ae1 all exist.
Repeatedly substituting a eland a for a and b respectively in the last paragraph
produces values vx/y Vx, YEO, 1, ... , k + 1. There is, therefore, at least one value
that can be generated by the unit in both states of every possible pair of states.
Since there are
(k! 1) distinct pairs but only kdistinct output values, and since
when k > 1, then not all of these pairs can be represented by unique V values. At
least two of these pairs must share the same output value, and this implies that
some v a / b/ e exists that can be output by the unit in any of the three FSA states
a, b, and c.
Starting with
(V a / b/ e , c)
::::}
va/b/e$l,
and following the same argument given above for V a / b , there must be a vx/y/z for
all triples of states x,
Y, and z.
Since there are
distinct output values, and since
where k > 3, some
v a / b/ e / d
fi+ll
(k ~ 1) distinct triples but only k
> 1,
must also exist.
rer)l
This argument can be followed repeatedly since:
>1,
+ 1, including when m = k. Therefore, there is at least one
that can be output by the unit in all k + 1 states of the RFk+l.
Call this value and any other that can be generated in all FSA states ~,k. All
Vk>s are reachable (else they could be discarded and the above proof applied for
DRU I , / < k). When a Vk is output by a DRU k , it does not distinguish any states
of the RFH 1 .
for all m
<
k
VO/l/2f..fk/k+l
Lemma 2 Once a DRUk outputs a V k , all future outputs will a/so be Vk's.
Proof: The proof is simply by inspection, and is shown in the following table:
Actual State
Transition
Next State
x
xEB1
xEB2
xEB3
x
x
x
x
xEB1
x
xEB2
xEB3
x82
x81
x
x
x82
x81
M. Ring
624
If the unit's last output value was a Vk, then the FSA might be in any of its k + 1
possible states. As can be seen, if at this point any of the possible transitions is
given as input, the next state can also be any of the k + 1 possible states. Therefore,
no future inp'ut can ever serve to lessen the unit's ambiguity.
Theorem 1 An RGG network composed of any finite number of DR Uk 's cannot
model an Rpk+l.
Proof: Let us describe the transitions of an RCC network of N units by using the
following notation:
((VN-I , VN-2, ... , VI, Va), i) ~ (VN- I , VN- 2, ... , V{, V~),
where Vrn is the output value of the m'th unit (i.e., Urn) before the given input,
i, is seen by the network, and V~ is Urn's value after i has been processed by the
network. The first unit, Uo, receives only i and Va as input. Every other unit Ux
receives as input i and Vx as well as v~, y < x.
Lemma 1 shows that the first unit, Uo, will eventually generate a value vl, which
can be generated in any of the RFk+1 states. From Lemma 2, the unit will continue
to produce vl values after this point.
Given any finite number N of DRUk,s, Urn-I, ... , Uo that are producing their Vk
values, V~ -1' .. . , Vt, the next higher unit , UN, will be incapable of disambiguating
all states by itself, i.e., at least two FSA states , a and b, will have overlapping
output values,
Since none of the units UN-I, ... , o can distinguish between
any states (including a and b),
k
k
k) b
(a / b(JJ 1 k
Vk
k)
( (VNa/ b ,VN-I,??
?, VI'VO ' )~ VN
'VN- I '?? ?, I'VO '
assuming that be a ~ ae b and k > 1. The remainder of the prooffollows identically
along the lines developed for Lemmas 1 and 2. The result of this development is
that UN also has a set of reachable output values V~ that can be produced in any
state of the FSA. Once one such value is produced, no less-ambiguous value is ever
generated. Since no RCC network containing any number of DRU k's can over time
distinguish among any states of an RFHI, no such RCC network can model such
an FSA.
V;,p.
4
U
Continuous Transfer Functions
Sigmoid functions can generate a theoretically infinite number of output values;
if represented with 32 bits, they can generate 232 outputs. This hardly means,
however, that all such values are of use. In fact, as was shown by Giles et al. (1995),
if the input remains constant for a long enough period of time (as it can in all
RFHI'S) , the output of sigmoid units will converge to a constant value (a fixed
point) or oscillate between two values. This means that a unit with a sigmoid
transfer function is in principle a DRU 2 . Most useful continuous transfer functions
(radial-basis functions, for example), exhibit the same property, reducing to only a
small number of distinct output values when given the same input repeatedly. The
results shown here are therefore not merely theoretical, but are of real practical
significance and apply to any network whose recurrent links are restricted to self
connections.
5
Concl usion
No RCC network can model any FSA containing an RF k+1 (such as that shown
in Figure 3), given units limited to generating k possible output values, regardless
RCC Cannot Compute Certain FSA, Even with Arbitrary Transfer Functions
625
of the sophistication of the transfer function that generates these values. This
places an upper bound on the computational capabilities of an RCC network. Less
sophisticated transfer functions, such as the sigmoid units investigated by Giles et
al. and Kremer may have even greater limitations. Figure 2, for example, could
be modeled by a single sufficiently sophisticated DRU 2 , but cannot be modeled
by an RCe network composed of hard-threshold or sigmoidal units (Giles et al.,
1995; Kremer, 1996) because these units cannot exploit all mappings from inputs
to outputs. By not assuming arbitrary transfer functions, previous proofs could not
isolate the network's structure as the source of RCC's limitations.
References
Bachrach, J. R. (1988). Learning to represent state. Master's thesis, Department
of Computer and Information Sciences, University of Massachusetts, Amherst,
MA 01003.
Fahlman, S. E. (1991) . The recurrent cascade-correlation architecture. In Lippmann, R. P., Moody, J. E., and Touretzky, D. S., editors, Advances in Neural
Information Processing Systems 3, pages 190-196, San Mateo, California. Morgan Kaufmann Publishers.
Giles, C., Chen, D., Sun, G., Chen, H., Lee, Y., and Goudreau, M. (1995). Constructive learning of recurrent neural networks: Problems with recurrent cascade correlation and a simple solution. IEEE Transactions on Neural Networks,
6(4):829.
Kremer, S. C. (1996). Finite state automata that recurrent cascade-correlation
cannot represent. In Touretzky, D. S., Mozer, M. C., and Hasselno, M. E.,
editors, Advances in Neural Information Processing Systems 8, pages 679-686.
MIT Press. In Press.
Mozer, M. C. (1988). A focused back-propagation algorithm for temporal pattern
recognition. Technical Report CRG-TR-88-3, Department of Psychology, University of Toronto.
| 1426 |@word tr:1 contains:1 existing:2 current:5 x81:2 activation:1 must:4 designed:1 inspection:1 accepting:1 node:3 toronto:1 sigmoidal:6 along:1 consists:1 paragraph:1 manner:2 theoretically:1 expected:1 actual:1 becomes:2 provided:1 begin:1 notation:1 kind:1 unspecified:1 sankt:1 developed:1 temporal:1 every:3 oscillates:2 demonstrates:1 uk:3 unit:48 uo:4 rcc:20 producing:1 before:1 limit:1 ree:7 might:1 eb:2 mateo:1 ease:1 limited:3 unique:1 practical:1 cascade:5 word:1 radial:2 inp:1 cannot:13 impossible:1 center:1 go:1 regardless:5 starting:1 automaton:6 focused:1 bachrach:4 proving:1 distinguishing:1 recognition:1 rec:1 labeled:2 mike:1 rpk:1 connected:2 sun:1 mozer:6 serve:1 ae1:1 efficiency:1 basis:2 represented:3 distinct:10 fast:1 describe:2 refined:1 whose:9 modular:1 say:1 otherwise:1 itself:5 fsa:26 sequence:2 vrn:1 net:1 ambiguously:1 remainder:1 produce:3 generating:2 ring:5 recurrent:12 odd:1 minor:1 implemented:2 come:1 indicate:1 implies:1 functionality:1 vx:8 generalization:1 crg:1 sufficiently:2 considered:1 ae2:1 mapping:2 substituting:1 purpose:1 label:2 augustin:1 hope:1 mit:1 clearly:1 vk:7 indicates:1 vl:2 her:1 germany:1 among:3 unreachable:1 development:1 once:3 having:1 future:2 report:1 composed:2 national:1 lessen:1 intended:1 interest:1 usion:1 allocating:1 capable:2 circle:2 theoretical:2 modeling:3 giles:8 cost:1 amherst:1 lee:1 together:1 moody:1 thesis:1 ambiguity:1 containing:2 dr:3 yeo:1 suggesting:1 halting:2 de:1 includes:1 matter:1 explicitly:2 vi:3 capability:2 unrepresentable:2 kaufmann:1 ofthe:1 produced:2 none:1 touretzky:2 email:1 x82:2 proof:16 massachusetts:1 ut:1 sophisticated:3 actually:1 back:1 higher:1 rfk:10 arranged:1 furthermore:2 correlation:5 receives:2 replacing:1 overlapping:1 propagation:1 name:3 inductive:2 laboratory:1 ll:2 self:1 ambiguous:1 generalized:1 demonstrate:1 vo:6 meaning:1 fi:1 sigmoid:5 discussed:1 ai:1 fk:1 had:1 reachable:2 concl:1 own:2 showed:1 certain:6 incapable:5 continue:1 vt:1 seen:3 morgan:1 greater:3 converge:1 schloss:1 period:8 dashed:1 arithmetic:1 technical:1 offer:1 long:1 va:8 ae:1 represent:4 rgg:1 receive:1 else:1 source:1 publisher:1 isolate:1 call:3 identically:1 enough:1 psychology:1 architecture:1 inner:1 reduce:1 whether:1 assist:1 oscillate:3 jj:1 repeatedly:4 hardly:1 useful:1 processed:1 category:1 gmd:3 generate:4 exist:2 vy:2 discrete:5 four:1 demonstrating:1 threshold:7 merely:1 master:1 place:1 rer:1 vn:7 oscillation:1 bit:1 bound:1 hi:1 followed:1 distinguish:5 generates:1 argument:2 urn:3 department:2 representable:2 making:2 restricted:1 computationally:1 equation:1 remains:1 german:1 eventually:1 mind:1 serf:1 dru:11 apply:1 appropriate:1 alternative:1 rp:2 remaining:1 rce:1 exploit:1 prof:3 x8y:1 exhibit:1 conceivable:1 link:2 assuming:2 modeled:4 reliably:1 upper:1 discarded:2 arc:2 finite:12 ever:2 arbitrary:6 pair:4 required:1 connection:3 bel:1 california:1 distinction:1 pattern:1 rf:1 including:2 power:2 representing:1 technology:1 fully:1 interesting:1 limitation:6 proven:2 triple:2 foundation:1 principle:2 editor:2 share:2 repeat:1 parity:2 last:4 fahlman:3 kremer:7 allow:1 fall:1 transition:9 computes:1 made:1 san:1 simplified:1 far:1 transaction:1 lippmann:1 assumed:1 vna:1 continuous:4 un:4 table:1 transfer:24 investigated:1 significance:1 whole:1 nothing:1 theorem:1 specific:1 showing:1 goudreau:1 exists:1 restricting:1 chen:2 sophistication:2 simply:2 tfj:1 expressed:1 ux:2 cyclical:2 applies:1 ma:1 exposition:1 disambiguating:1 feasible:1 hard:6 determined:1 specifically:1 reducing:3 infinite:1 lemma:5 mark:2 rwcp:1 constructive:1 |
469 | 1,427 | Enhancing Q-Learning for
Optimal Asset Allocation
Ralph Neuneier
Siemens AG, Corporate Technology
D-81730 MUnchen, Germany
Ralph.Neuneier@mchp.siemens.de
Abstract
This paper enhances the Q-Iearning algorithm for optimal asset allocation proposed in (Neuneier, 1996 [6]). The new formulation simplifies
the approach by using only one value-function for many assets and allows model-free policy-iteration. After testing the new algorithm on
real data, the possibility of risk management within the framework of
Markov decision problems is analyzed. The proposed methods allows
the construction of a multi-period portfolio management system which
takes into account transaction costs, the risk preferences of the investor,
and several constraints on the allocation.
1
Introduction
Asset allocation and portfolio management deal with the distribution of capital to various
investment opportunities like stocks, bonds, foreign exchanges and others. The aim is to
construct a portfolio with a maximal expected return for a given risk level and time horizon
while simultaneously obeying institutional or legally required constraints. To find such an
optimal portfolio the investor has to solve a difficult optimization problem consisting of
two phases [4]. First, the expected yields together with a certainty measure has to be predicted. Second, based on these estimates, mean-variance techniques are typically applied
to find an appropriate fund allocation. The problem is further complicated if the investor
wants to revise herlhis decision at every time step and if transaction costs for changing the
allocations must be considered.
disturbanc ies
- ,---
financial market
I---
j return
investmen ts
'----
investor
rates, prices
I---
Markov Decision Problem:
Xt = ($t, J(t}' state: market $t
and portfolio J(t
policy p, actions
at = p(xt)
transition probabilities
p(Xt+ll x d
r( Xt, at, $t+l) return function
Within the framework of Markov Decision Problems, MDPs, the modeling phase and the
search for an optimal portfolio can be combined (fig. above). Furthermore, transaction
costs, constraints, and decision revision are naturally integrated. The theory ofMDPs formalizes control problems within stochastic environments [1]. If the discrete state space is
small and if an accurate model of the system is available, MDPs can be solved by con-
937
Enhancing Q-Leaming for Optimal Asset Allocation
ventional Dynamic Programming, DP. On the other extreme, reinforcement learning methods using function approximator and stochastic approximation for computing the relevant
expectation values can be applied to problems with large (continuous) state spaces and
without an appropriate model available [2, 10].
In [6], asset allocation is fonnalized as a MDP under the following assumptions which
clarify the relationship between MDP and portfolio optimization:
1. The investor may trade at each time step for an infinite time horizon.
2. The investor is not able to influence the market by her/his trading.
3. There are only two possible assets for investing the capital.
4. The investor has no risk aversion and always invests the total amount.
The reinforcement algorithm Q-Learning, QL, has been tested on the task to invest liquid
capital in the Gennan stock market DAX, using neural networks as value function approximators for the Q-values Q(x, a). The resulting allocation strategy generated more profit
than a heuristic benchmark policy [6].
Here, a new fonnulation of the QL algorithm is proposed which allows to relax the third
assumption. Furthennore, in section 3 the possibility of risk control within the MDP framework is analyzed which relaxes assumption four.
2 Q-Learning with uncontrollable state elements
This section explains how the QL algorithm can be simplified by the introduction of an
artificial detenninistic transition step. Using real data, the successful application of the
new algorithm is demonstrated.
2.1
Q-Leaming for asset allocation
=
The situation of an investor is fonnalized at time step t by the state vector Xt
($t, Kt),
which consists of elements $t describing the financial market (e. g. interest rates, stock
indices), and of elements K t describing the investor's current allocation of the capital
(e. g. how much capital is invested in which asset). The investor's decision at for a new allocation and the dynamics on the financial market let the state switch to Xt+l = ($t+l' K t+1 )
according to the transition probability p(Xt+lIXto at). Each transition results in an immer(xt, Xt+l. at} which incorporates possible transaction costs depending
diate return rt
on the decision at and the change of the value of K t due to the new values of the assets at time t + 1. The aim is to maximize the expected discounted sum of the returns,
V* (x) = E(2::~o It rt Ixo = x). by following an optimal stationary policy J.l. (xt) = at.
For a discrete finite state space the solution can be stated as the recursive Bellman equation:
=
V? (xd = m:-x
[L
Xt+l
p(xt+llxt, a)rt
+
~I
L
p(xt+llxt. a) V* (Xt+l)] .
(1)
X.+l
A more useful fonnulationdefines a Q-function Q?(x, a) of state-action pairs (Xt. at),
to allow the application ofan iterative stochastic approximation scheme, called Q-Learning
[11]. The Q-value Q*(xt,a,) quantifies the expected discounted sum of returns if one
executes action at in state Xt and follows an optimal policy thereafter, i. e. V* (xt) =
max a Q* (Xt, a). Observing the tuple (Xt, Xt+l, at, rd, the tabulated Q-values are updated
938
R. Neuneier
in the k + 1 iteration step with learning rate 17k according to:
It can be shown, that the sequence of Qk converges under certain assumptions to Q* . If the
Q-values Q* (x, a) are approximated by separate neural networks with weight veCtor w a
for different actions a, Q* (x, a) ~ Q(x; w a ) , the adaptations (called NN-QL) are based on
the temporal differences d t :
dt := r(xt, at , Xt+l)
+ ,),maxQ(Xt+l;
w~) aEA
Q(Xt; wZ t )
,
Note, that although the market dependent part $t of the state vector is independent of the
investor's decisions, the future wealth Kt+l and the returns rt are not. Therefore, asset
allocation is a multi-stage decision problem and may not be reduced to pure prediction
if transaction costs must be considered. On the other hand, the attractive feature that the
decisions do not influence the market allows to approximate the Q-values using historical
data of the financial market. We need not to invest real money during the training phase.
2.2 Introduction of an artificial deterministic transition
Now, the Q-values are reformulated in order to make them independent of the actions chosen at the time step t. Due to assumption 2, which states that the investor can not influence
the market by the trading decisions, the stochastic process of the dynamics of $t is an uncontrollable Markov chain. This allows the introduction of a deterministic intermediate
step between the transition from Xt to Xt+1 (see fig. below). After the investor has "hosen an action at, the capital K t changes to K: because he/she may have paid transaction
c(Kt, at) and K; reflects the new allocation whereas the state of the market,
costs Ct
$t, remains the same. Because the costs Ct are known in advance, this transition is deterministic and controllable. Then, the market switches stochastically to $t+1 and generates
r' ($t, K:, $t+1) i.e., rt
Ct + r~ . The capital changes to
the immediate return r~
Kt+1
r~ + K; . This transition is uncontrollable by the investor. V* ($, K)
V* (x) is
now computed using the costs Ct and returns r~ (compare also eq. 1)
=
=
=
=
110<'_
...torml.lode
tn.sid.. ,
St
Kt
at
t+l
St
St+l
K't
Kt+l
Ct
Q(SbK~)
Defining Q* ($t,
=
r:
Kn as the Q-values of the intermediate time step
Q* ($t , K:)
E [r' ($t , K: , $t+1)
+ ')'V* ($t+1 ' Kt+d]
Enhancing Q-Leaming for Optimal Asset Allocation
939
gives rise to the optimal value function and policy (time indices are suppressed),
V* ($, K)
= max[c(K,
a) + Q* ($, K')],
a
Jl*($, K) = argmax[c(K, a)
a
+ Q*($, K')].
Defining the temporal differences d t for the approximation Qk as
dt := r' ($t, K:, $t+1)
+ ,max[c(Kt+b
a) + Q(k)($t+1, K:+ 1 )] a
Q(k)($t, KD
leads to the update equations for the Q-values represented by tables or networks:
QLU:
NN-QLU:
Q(k+l)($t,K;)
w(k+l)
Q(k)($t, K:) + 1/kdt ,
w(k)
+ 1/kdtV'Q($, K'; w(k?)
.
The simplification is now obvious, because (NN-)QLU only needs one table or neural network no matter how many assets are concerned. This may lead to a faster convergence and
better results. The training algorithm boils down to the iteration of the following steps:
QLU for optimal investment decisions
1. draw randomly patterns $t, $t+ 1 from the data set,
draw randomly an asset allocation
K:
2. for all possible actions a:
compute rf, c(Kt+b a), Q(k)($t+b K:+I)
3. compute temporal difference dt
4. compute new value Q(k+1)($t, Kn resp. Q($t, K:; w(k+1?)
5. stop, ifQ-values have converged, otherwise go to 1
Since QLU is equivalent to Q-Leaming, QLU converges to the optimal Q-values under the
same conditions as QL (e. g [2]). The main advantage of (NN- )QLU is that this algorithm
only needs one value function no matter how many assets are concerned and how fine the
grid of actions are:
Q*(($,K),a) = c(K,a) + Q*($,K').
Interestingly, the convergence to an optimal policy of QLU does not rely on an explicit
in step 1 simulates a random
exploration strategy because the randomly chosen capital
action which was responsible for the transition from K t . In combination with the randomly
chosen market state $t, a sufficient exploration of the action and state space is guaranteed.
K:
2.3 M\ldel-free policy-iteration
The refonnulation also allows the design of a policy iteration algorithm by alternating a
policy evaluation phase (PE) and a policy improvement (PI) step. Defining the temporal
differences dt for the approximation Q~I of the policy JlI in the k step ofPE
dt := r' ($t, K;, $t+d
+ ,[c(Kt+I, JlI ($t+l, K t+1 )) + Q(k) (K:+ 1 , $t+d] -
leads to the following update equation for tabulated Q-values
(k+l)($ t, K')
Q JJI
t
= Q(k)($
IJ.I
t, K") + 1/k d t?
t
Q(k)(K;, $t}
940
R. Neuneier
After convergence, one can improve the policy J-li to J-lI+l by
J-l1+I($t, Kt}
, a) + QJ.'I ($t, KD] .
= arg max[c(Kt
a
By alternating the two steps PE and PI, the sequence of policies [J-l1 (x )]1=0,... converges
under the typical assumptions to the optimal policy J-l* (x) [2] .
Note, that policy iteration is normally not possible using classical QL, if one has not an
appropriate model at hand. The introduction of the detenninistic intermediate step allows
to start with an initial strategy (e. g. given by a broker), which can be subsequently optimized by model-free policy iteration trained with historical data of the financial market.
Generalization to parameterized value functions is straightforward.
2.4 Experiments on the German Stock Index DAX
The NN-QLU algorithm is now tested on a real world task: assume that an investor wishes
to invest herihis capital into a portfolio of stocks which behaves like the German stock
index DAX. Herihis alternative is to keep the capital in the certain asset cash, referred to
as DM. We compare the resulting strategy with three benchmarks, namely Neuro-Fuzzy,
Buy&Hold and the naive prediction. The Buy&Hold strategy invests at the first time step
in the DAX and only sells at the end. The naive prediction invests if the past return of the
DAX has been positive and v. v. The third is based on a Neuro-Fuzzy model which was
optimized to predict the daily changes of the DAX [8]. The heuristic benchmark strategy is
then constructed by taking the sign of the prediction as a trading signal, such that a positive
prediction leads to an investment in stocks. The input vector of the Neuro-Fuzzy model,
which consists of the DAX itself and 11 other influencing market variables, was carefully
optimized for optimal prediction. These inputs also constitutes the $t part of the state vector
which describes the market within the NN-QLU algorithm. The data is split into a training
(from 2. Jan. 1986 to 31. Dec. 1994) and a test set (from 2. Jan. 1993 to 1. Aug. 1996). The
transaction costs (Ct) are 0.2% of the invested capital if K t is changed from DM to DAX,
which are realistic for financial institutions. Referring to an epoch as one loop over all
training patterns, the training proceeds as outlined in the previous section for 10000 epochs
with T}k = "'0 . 0.999 k with start value "'0 = 0.05.
Table 1: Comparison of the profitability of the strategies, the number of position changes
and investments in DAX for the test (training) data.
I strategy
NN-QLU
N euro-Fuzzy
Naive Prediction
Buy&Hold
profit
1.60 (3.74)
1.35 (1.98)
0.80 (1.06)
1.21 (1.46)
I investments in DAX I position changes I
70 (73)%
53 (53)%
51 (51)%
100 (100)%
30 (29)%
50 (52)%
51 (48)%
0(0)%
The strategy constructed with the NN-QLU algorithm, using a neural network with 8 hidden neurons and a linear output, clearly beats the benchmarks. The capital at the end of the
test set (training set) exceeds the second best strategy Neuro-Fuzzy by about 18.5% (89%)
(fig. 1). One reason for this success is, that QLU changes less often the position and thus,
avoids expensive transaction costs. The Neuro-Fuzzy policy changes almost every second
day whereas NN-QLU changes only every third day (see tab. 1).
It is interesting to analyze the learning behavior during training by evaluating the strategies
ofNN-QLU after each epoch. At the beginning, the policies suggest to change almost never
or each time to invest in DAX. After some thousand epochs, these bang-bang strategies
starts to differentiate. Simultaneously, the more complex the strategies become the more
profit they generate (fig. 2).
941
Enhancing Q-Leaming for Optimal Asset Allocation
the Capital
de~lopment 01
3.5
NN-QLU
2.5
09
.
o8
'. '
NaIVe PredlCllon
1 3.94
18.96
time
lime
Figure 1: Comparison of the development of the capital for the test set (left) and the training
set (right). The NN-QLU strategy clearly beats all the benchmarks.
DAX-mvestrnents In
".
o
8000
r8ILm CNGf 60 days
i
10000
2000
opoehs
4000
6000
epochs
8000
10000
Figure 2: Training course: percentage ofDAX investments (left), profitability measured as
the average return over 60 days on the training set (right).
3 Controlling the Variance of the Investment Strategies
3.1
Risk-adjusted MDPs
People are not only interested in maximizing the return, but also in controlling the risk of
their investments. This has been formalized in the Markowitz portfolio-selection, which
aims for an allocation with the maximal expected return for a given risk level [4]. Given a
stationary fo1icy f..L( x) with finite state space, the associated value function V JI. (x) and its
variance (T (V JI. ( X )) can be defined as
V"(x)
E
[t. ~'r(x"I",
E [
(t. ~'r(x"
x'+1)
xo ~ xl,
p" X'+1) -
V"(X)),
Xo
= x] .
Then, an optimal strategy f..L* (x ; ,\) for a risk-adjusted MDP (see [9], S. 410 for variancepenalized MDPs) is
f..L*(x;,\)
= argmax[VJI.(x)
JI.
,\(T2(VJI.(x))]
for'\
> O.
By variation of '\, one can construct so-called efficient portfolios which have minimal risk
for each achievable level of expected return. But in comparison to classical portfolio theory,
this approach manages multi-period portfolio management systems including transaction
costs. Furthermore, typical min-max requirements on the trading volume and other allocation constraints can be easily implemented by constraining the action space.
942
3.2
R. Neuneier
Non-linear Utility Functions
In general, it is not possible to compute (J"2 (V If. (x)) with (approximate) dynamic programming or reinforcement techniques, because (J"2 (VJ.I (x)) can not be written in a recursive
Bellman equation. One solution to this problem is the use of a return function rt, which penalizes high variance. In financial analysis, the Sharpe-ratio, which relates the mean of the
single returns to their variance i. e., r/(J"(r), is often employed to describe the smoothness
of an equity curve. For example, Moody has developed a Sharpe-ratio based error function
and combines it with a recursive training procedure [5] (see also [3]). The limitation of the
Sharpe-ratio is, that it penalizes also upside volatility. For this reason, the use of an utility
function with a negative second derivative, typical for risk averse investors, seems to be
more promising. For such return functions an additional unit increase is less valuable than
the last unit increase [4]. An example is r = log (new portfolio value I old portfolio value)
which also penalizes losses much stronger than gains. The Q-function Q(x, a) may lead to
intermediate values of a* as shown in the figure below.
e"I ---'--'--_~~~_
.0
I
~"7Jr
~J
..
'"
,
1 ' ''~
,_l
1II?'i
I
--~
.1
O.
~ - ~.- -~-~ -
"
01
\
.J
,.
"
rtMaM change 01 the pcwtIoko I4l1A ... %
---'-- 't
"
.-
,,----.-:;----0; --;. -
:i - .?-y:- -
? ? -~
% of l'N'8Sur'8n11n UncertlWl asset
4 Conclusion and Future Work
Two improvements of Q-Ieaming have been proposed to bridge the gap between classical portfolio management and asset allocation with adaptive dynamic programming. It is
planned to apply these techniques within the framework of a European Community sponsored research project in order to design a decision support system for strategic asset allocation [7). Future work includes approximations and variational methods to compute
explicitly the risk (J"2 (V If. (x)) of a policy.
References
[I J D. P. Bertsekas. Dynamic Programming and Optimal Control, vol. I. Athena Scientific, 1995.
[2] D. P. Bertsekas and J. N. Tsitsiklis. Neuro-Dynamic Programming. Athena Scientific, 1996.
[3J M. Choey and A. S. Weigend. Nonlinear trading models through Sharpe Ratio maximization.
In proc. ofNNCM'96, 1997. World Scientific.
[4J E. J. Elton and M. J. Gruber. Modern Portfolio Theory and Investment Analysis. 1995.
[5J J. Moody, L. Whu, Y. Liao, and M. Saffell. Performance Functions and Reinforcement Learning
for Trading Systems and Portfolios. Journal of Forecasting, 1998. forthcoming,
[6J R. Neuneier. Optimal asset allocation using adaptive dynamic programming. In proc. of Advances in Neural Information Processing Systems, vol. 8, 1996.
[7J R. Neuneier, H. G. Zimmermann, P. Hierve, and P. Nairn. Advanced Adaptive Asset Allocation.
EU Neuro-Demonstrator, 1997,
[8J R. Neuneier, H. G. Zimmermann, and S. Siekmann. Advanced Neuro-Fuzzy in Finance: Predicting the German Stock Index DAX, 1996. Invited presentation at ICONIP'96, Hong Kong,
availabel by email fromRalph.Neuneier@mchp.siemens.de.
[9J M. L. Puterman. Markov Decision Processes. John Wiley & Sons, 1994.
[IOJ S. P. Singh. Learning to Solve Markovian Decision Processes, CMPSCI TR 93-77, University
of Massachusetts, November 1993.
[I I J C. J. C. H. Watkins and P. Dayan. Technical Note: Q-Learning. Machine Learning: Special
Issue on Reinforcement Learning, 8,3/4:279-292, May 1992.
| 1427 |@word kong:1 achievable:1 seems:1 stronger:1 paid:1 profit:3 tr:1 initial:1 liquid:1 interestingly:1 past:1 neuneier:10 current:1 must:2 written:1 john:1 realistic:1 sponsored:1 fund:1 update:2 stationary:2 beginning:1 institution:1 preference:1 constructed:2 become:1 consists:2 combine:1 market:16 expected:6 behavior:1 multi:3 bellman:2 discounted:2 revision:1 project:1 dax:13 fuzzy:7 developed:1 ag:1 formalizes:1 certainty:1 temporal:4 every:3 xd:1 iearning:1 finance:1 control:3 normally:1 unit:2 bertsekas:2 positive:2 influencing:1 responsible:1 testing:1 investment:9 recursive:3 procedure:1 jan:2 suggest:1 selection:1 risk:12 influence:3 equivalent:1 deterministic:3 demonstrated:1 maximizing:1 go:1 straightforward:1 formalized:1 pure:1 financial:7 his:1 jli:2 variation:1 updated:1 resp:1 construction:1 controlling:2 programming:6 element:3 approximated:1 expensive:1 legally:1 solved:1 thousand:1 averse:1 eu:1 trade:1 valuable:1 environment:1 saffell:1 dynamic:8 trained:1 singh:1 easily:1 stock:8 various:1 represented:1 describe:1 artificial:2 heuristic:2 solve:2 relax:1 furthennore:1 otherwise:1 invested:2 itself:1 differentiate:1 sequence:2 advantage:1 maximal:2 adaptation:1 relevant:1 loop:1 invest:4 convergence:3 requirement:1 converges:3 volatility:1 depending:1 measured:1 ij:1 aug:1 eq:1 implemented:1 predicted:1 trading:6 stochastic:4 subsequently:1 exploration:2 explains:1 exchange:1 generalization:1 uncontrollable:3 adjusted:2 clarify:1 hold:3 considered:2 predict:1 institutional:1 proc:2 bond:1 bridge:1 reflects:1 clearly:2 always:1 aim:3 cash:1 gennan:1 she:1 improvement:2 dependent:1 dayan:1 foreign:1 nn:11 typically:1 integrated:1 cmpsci:1 her:1 hidden:1 interested:1 germany:1 ralph:2 arg:1 issue:1 development:1 special:1 construct:2 never:1 sell:1 nairn:1 constitutes:1 future:3 markowitz:1 others:1 t2:1 modern:1 randomly:4 simultaneously:2 phase:4 consisting:1 argmax:2 interest:1 possibility:2 evaluation:1 analyzed:2 extreme:1 sharpe:4 chain:1 accurate:1 kt:12 ieaming:1 tuple:1 detenninistic:2 daily:1 old:1 penalizes:3 minimal:1 modeling:1 markovian:1 planned:1 maximization:1 cost:11 strategic:1 successful:1 kn:2 combined:1 referring:1 st:3 together:1 moody:2 management:5 stochastically:1 derivative:1 return:17 li:2 account:1 de:3 includes:1 ioj:1 matter:2 explicitly:1 observing:1 tab:1 analyze:1 start:3 investor:16 complicated:1 variance:5 qk:2 yield:1 sid:1 manages:1 asset:22 executes:1 converged:1 email:1 obvious:1 dm:2 naturally:1 associated:1 con:1 boil:1 stop:1 gain:1 massachusetts:1 revise:1 carefully:1 dt:5 day:4 formulation:1 furthermore:2 profitability:2 stage:1 hand:2 nonlinear:1 llxt:2 sbk:1 scientific:3 mdp:4 alternating:2 deal:1 attractive:1 puterman:1 ll:1 during:2 hong:1 iconip:1 tn:1 ventional:1 l1:2 variational:1 behaves:1 ji:3 volume:1 jl:1 he:1 smoothness:1 rd:1 grid:1 outlined:1 portfolio:17 money:1 certain:2 success:1 approximators:1 additional:1 employed:1 maximize:1 period:2 kdt:1 signal:1 ii:1 relates:1 upside:1 corporate:1 exceeds:1 technical:1 faster:1 y:1 prediction:7 neuro:8 liao:1 enhancing:4 expectation:1 iteration:7 dec:1 whereas:2 want:1 fine:1 wealth:1 invited:1 fonnulation:1 simulates:1 incorporates:1 intermediate:4 split:1 constraining:1 relaxes:1 concerned:2 switch:2 forthcoming:1 simplifies:1 qj:1 utility:2 forecasting:1 tabulated:2 aea:1 reformulated:1 action:11 useful:1 elton:1 amount:1 reduced:1 hosen:1 generate:1 demonstrator:1 percentage:1 sign:1 discrete:2 vol:2 thereafter:1 four:1 capital:14 changing:1 sum:2 weigend:1 parameterized:1 almost:2 draw:2 decision:15 lime:1 ct:6 guaranteed:1 simplification:1 jji:1 constraint:4 generates:1 min:1 according:2 combination:1 kd:2 jr:1 describes:1 son:1 suppressed:1 vji:2 xo:2 zimmermann:2 equation:4 remains:1 describing:2 german:3 end:2 available:2 apply:1 munchen:1 appropriate:3 alternative:1 opportunity:1 classical:3 strategy:16 rt:6 enhances:1 dp:1 separate:1 athena:2 reason:2 sur:1 index:5 relationship:1 ratio:4 difficult:1 ql:6 stated:1 rise:1 negative:1 design:2 policy:20 neuron:1 markov:5 benchmark:5 finite:2 november:1 t:1 beat:2 ixo:1 situation:1 immediate:1 defining:3 community:1 pair:1 required:1 namely:1 optimized:3 maxq:1 able:1 proceeds:1 below:2 pattern:2 rf:1 max:5 wz:1 including:1 rely:1 predicting:1 advanced:2 scheme:1 improve:1 technology:1 mdps:4 naive:4 epoch:5 loss:1 interesting:1 limitation:1 allocation:23 approximator:1 aversion:1 sufficient:1 o8:1 gruber:1 pi:2 invests:3 course:1 changed:1 last:1 free:3 tsitsiklis:1 allow:1 taking:1 mchp:2 curve:1 transition:9 world:2 avoids:1 evaluating:1 reinforcement:5 adaptive:3 simplified:1 historical:2 transaction:9 approximate:2 keep:1 buy:3 search:1 continuous:1 investing:1 iterative:1 quantifies:1 table:3 promising:1 controllable:1 complex:1 european:1 vj:1 main:1 fig:4 referred:1 euro:1 wiley:1 position:3 diate:1 obeying:1 explicit:1 wish:1 xl:1 pe:2 watkins:1 third:3 down:1 xt:27 horizon:2 gap:1 presentation:1 bang:2 leaming:5 price:1 change:11 broker:1 infinite:1 typical:3 total:1 called:3 siemens:3 equity:1 people:1 support:1 tested:2 |
470 | 1,429 | Modelling Seasonality and Trends in Daily
Rainfall Data
Peter M Williams
School of Cognitive and Computing Sciences
University of Sussex
Falmer, Brighton BN1 9QH, UK.
email: peterw@cogs.susx.ac.uk
Abstract
This paper presents a new approach to the problem of modelling daily
rainfall using neural networks. We first model the conditional distributions of rainfall amounts, in such a way that the model itself determines
the order of the process, and the time-dependent shape and scale of the
conditional distributions. After integrating over particular weather patterns, we are able to extract seasonal variations and long-term trends.
1 Introduction
Analysis of rainfall data is important for many agricultural, ecological and engineering
activities. Design of irrigation and drainage systems, for instance, needs to take account
not only of mean expected rainfall, but also of rainfall volatility. In agricultural planning,
changes in the annual cycle, e.g. advances in the onset of winter rain, are significant in
determining the optimum time for planting crops. Estimates of crop yields also depend
on the distribution of rainfall during the growing season, as well as on the overall amount.
Such problems require the extrapolation of longer term trends as well as the provision of
short or medium term forecasts.
2 Occurrence and amount processes
Models of daily precipitation commonly distinguish between the occurrence process, i.e.
whether or not it rains, and the amount process, i.e. how much it rains, if it does. The
occurrence process is often modelled as a two-state Markov chain of first or higher order.
In discussion of [12], Katz traces this approach back to Quetelet in 1852. A first order
chain has been considered adequate for some weather stations, but second or higher order
models may be required for others, or at different times of year. Non-stationary Markov
chains have been used by a number of investigators, and several approaches have been taken
P. M Williams
986
to the problem of seasonal variation, e.g. using Fourier series to model daily variation of
parameters [16, 12, 15].
The amount of rain X on a given day, assuming it rains, normally has a roughly exponential
distribution. Smaller amounts of rain are generally more likely than larger amounts. Several
models have been used for the amount process. Katz & Parlange [9], for example, assume
that \IX has a normal distribution, where n is a positive integer empirically chosen to
minimise the skewness of the resulting historical distribution. But use has more commonly
been made of a gamma distribution [7,8, 12] or a mixture of two exponentials [16, 15].
3 Stochastic model
The present approach is to deal with the occurrence and amount processes jointly, by assuming that the distribution of the amount of rain on a given day is a mixture of a discrete
and continuous component. The discrete component relates to rainfall occurrence and the
continuous component relates to rainfall amount on rainy days.
We use a gamma distribution for the continuous component. l This has density proportional
to x v - 1 e- x to within an adjustable scaling of the x-axis. The shape parameter v > 0
controls the ratio of standard deviation to mean. It also determines the location of the
mode, which is strictly positive if v > 1. For certain patterns of past precipitation, larger
amounts may be more likely on the following day than smaller amounts. Specifically the
distribution of the amount X of rain on a given day is modelled by the three parameter
family
if x < 0
(1)
if x ~ 0
where 0 ~ a
~
1 and v,O > 0 and
r(v,z) = r(V)-l
1
00
yv-l e- y dy
is the incomplete gamma function. For a < 1, there is a discontinuity at x = 0 corresponding to the discrete component. Putting x = 0, it is seen that a = P(X > 0) is the
probability of rain on the day in question. The mean daily rainfall amount is avO and the
variance is aV{l + v(l - a)}02.
4 Modelling time dependency
The parameters a, v, 0 determining the conditional distribution for a given day, are understood to depend on the preceding pattern of precipitation, the time of year etc. To model
this dependency we use a neural network with inputs corresponding to the conditioning
events, and three outputs corresponding to the distributional parameters. 2 Referring to the
activations of the three output units as zO:, ZV and zO, we relate these to the distributional
parameters by
1
a=---1 + expzO:
v = expzv
0= expzo
(2)
in order to ensure an unconstrained parametrization with 0 < a < 1 and v,O > 0 for any
real values of zO:, zV, zO.
1It would be straightforward to use a mixture of gammas, or exponentials, with time-dependent
mixture components. A single gamma was chosen for simplicity to illustrate the approach.
2 A similar approach to modelling conditional distributions, by having the network output distributional parameters, is used, for example, by Ghabramani & Jordan [6], Nix & Weigend [10], Bishop
& Legleye [3], Williams [14], Baldi & Chauvin [2].
Modelling Seasonality and Trends in Daily Rainfall Data
987
On the input side, we first need to make additional assumptions about the statistical properties of the process. Specifically it is assumed that the present is stochastically independent
of the distant past in the sense that
(t > T)
(3)
for a sufficiently large number of days T. In fact the stronger assumption will be made that
P(Xt>X!Xt-1,,,,,XO) = P(Xt>X!Rt-1, ,, ,,Rt-T)
(t>T)
(4)
where R t = (X t > 0) is the event of rain on day t. This assumes that today's rainfall
amount depends stochastically only on the occurrence or non-occurrence of rain in the
recent past, and not on the actual amounts. Such a simplification is in line with previous
approaches [8, 16, 12J. For the present study T was taken to be 10.
To assist in modelling seasonal variations, cyclic variables sin T and cos T were also provided as inputs, where T = 21rt/ D and D = 365.2422 is the length of the tropical year.
This corresponds to using Fourier series to model seasonality [16, 12J but with the number of harmonics adaptively determined by the model. 3 To allow for non-periodic nonstationarity, the current value of t was also provided as input.
5 Model fitting
Suppose we are given a sequence of daily rainfall data of length N. Equation (4) implies
that the likelihood of the full data sequence (x N -1
Xo) factorises as
I ??? I
N-1
p(XN-1 , .. . I Xo; w) = p(XT-1I" . I Xo)
II p(Xt !Tt-1
I' .. I Tt-T;
w)
(5)
t=T
where the likelihood p(XT-1I'" IXO) of the initial sequence is not modelled and can be
considered as a constant (compare [14]). Our interest is in the likelihood (5) of the actual
sequence of observations, which is understood to depend on the variable weights w of the
neural network. Note that p(Xt !Tt-1
Tt-T; w) is computed by means of the neural
network outputs zf I zf I zf, using weights wand the inputs corresponding to time t.
I ' ?? I
The log likelihood of the data can therefore be written, to within a constant, as
N-1
logp(xN-1 I' ..
IXO;
w) =
L logp(xt !Tt-1,? ..
I Tt-T;
w)
t=T
or, more simply,
N-1
L(w)
=L
(6)
Lt(w)
t=T
where from (1)
L ()
{log(1 - at)
t w =
log at + (lit -1) logxt -lit logOt -logr(lIt) - xt/Ot
if Xt = 0
if Xt > 0
(7)
where dependence of at, lit Ot on w, and also on the data, is implicit.
I
To fit the model, it is useful to know the gradient 'VL(w). This can be computed using
backpropagation if we know the partial derivatives of L(w) with respect to network outputs. In view of (6) we can concentrate on a single observation and perform a summation.
3Note that both sin nr and cos nr can be expressed as non-linear functions of sin r and cos r.
which can be approximated by the network.
P. M. Williams
988
Omitting subscript references to t for simplicity, and recalling the links between network
outputs and distributional parameters given by (2), we have
8L
8z Q
8L
8z v
8L
8zo
-a
{1-0
=
{
{
if x = 0
if x > 0
0
x
v'I/J (v) - v log (j
0
x
v-()
if x = 0
ifx> 0
(8)
if x = 0
if x > 0
where
d
r'(v)
'I/J(v) = -logr(v) = - dv
r(v)
is the digamma function of v. Efficient algorithms for computing log r(v) in (7) and 'I/J(v)
in (8) can be found in Press et al. [11] and Amos [1].
6
Regularization
Since neural networks are universal approximators, some form of regularization is needed.
As in all statistical modelling, it is important to strike the right balance between jumping
to conclusions (overfitting) and refusing to learn from experience (underfitting). For this
purpose, each model was fitted using the techniques of [13] which automatically adapt the
complexity of the model to the information content of the data, though other comparable
techniques might be used. The natural interpretation of the regularizer is as a Bayesian
prior. The Bayesian analysis is completed by integration over weight space. In the present
case, this was achieved by fitting several models and taking a suitable mixture as the solution. On account of the large datasets used, however, the results are not particularly
sensitive to this aspect of the modelling process.
7
Results for conditional distributions
The process was applied to daily rainfall data from 5 stations in south east England and 5
stations in central Italy.4 The data covered approximately 40 years providing some 15,000
observations for each station. A simple fully connected network was used with a single
layer of 13 input units, 20 hidden units and 3 output units corresponding to the 3 parameters
of the conditional distribution shown in (2). As a consequence of the pruning features of the
regularizer, the models described here used an average of roughly 65 of the 343 available
parameters.
To illustrate the general nature of the results, Figure 1 shows an example from the analysis
of an early part of the Falmer series. It is worth observing the succession of 16 rainy days
from day 39 to day 54. The lefthand figure shows that the conditional probability of rain
increases rapidly at first, and then levels out after about 5-7 days.s Similar behaviour is
observed for successive dry days, for example between days 13 and 23. This suggests
that the choice of 10 time lags was sufficient. Previous studies have used mainly first
or second order Markov chains [16, 12]. Figure 1 confirms that conditional dependence
4The English stations were at Cromptons, FaImer, Kemsing, Petworth, Rothertield; the Italian
stations were at Monte Oliveto, Pisticci, Pomarico, Siena, Taverno d' Arbia.
sIn view of the number of lags used as inputs, the conditional probability would necessarily be
constant after 10 days apart from seasonal effects. In fact this is the last quarter of 1951 and the
incidence of rain is increasing here at that time of year.
Modelling Seasonality and Trends in Daily Rainfall Data
989
FALMER: conditional probability
ii
II
0.8
II
II
I
0.6
i
I
0.4
FALMER: conditional mean
20
I
II
15
II
10
5
0.2
I
II
0
0
10
20
30
40
50
60
70
10
20
30
40
50
60
70
Figure 1: Results for the 10 weeks from 18 September to 27 November, 1951. The lefthand
figure shows the conditional probability of rain for each day, with days on which rain
occurred indicated by vertical lines. The righthand figure shows the conditional expected
amount of rain in millimeters for the same period, together with the actual amount recorded.
decays rapidly at this station, at this time of year, but also indicates that it can persist for up
to at least 5 days (compare [5,4]).
8 Seasonality and trends
Conditional probabilities and expectations displayed in Figure 1 show considerable noise
since they are realisations of random variables depending on the rainfall pattern for the last
10 days. For the purpose of analysing seasonal effects and longer term trends, it is more
indicative to integrate out the noise resulting from individual weather patterns as follows.
Let R t denote the event (X t > 0) and let R t denote the complementary event (X t = 0).
The expected value of X t can then be expressed as
E(Xt )
=L
E(X t I A t - 1, . .. ,At-T) P(At- 1, . .. ,At-T)
(9)
where each event At stands for either R t or R t , and summation is over the 2T possible
combinations. Equation (9) takes the full modelled jOint distribution over the variables
X N -1, .. . ,X0 and extracts the marginal distribution for X t . This should be distinguished
from an unconditional distribution which might be estimated by pooling the data over all
40 years. E(Xt ) relates to a specific day t. Note that (9) also holds if X t is replaced by
any integrable function of X t , in particular by the indicator function of the event (X t > 0)
in which case (9) expresses the probability of rain on that day.
Examining (9) we see that the conditional expectations in the first term on the right are
known from the model, which supplies a conditional distribution not only for the sequence
of events which actually occurred, but for any possible sequence over the previous T days.
It therefore only remains to calculate the probabilities P( A t - 1, ... , At-T) of T -day sequences preceding a given day t. Note that these are again time-dependent marginal probabilities, which can be calculated recursively from
P(At , .. . , At-T+t} =
P(At I At- 1,? . . , A t -T+1 R t-T) P(At- 1 , . .? , At-T+1 R t-T)
+
P(At I At-I,?? ., At-T+IRt-T) P(At- 1, . .. , At-T+IRt-T)
provided we assume a prior distribution over the 2T initial sequences (AT-I, . .. , Ao) as
a base for the recursion. The conditional probabilities on the right are given by the model,
990
P. M. Williams
POMARICO: mean and standard deviation
POMARICO: probability of rain
0.35
0.3
.----r---..---.---~-___r-__,
HHHI"*-lHI-IHHHH1?-lH?lI-H~-f+H
0.25
M-I++H++I-tI+i1~++HHI#H-fHHHHI-{HI-1HHHHI?H-II-fHli
0.2
I\I*H*I*+H1flHl-+l!#~iHHIl-Iffi~+++H,HIf~KH+IHIH?HH1ffHII
2~~~~~~~~~~~~~-~
0.15 H+1-HrH-l1-f ,
I
0.1
1955
!
1
_ __'__
1970 1975
L--_-'--_-'--_~
1960
1965
_'__~
1980
1985
OL----'----'---~---'---'---~
1955
1960
1965
1970
1975
1980
1985
Figure 2: Integrated results for Pomarico from 1955-1985. The lefthand figure shows the
daily probability of rain, indicating seasonal variation from a summer minimum to a winter
maximum. The righthand figure shows the daily mean (above) and standard deviation
(below) of rainfall amount in millimeters.
as before, and the unconditional probabilities are given by the recursion. It turns out that
results are insensitive to the choice of initial distribution after about 50 iterations, verifying
that the occurrence process, as modelled here, is in fact ergodic.
9 Integrated results
Results for the integrated distribution at one of the Italian stations are shown in Figure 2.
By integrating out the random shocks we are left with a smooth representation of time
dependency alone. The annual cycles are clear. Trends are also evident over the 30 year
period. The mean rainfall amount is decreasing significantly, although the probability of
rain on a given day of the year remains much the same. Rain is occurring no less frequently,
but it is occurring in smaller amounts. Note also that the winter rainfall (the upper envelope
of the mean) is decreasing more rapidly than the summer rainfall (the lower envelope of
the mean) so that the difference between the two is narrowing.
10 Conclusions
This paper provides a new example of time series modelling using neural networks. The
use of a mixture of a discrete distribution and a gamma distribution emphasises the general
principle that the "error function" for a neural network depends on the particular statistical
model used for the target data. The use of cyclic variables sin T and cos T as inputs shows
how the problem of selecting the number of harmonics required for a Fourier series analysis
of seasonality can be solved adaptively. Long term trends can also be modelled by the use
of a linear time variable, although both this and the last feature require the presence of
a suitable regularizer to avoid overfitting. Lastly we have seen how a suitable form of
integration can be used to extract the underlying cycles and trends from noisy data. These
techniques can be adapted to the analysis of time series drawn from other domains.
Modelling Seasonality and Trends in Daily Rainfall Data
991
Acknowledgement
I am indebted to Professor Helen Rendell of the School of Chemistry, Physics and Environmental Sciences, University of Sussex, for kindly supplying the rainfall data and for
valuable discussions.
References
[1] D. E. Amos. A portable fortran subroutine for derivatives of the psi function. ACM
Transactions on Mathematical Software, 9:49~502, 1983.
[2] P. Baldi and Y. Chauvin. Hybrid modeling, HMM/NN architectures, and protein
applications. Neural Computation, 8:1541-1565, 1996.
[3] C. M. Bishop and C. Legleye. Estimating conditional probability densities for periodic variables. In G. Tesauro, D. Touretzky, and T. Leen, editors, Advances in Neural
Information Processing Systems 7, pages 641-648. The MIT Press, 1995.
[4] E. H. Chin. Modelling daily precipitation occurrence process with Markov chain.
Water Resources Research, 13:949-956,1977.
[5] P. Gates and H. Tong. On Markov chain modelling to some weather data. Journal of
AppliedMeteorology, 15:1145-1151, 1976.
[6] Z. Ghahramani and M. 1. Jordan. Supervised learning from incomplete data via an
EM approach. In Jack D. Cowan, Gerald Tesauro, and Joshua Alspector, editors, Advances in Neural Information Processing Systems 6, pages 120-127. Morgan Kaufmann, 1994.
[7] N. T. Ison, A. M. Feyerherm, andL. D. Bark. Wet period precipitation and the gamma
distribution. Journal of Applied Meteorology, 10:658-665, 1971.
[8] R. W. Katz. Precipitation as a chain-dependent process. Journal of Applied Meteorology, 16:671-676,1977.
[9] R. W. Katz and M. B. Parlange. Effects of an index of atmospheric circulation on
stochastic properties of precipitation. Water Resources Research, 29:2335-2344,
1993.
[10] D. A. Nix and A. S. Weigend. Learning local error bars for nonlinear regression. In
Gerald Tesauro, David S. Touretzky, and Todd K. Leen, editors, Advances in Neural
Information Processing Systems 7, pages 489-496. MIT Press, 1995.
[11] W. H. Press, B. P. Flannery, S. A. Teukolsky, and W. T. Vetterling. Numerical Recipes
in C. Cambridge University Press, 2nd edition, 1992.
[12] R. D. Stern and R. Coe. A model fitting analysis of daily rainfall data, with discussion.
Journal of the Royal Statistical Society A, 147(Part 1):1-34,1984.
[13] P. M. Williams. Bayesian regularization and pruning using a Laplace prior. Neural
Computation, 7:117-143,1995.
[14] P. M. Williams. Using neural networks to model conditional multivariate densities.
Neural Computation, 8:843-854, 1996.
[15] D. A. Woolhiser. Modelling daily precipitation-progress and problems. In Andrew T. Walden and Peter Guttorp, editors, Statistics in the Environmental and Earth
Sciences, chapter 5, pages 71-89. Edward Arnold, 1992.
[16] D. A. Woolhiser and G. G. S. Pegram. Maximum likelihood estimation of Fourier coefficients to describe seasonal variation of parameters in stochastic daily precipitation
models. Journal of Applied Meteorology, 18:34-42, 1979.
| 1429 |@word stronger:1 nd:1 confirms:1 recursively:1 initial:3 cyclic:2 series:6 selecting:1 past:3 current:1 legleye:2 incidence:1 activation:1 written:1 numerical:1 distant:1 shape:2 stationary:1 alone:1 indicative:1 parametrization:1 short:1 ifx:1 supplying:1 provides:1 location:1 successive:1 mathematical:1 supply:1 fitting:3 baldi:2 underfitting:1 x0:1 expected:3 alspector:1 roughly:2 planning:1 growing:1 frequently:1 ol:1 decreasing:2 automatically:1 actual:3 agricultural:2 precipitation:9 provided:3 increasing:1 underlying:1 estimating:1 medium:1 skewness:1 ti:1 rainfall:23 uk:2 control:1 normally:1 unit:4 positive:2 before:1 engineering:1 understood:2 local:1 todd:1 consequence:1 subscript:1 approximately:1 might:2 emphasis:1 suggests:1 co:4 irrigation:1 backpropagation:1 universal:1 significantly:1 weather:4 integrating:2 protein:1 helen:1 williams:7 straightforward:1 ergodic:1 simplicity:2 variation:6 laplace:1 qh:1 today:1 suppose:1 target:1 trend:11 approximated:1 particularly:1 hrh:1 distributional:4 persist:1 observed:1 narrowing:1 solved:1 verifying:1 calculate:1 cycle:3 connected:1 valuable:1 complexity:1 gerald:2 depend:3 joint:1 chapter:1 regularizer:3 zo:5 describe:1 monte:1 lag:2 larger:2 statistic:1 jointly:1 itself:1 noisy:1 sequence:8 rapidly:3 lefthand:3 kh:1 recipe:1 optimum:1 volatility:1 illustrate:2 depending:1 ac:1 andrew:1 school:2 progress:1 bn1:1 edward:1 implies:1 concentrate:1 stochastic:3 require:2 behaviour:1 ao:1 summation:2 strictly:1 hold:1 sufficiently:1 considered:2 normal:1 week:1 early:1 purpose:2 earth:1 estimation:1 wet:1 sensitive:1 amos:2 mit:2 avoid:1 season:1 seasonal:7 modelling:14 likelihood:5 mainly:1 indicates:1 digamma:1 sense:1 am:1 dependent:4 nn:1 vl:1 integrated:3 vetterling:1 hidden:1 italian:2 subroutine:1 i1:1 overall:1 integration:2 marginal:2 having:1 lit:4 others:1 realisation:1 rainy:2 winter:3 gamma:7 individual:1 replaced:1 recalling:1 interest:1 righthand:2 mixture:6 unconditional:2 chain:7 partial:1 daily:16 experience:1 lh:1 jumping:1 incomplete:2 fitted:1 instance:1 modeling:1 ison:1 andl:1 logp:2 deviation:3 examining:1 dependency:3 periodic:2 referring:1 adaptively:2 density:3 physic:1 together:1 again:1 central:1 recorded:1 cognitive:1 stochastically:2 derivative:2 logr:2 lhi:1 li:1 account:2 chemistry:1 coefficient:1 onset:1 depends:2 view:2 extrapolation:1 observing:1 yv:1 circulation:1 variance:1 kaufmann:1 succession:1 yield:1 dry:1 millimeter:2 modelled:6 bayesian:3 refusing:1 worth:1 indebted:1 touretzky:2 nonstationarity:1 email:1 psi:1 provision:1 actually:1 back:1 higher:2 day:26 supervised:1 leen:2 though:1 implicit:1 lastly:1 tropical:1 nonlinear:1 mode:1 indicated:1 omitting:1 effect:3 avo:1 regularization:3 deal:1 sin:5 during:1 sussex:2 brighton:1 chin:1 evident:1 tt:6 l1:1 rendell:1 harmonic:2 jack:1 planting:1 quarter:1 empirically:1 conditioning:1 insensitive:1 interpretation:1 occurred:2 katz:4 significant:1 cambridge:1 unconstrained:1 longer:2 etc:1 base:1 multivariate:1 recent:1 italy:1 apart:1 tesauro:3 certain:1 ecological:1 approximators:1 joshua:1 integrable:1 seen:2 morgan:1 additional:1 minimum:1 preceding:2 strike:1 period:3 ii:9 relates:3 full:2 smooth:1 adapt:1 england:1 long:2 crop:2 regression:1 expectation:2 iteration:1 achieved:1 ot:2 envelope:2 south:1 pooling:1 cowan:1 jordan:2 integer:1 presence:1 fit:1 architecture:1 minimise:1 whether:1 assist:1 peter:2 adequate:1 generally:1 useful:1 covered:1 clear:1 amount:22 meteorology:3 seasonality:7 estimated:1 discrete:4 express:1 zv:2 putting:1 drawn:1 falmer:4 shock:1 year:9 weigend:2 wand:1 family:1 dy:1 scaling:1 comparable:1 layer:1 hi:1 summer:2 simplification:1 distinguish:1 hhi:1 annual:2 activity:1 adapted:1 software:1 fourier:4 aspect:1 combination:1 smaller:3 em:1 dv:1 xo:4 taken:2 equation:2 resource:2 remains:2 turn:1 needed:1 know:2 fortran:1 available:1 occurrence:9 distinguished:1 gate:1 assumes:1 rain:21 ensure:1 completed:1 coe:1 ghahramani:1 society:1 question:1 rt:3 dependence:2 nr:2 september:1 gradient:1 link:1 hmm:1 portable:1 chauvin:2 water:2 assuming:2 length:2 index:1 ratio:1 balance:1 providing:1 relate:1 trace:1 irt:2 design:1 stern:1 adjustable:1 perform:1 upper:1 zf:3 av:1 observation:3 vertical:1 markov:5 nix:2 datasets:1 november:1 displayed:1 ixo:2 station:8 atmospheric:1 david:1 required:2 discontinuity:1 able:1 bar:1 below:1 pattern:5 royal:1 event:7 suitable:3 natural:1 hybrid:1 indicator:1 recursion:2 factorises:1 axis:1 extract:3 siena:1 prior:3 acknowledgement:1 bark:1 determining:2 hif:1 fully:1 proportional:1 integrate:1 sufficient:1 principle:1 editor:4 last:3 english:1 side:1 allow:1 arnold:1 taking:1 calculated:1 xn:2 stand:1 commonly:2 made:2 historical:1 transaction:1 pruning:2 overfitting:2 assumed:1 continuous:3 guttorp:1 learn:1 nature:1 necessarily:1 domain:1 kindly:1 noise:2 edition:1 complementary:1 tong:1 exponential:3 ix:1 cog:1 bishop:2 xt:13 specific:1 decay:1 occurring:2 forecast:1 flannery:1 lt:1 simply:1 likely:2 expressed:2 corresponds:1 determines:2 environmental:2 acm:1 teukolsky:1 conditional:19 professor:1 content:1 change:1 considerable:1 analysing:1 specifically:2 determined:1 east:1 indicating:1 investigator:1 |
471 | 143 | 240
TEMPORAL REPRESENTATIONS IN A
CONNECTIONIST SPEECH SYSTEM
Erich J. Smythe
207 Greenmanville Ave, #6
Mystic, CT 06355
ABSTRACT
SYREN is a connectionist model that uses temporal information
in a speech signal for syllable recognition. It classifies the rates
and directions of formant center transitions, and uses an adaptive
method to associate transition events with each syllable. The
system uses explicit spatial temporal representations through delay lines. SYREN uses implicit parametric temporal representations in formant transition classification through node activation
onset, decay, and transition delays in sub-networks analogous to
visual motion detector cells. SYREN recognizes 79% of six repetitions of 24 consonant-vowel syllables when tested on unseen
data, and recognizes 100% of its training syllables.
INTRODUCTION
Living organisms exist in a dynamic environment. Problem solving systems, both
natural and synthetic, must relate and interpret events that occur over time.
Although connectionist models are based on metaphors from the brain, few have
been designed to capture temporal and sequential information common to even
the most primitive nervous systems. Yet some of the most popular areas of application of these models, including speech recognition, vision, and motor control,
require some form of temporal processing.
The variation in a speech signal contains considerable information. Changes in
format location or other acoustic parameters (Delattre, et al., 1955; Pols and
Schouten, 1982) can determine the identity of constituents of speech, even when
segmentation information is obscure. Speech recognition systems have shown
good results when they incorporate some temporal information (Waible, et al.,
1988, Anderson, et al., 1988). Successful speech systems must incorporate temporal processing.
Natural organisms have sensory organs that are continuously updated and can do
only limited buffering of input stimuli. Synthetic implementations can buffer their
input, transforming time into space. Often the size and complexity of the input
representations place limits on the amount of input that can be buffered, especially when data is coming from hundreds or thousands of sensors, and other
methods must be found to integrate temporal information.
Temporal Representations in a Connectionist Speech System
This paper describes SYREN (SYllable REcognition Network), a connectionist
network that incorporates various temporal representations for consonant-vowel
(CV) syllable recognition by the classification of formant center transitions. Input
is presented sequentially, one time slice at a time. The network is described,
including the temporal processing used in formant transition classification, learning, and syllable recognition. The results of syllable recognition experiments are
discussed in the final section.
TEMPORAL REPRESENTATIONS
Various types of temporal representations may be used to incorporate time in
connectionist models. They range from explicit spatial representations where
time is convened into space, to implicit parametric representations where time is
incorporated using network computational parameters. Spatiotemporal representations are a middle ground combining the two extremes. The categories represent a continuum rather than absolute distinctions. Several of these types are
found in SYREN.
EXPLICIT SPATIAL REPRESENTATIONS
In a purely spatial representation temporal information is preserved by spreading
time steps over space through the network topology. These representations include input buffers, delay lines, and recurrent networks.
Fixed input buffers allow interaction between time slices of input. Parts of the
network are copied to represent states at panicular time slices. Other methods
use sliding input buffers in the form of a queue. Tapped delay lines and delay
filters are means of spreading network node activations over time. Composed of
chains of network nodes or delay functions, they can preserve the sequential
structure of information. A value on a connection from a delay line represents
events that have occurred in the past. Delay lines and filters have been used in
speech recognition systems by Waible, et al. (1988), and Tank and Hopfield
(1987) .
Recurrent networks are similar to delay lines in that information is preserved by
propagating activation through the network. They can store information indefinitely or generate potentially infinite sequences of behaviors through feedback
cycles, whereas delay lines without cycles are limited by their fixed length. Recurrent networks pose problems for learning, although researchers are working on
recurrent back propagation networks (Jordan, 1988).
Spatial representations are good for explicitly preserving sequences of events, and
can simplify the learning of temporal patterns. Resource constraints place a limit
on the size of fixed length buffers and delay lines, however. Input data from
thousands of sensors place limits on the length of time represented in the buffer,
and may not be able to retain information long enough to be of use. Fixed input
buffers may introduce edge effects. Interaction is lost between the edges of the
buffer and data from preceding and succeeding buffers unless the input is properly segmented. Long delay lines may be computationally expensive as well.
241
242
Smythe
SPATIOTEMPORAL REPRESENTATIONS
Implicit parametric methods represent time in connectionist models by the behavior of network nodes . State information stored in individual nodes allows more
complex activation functions and the accumulation of statistical information.
This method may be used to regulate the flow of activation in the network, provide a trace of previous activation, and learn from data separated in time.
Adjusting the parameters of functions such as the interactive activation equation
of McClelland and Rumelhart (1982) can control the strength of input, affecting
the rate that activation reaches saturation. This leads to pulse trains used in
synchronization. Variations in decay parameters control the duration of an activation trace.
State and statistical information is useful in learning. Eligibility traces from classical conditioning models provide decaying memory of past connection activation.
Temporally weighted averages may be used for weight computations.
Spatiotemporal representations combine implicit parametric representations with
explicit spatial representations. These include the regulation of propagation time
and pulse trains through parameter adjustment. Gating behavior that controls the
flow of activation through a network is another spatiotemporal method.
SYREN DESCRIPTION
SYREN is a connectionist model that incorporates temporal processing in isolated
syllable recognition using formant center transitions. Formant center tracts are
presented in 5 ms time slices. Input nodes are updated once per time slice. The
network classifies the rates and directions of formant transitions. Transition data
are used by an adaptive network to associate transition patterns with syllables. A
recognition network uses output of the adaptive network to identify a syllable.
Complete details of the system maybe found in Smythe (1988).
DATA CORPUS
Input data consist of formant centers from five repetitions of twenty-four consonant-vowel syllables (the stop consonants Ib, d, gl paired with the vowels Iii, ey,
ih, eh, ae, ah, ou, uu/), and an averaged set of each of the five repetitions from
work performed by Kewley Port (1982). Each repetition is presented as a binary
matrix with a row representing frequency in 20 Hz units, and a column representing time in 5 ms slices. The matrix is given to the input units one column at a
time. A '1' in a cell of a matrix represents a formant center at a particular
frequency during a particular time slice.
FORMANT TRANSITION CLASSIFICATION
In the first stage of processing SYREN determines the rate and direction of formant center transitions. Formant transition detectors are subnetworks designed
to respond to transitions of one of six rates in either rising or falling directions,
Temporal Representations in a Connectionist Speech System
and also to steady-state events. The method used is motivated by a mechanism
for visual motion detection in the retina that combines interactions between subunits of a dendritic tree and shunting, veto inhibition (Koch et ai, 1982). Formant motion is analogous to visual motion, and formant transitions are treated as
a one dimensional case of visual motion.
Preferred
Transition
I Branch Nodes
'
DIsta
Proximal
Figure 1. Formant transition detector subnetwork and its preferred
descending transition type. The vertical axis is frequency (one row for
each input unit) and the horizontal axis is time in 5 ms slices.
A detector subnetwork for a slow transition is shown in figure 1, along with its
preferred transition. Branch nodes are analogous to dendritic subunits, and serve
as activation transmission lines. Their activation is computed by the equation:
af+l
= af(1- 0) + netf(l -
aD
Where a is the activation of unit i at time t, net is the weighted input, t is an
update cycle (there are 7 updates per time slice), and e is a decay constant.
Input to a branch node drives the activation to a maximum value, the rate of
which is determined by the strength of the input, In the absence of input the
activation decays to O.
For the preferred direction, input nodes are activated for two time slices (10 ms)
in order from top to bottom. An input node causes the activation of the most
distal branch node to rise to a maximum value. This in turn causes the next node
to activate, slightly delayed with respect to the first, and so on for the rest of the
branch. This results in a pulse of activation flowing along the branch with a
transmission delay of roughly one time slice (7 update cycles) from the distal to
the proximal end. The most proximal branch node also has a connection to the
input node. This connection serves to prime the node for slower transitions.
Activation from an input node that lasts for only one time slice will decay in the
proximal branch node before the activation from the distal region arrives. If
input is present for two time steps the extra activation from the input connection
primes the node, quickly driving it to a maximal value when the distal activation
arrives.
243
244
Smythe
An S-node provides the output of the detector. It computes a sigmoid squash
function and fires (a sudden increase in activation) when sufficient activation is in
the proximal branch nodes. For this particular detector, if the transition is too
fast (i. e. one time step for each input unit) the proximal nodes will not attain a
high enough activation; if the transition is too slow (i.e. three time steps for each
input unit) activation on proximal branch nodes from earlier time steps will have
decayed before the transition is complete. This architecture is tuned to a slower
transition by increasing the transmission time on the branches by varying the connection weights, and by reducing the decay rate by lowering the decay constant.
This illustrates the use of parametric manipulations to control temporal behavior
in for rate sensitivity.
Veto inhibition is used in this detector for direction selectivity. Veto nodes provide inhibition and are activated by input nodes, and use the interactive activation
equation for a decaying memory. Had the transition in figure 1. been in the
opposite direction, activation from previous time slices on a veto connection
would prevent the input node from activating its distal branch node, preventing
the flow of activation and the firing of the S-node. Here a veto connection acts
as a gate, serving to select input for processing.
Detectors are constructed for faster transitions by shortening the transmission
lines and by using veto connections for rate sensitivity. A transition detector for a
faster transition is shown in figure 2. Here the receptive field is larger, and veto
connections are used to select transitions that skip one input unit at each time
slice. Veto connections are still used for direction selectivity. Detectors for even
faster transitions are created by widening the receptive field and increasing the
number of veto connections for rate sensitivity.
Detectors are designed to respond to a specific transition type and not to respond
to the transitions of other detectors. They will respond to transitions with rates
between their own and the next type of detector. For slower transitions the firing
of two detectors indicates an intermediate rate. For faster transitions special
detectors are designed to fire for only one precise rate by eliminating some of the
branches. Different firing patterns of precise and more general detectors distinguish rates. This gives a very fine rate sensitivity throughout the range of transitions.
Detector networks are copied to span the entire frequency range with overlapping
receptive fields. This yields an array of S-nodes for each transition type. giving
excellent spatial resolution of the frequency range. There are 200 S-nodes for
each detector type. each signaling a transition that starts and ends at a particular
frequency unit.
ADAPTIVE NE1WORK
The adaptive network learns to associate patterns of formant transitions with specific syllables. To do this it must be able to store at least part of the unfolding
patterns or else it is forced to respond to information from only one time slice.
Temporal Representations in a Connectionist Speech System
Preferred
Transition
Veto
Nodes
Branch Nodes
Figure 2. Formant transition detector subnetwork for a faster transition. Only the veto connections used for rate sensitivity are shown.
The learning algorithm must also deal with past activation histories of connections
or else it can only learn from one time slice. The network accomplishes this
through tapped delay lines and decaying eligibility traces.
There are twenty-four nodes in the adaptive network, each assigned to one syllable. It is a single layer network, trained using a hybrid supervised learning algorithm that merges Widrow-Hoff type learning with a classical conditioning model
(Sutton and Barto, 1987).
Storage of temporal patterns
Tapped delay lines are used to briefly store sequences of formant transition patterns. S-nodes from each transition detector are connected to a tapped delay
line of five nodes. Each delay node simply passes on its S-node's activation value
once per 5 ms time slice, allowing the delay matrix to store 25 ms (five time
slices) of transition patterns.
The delay matrix consists of delay lines for each transition detector at each receptive field. Adaptive nodes are connected to every node in the delay matrix. The
delay lines do not perform input buffering; information in the delay matrix has
been subject to one level of processing. The amount of information stored (the
length of the delay line) is limited by efficiency considerations.
Adaptive Algorithm
Nodes in the adaptive network compute their activation using a sigmoid squash
function and adjust their weights according to the equation:
W~"!"l
IJ
=
w~? + a(z~ - s~)e~
IJ
I
I
J
where w is the weight from a connection from node j to node i at time t. a is a
learning constant. z is the expected value of node i, s is the weighted sum of the
connections of node i. and e is the exponentially decaying canonical eligibility of
245
246
Smythe
connection j. The eligibility constant gives some variation in the exact timing of
transition patterns, allowing limited time warping between training and testing.
FINAL RECOGNITION NETWORK
The adaptive network is not perfect and results in a number of false alarm errors.
Many of these are eliminated by using firing patterns of other adaptive nodes.
For example, a node that consistently misfires on one syllable could be blocked
by the firing of the correct node for that syllable. Adaptive nodes are connected
to a veto recognition network. Since an adaptive node may fire at any time (and
at different times) throughout input presentation, delay lines are used to preserve
patterns of adaptive node behavior, and veto inhibition is used to block false
alarms. Connections in the veto network are enabled or disabled after training.
Clearly this is an ad hoc solution, but it suggests the use of representations that
are distributed both spatially and temporally.
RESULTS AND DISCUSSION
In each experiment syllable repetitions were divided into mutually exclusive training and testing sets. A training cycle consisted of one presentation of each member of the training set. In both experiments the networks were trained until
adequate performance was achieved, usually after four to ten training cycles.
In the first experiment the network was trained on the five raw repetitions and
tested on the averaged set. It achieved 92% recognition on the testing set and
100% recognition on the training set. The network had two miss errors on the
training set.
In the second experiment, the network was trained on four of the raw repetitions
and tested on the fifth. Five separate training runs were performed to test the
network on each repetition. The network achieved 76% recognition on the testing set for all training runs, and 100% recognition on the training set.
In all experiments most of the adaptive nodes responded when there was transition information in the delay matrix. Many responded when both transition and
steady-state information was present, using clues from both the consonant and
the vowel. This situation occurs only briefly for each formant, since the delay
matrix holds information for 5 time slices, and it takes four time slices to signal a
steady-state event. Transition information will be at the end of the delay matrix
while steady-state is at the beginning. Many nodes were strongly inhibited in the
absence of transition information even for their correct syllable, although they
had fired earlier in the data presentation.
CONCLUSIONS
We have shown how different temporal representations and processing methods
are used in a connectionist model for syllable recognition. Hybrid connectionist
architectures with only slightly more elaborate processing methods can classify
acoustic motion and associate sequences of transition events with syllables. The
Temporal Representations in a Connectionist Speech System
system is not designed as a general speech recognition system, especially since the
accurate measurement of formant center frequencies is impractical. Other signal
processing techniques, such as spectral peak estimation, can be used without
changes in the architecture. This could provide information to a larger speech
recognition system .
SYREN was influenced by a neurophysiological model for visual motion detection, and shows how knowledge from one processing modality is applied to other
problems. The merging of ideas from real nervous systems with existing techniques can add to the connectionist tool kit, resulting in more powerful processing
systems.
Acknowledgments
This research was performed at Indiana University Computer Science Department as part of the author's Ph.D. thesis. The author would like to thank committee members John Barnden and Robert Port for their help and direction, and
Donald Lee and Peter Brodeur for their assistance in preparing the manuscript.
References
Delattre, P. C., Liberman, A. M., Cooper. F. S., 1955, "Acoustic loci and transitional cues for stop consonants," J. Acous. Soc. Am .? 27. 769-773.
Jordan, M . I., 1986 "Serial order: A parallel distributed processing approach,"
ICS Report 8604, UCSD, San Diego.
Kewley-Port, D., 1982, "Measurement of formant transitions in naturally produced consonant-vowel syllables," J. Acous. Soc. Am, 72, 379-389.
Koch, C., Poggio, T., Torre, V., 1982, "Retinal ganglion cells: A functional
interpretation of dendritic morphology," Phil. Trans. R. Soc. Lon.: Series B,
298, 227-264.
McClelland, J. L.. Rumelhart, D. E., 1982, "An interactive activation model of
context effects in letter perception," Psychological Review, 88, 375-407.
Pols, L. C. W., Schouten, M. F. H., 1982, "Perceptual relevance of coarticulation, " in: Carlson, R., and Grandstrom, B., The Representation of Speech in
the Peripheral Auditory System, Elsevier, 203-208.
Anderson, S., Merrill, J.W.L, Port, R., 1988, "Dynamic speech characterization
with recurrent networks," Indiana University Dept. of Computer Science TR.
No. 258, Bloomington, In.
Smythe, E. J., 1988, "Temporal computation in connectionist models," Indiana
University Dept. of Computer Science TR. No. 251, Bloomington, In.
Sutton, R. S., Barto, A. G., 1987, "A temporal difference model of classical
conditioning," GTE TR87-509.2.
Tank, D. W., Hopfield, J. J., 1987, "Concentrating information in time," Proceedings of the IEEE Conference on Neural Networks, San Diego, IV-455-468.
Waible, A., Hanazawa, T., Hinton, G., Shikana, K, Lang, K, 1988, "Phoneme
recognition: Neural networks vs. Hidden Markov Models," Proc. Int. Conf.
Acoustics, Speech, and Signal Processing, 107-110.
247
| 143 |@word merrill:1 briefly:2 middle:1 rising:1 eliminating:1 pulse:3 tr:2 ne1work:1 contains:1 series:1 tuned:1 past:3 existing:1 activation:33 yet:1 lang:1 must:5 john:1 motor:1 designed:5 succeeding:1 update:3 v:1 cue:1 nervous:2 beginning:1 indefinitely:1 sudden:1 provides:1 characterization:1 node:52 location:1 five:6 along:2 constructed:1 consists:1 combine:2 introduce:1 expected:1 roughly:1 behavior:5 morphology:1 brain:1 metaphor:1 increasing:2 classifies:2 indiana:3 impractical:1 temporal:25 every:1 act:1 interactive:3 control:5 unit:8 before:2 timing:1 limit:3 sutton:2 firing:5 suggests:1 limited:4 range:4 averaged:2 acknowledgment:1 testing:4 lost:1 block:1 signaling:1 area:1 attain:1 donald:1 storage:1 context:1 descending:1 accumulation:1 center:8 phil:1 primitive:1 duration:1 resolution:1 array:1 enabled:1 variation:3 analogous:3 updated:2 diego:2 exact:1 us:5 tapped:4 associate:4 rumelhart:2 recognition:20 expensive:1 bottom:1 capture:1 thousand:2 region:1 cycle:6 connected:3 transitional:1 environment:1 transforming:1 complexity:1 pol:2 dynamic:2 trained:4 solving:1 purely:1 serve:1 efficiency:1 hopfield:2 various:2 represented:1 train:2 separated:1 forced:1 fast:1 activate:1 dista:1 larger:2 squash:2 formant:21 unseen:1 hanazawa:1 final:2 hoc:1 sequence:4 net:1 interaction:3 coming:1 maximal:1 combining:1 fired:1 description:1 constituent:1 transmission:4 tract:1 perfect:1 help:1 recurrent:5 widrow:1 propagating:1 pose:1 ij:2 soc:3 skip:1 uu:1 direction:9 coarticulation:1 correct:2 torre:1 filter:2 require:1 activating:1 dendritic:3 hold:1 koch:2 ground:1 ic:1 driving:1 continuum:1 estimation:1 proc:1 spreading:2 organ:1 repetition:8 tool:1 weighted:3 unfolding:1 clearly:1 sensor:2 rather:1 varying:1 barto:2 lon:1 properly:1 consistently:1 indicates:1 ave:1 am:2 elsevier:1 entire:1 hidden:1 tank:2 classification:4 spatial:7 special:1 hoff:1 field:4 once:2 eliminated:1 preparing:1 buffering:2 represents:2 connectionist:15 stimulus:1 simplify:1 inhibited:1 few:1 retina:1 report:1 composed:1 preserve:2 individual:1 delayed:1 fire:3 vowel:6 detection:2 adjust:1 extreme:1 arrives:2 activated:2 misfire:1 chain:1 kewley:2 accurate:1 edge:2 poggio:1 unless:1 tree:1 iv:1 isolated:1 psychological:1 column:2 earlier:2 classify:1 hundred:1 delay:28 successful:1 too:2 stored:2 spatiotemporal:4 proximal:7 synthetic:2 decayed:1 peak:1 sensitivity:5 retain:1 lee:1 continuously:1 quickly:1 thesis:1 conf:1 retinal:1 int:1 explicitly:1 onset:1 ad:2 performed:3 start:1 decaying:4 barnden:1 parallel:1 responded:2 phoneme:1 yield:1 identify:1 raw:2 produced:1 researcher:1 drive:1 ah:1 delattre:2 history:1 detector:21 reach:1 influenced:1 frequency:7 naturally:1 stop:2 auditory:1 bloomington:2 adjusting:1 popular:1 concentrating:1 knowledge:1 segmentation:1 ou:1 back:1 manuscript:1 supervised:1 flowing:1 strongly:1 anderson:2 implicit:4 stage:1 until:1 working:1 horizontal:1 overlapping:1 propagation:2 disabled:1 effect:2 consisted:1 assigned:1 spatially:1 deal:1 distal:5 assistance:1 during:1 eligibility:4 steady:4 m:6 complete:2 motion:7 consideration:1 common:1 sigmoid:2 functional:1 conditioning:3 exponentially:1 discussed:1 organism:2 occurred:1 interpretation:1 interpret:1 buffered:1 blocked:1 measurement:2 ai:1 cv:1 erich:1 had:3 inhibition:4 add:1 own:1 prime:2 manipulation:1 store:4 buffer:9 selectivity:2 binary:1 preserving:1 preceding:1 kit:1 ey:1 accomplishes:1 determine:1 signal:5 living:1 sliding:1 branch:14 segmented:1 faster:5 af:2 long:2 divided:1 serial:1 shunting:1 tr87:1 paired:1 ae:1 vision:1 represent:3 achieved:3 cell:3 preserved:2 whereas:1 affecting:1 fine:1 else:2 modality:1 extra:1 rest:1 pass:1 hz:1 subject:1 veto:14 member:2 incorporates:2 flow:3 jordan:2 intermediate:1 iii:1 enough:2 architecture:3 topology:1 opposite:1 idea:1 six:2 motivated:1 queue:1 peter:1 speech:17 cause:2 adequate:1 useful:1 maybe:1 amount:2 shortening:1 ten:1 ph:1 mcclelland:2 category:1 generate:1 exist:1 canonical:1 per:3 serving:1 four:5 falling:1 prevent:1 lowering:1 sum:1 run:2 letter:1 powerful:1 respond:5 place:3 throughout:2 layer:1 ct:1 syllable:21 distinguish:1 copied:2 strength:2 occur:1 constraint:1 span:1 format:1 department:1 according:1 peripheral:1 describes:1 slightly:2 computationally:1 resource:1 equation:4 mutually:1 turn:1 mechanism:1 committee:1 locus:1 end:3 subnetworks:1 serf:1 regulate:1 spectral:1 slower:3 gate:1 top:1 include:2 recognizes:2 carlson:1 giving:1 especially:2 classical:3 warping:1 occurs:1 parametric:5 receptive:4 exclusive:1 subnetwork:3 separate:1 thank:1 length:4 regulation:1 robert:1 potentially:1 relate:1 trace:4 rise:1 implementation:1 twenty:2 perform:1 allowing:2 vertical:1 markov:1 subunit:2 situation:1 incorporated:1 precise:2 hinton:1 ucsd:1 connection:18 acoustic:4 distinction:1 merges:1 trans:1 able:2 usually:1 pattern:11 perception:1 saturation:1 including:2 memory:2 event:7 natural:2 eh:1 treated:1 widening:1 hybrid:2 representing:2 temporally:2 axis:2 created:1 review:1 synchronization:1 integrate:1 sufficient:1 port:4 obscure:1 row:2 gl:1 last:1 schouten:2 allow:1 absolute:1 fifth:1 distributed:2 slice:20 feedback:1 transition:54 computes:1 sensory:1 preventing:1 author:2 adaptive:15 clue:1 acous:2 san:2 liberman:1 preferred:5 sequentially:1 corpus:1 consonant:7 learn:2 excellent:1 complex:1 alarm:2 elaborate:1 slow:2 cooper:1 sub:1 explicit:4 perceptual:1 ib:1 learns:1 specific:2 gating:1 decay:7 consist:1 ih:1 false:2 sequential:2 merging:1 illustrates:1 simply:1 ganglion:1 neurophysiological:1 visual:5 adjustment:1 determines:1 identity:1 presentation:3 absence:2 considerable:1 change:2 infinite:1 determined:1 reducing:1 miss:1 gte:1 select:2 relevance:1 incorporate:3 dept:2 tested:3 |
472 | 1,430 | Perturbative M-Sequences for Auditory
Systems Identification
Mark Kvale and Christoph E. Schreiner?
Sloan Center for Theoretical Neurobiology, Box 0444
University of California, San Francisco
513 Parnassus Ave, San Francisco, CA 94143
Abstract
In this paper we present a new method for studying auditory systems based on m-sequences. The method allows us to perturbatively study the linear response of the system in the presence of
various other stimuli, such as speech or sinusoidal modulations.
This allows one to construct linear kernels (receptive fields) at the
same time that other stimuli are being presented. Using the method
we calculate the modulation transfer function of single units in the
inferior colli cui us of the cat at different operating points and discuss
nonlinearities in the response.
1
Introduction
A popular approach to systems identification, i.e., identifying an accurate analytical model for the system behavior, is to use Volterra or Wiener expansions to
model behavior via functional Taylor or orthogonal polynomial series, respectively
[Marmarelis and Marmarelis1978]. Both approaches model the response r(t) as a
linear combination of small powers of the stimulus s(t). Although effective for mild
nonlinearities, deriving the linear combinations becomes numerically unstable for
highly nonlinear systems. A more serious problem is that many biological systems
are adaptive, i.e., the system behavior is dependent on the stimulus ensemble. For
instance, [Rieke et al.1995] found that in the auditory nerve of the bullfrog linearity
and information rates depended sensitively on whether a white noise or naturalistic
ensemble is used.
One approach to handling these difficulties is to forgo the full expansion, and simply compute the linear response to small (perturbative) stimuli in the presence
of various different ensembles, or operating points. By collecting linear responses
? Email: kvale@phy.ucsf.edu and chris@phy.ucsf.edu
Perturbative M-Sequences for Auditory Systems Identification
181
from different operating points, one may fit nonlinear responses as one fits a nonlinear function with a piecewise linear approximation. For adaptive systems the
same procedure would be applied, with different operating points corresponding to
different points along the time axis. Perturbative stimuli have wide application in
condensed-matter physics, where they are used to characterize linear responses such
as resistance, elasticity and viscosity, and in engineering, perturbative analyses are
used in circuit analysis (small signal models) and structural diagnostics (vibration
analysis). In neurophysiology, however, perturbative stimuli are unknown.
An effective stimulus for calculating the perturbative linear response of a system
is the m-sequence. M-sequences have a long history of use in engineering and the
physical sciences, with applications ranging from systems identification to cryptography and cellular communication. In physiology, m-sequences have been used
primarily to compute system kernels [Marmarelis and Marmarelisl978], especially
in the visual system [Pinter and Nabet1987]. In this work, we use perturbative msequences to study the linear response of single units in the inferior colli cui us of a
cat to amplitude-modulated (AM) stimuli. We add a small m-sequence signal to
an AM carrier, which allows us to study the linear behavior of the system near a
particular operating point in a non-destructive manner, i.e., without changing the
operating point. Perturbative m-sequences allow one to calculate linear responses
near the particular stimuli under study with only a little extra effort, and allow us
to characterize the system over a wide range of stimuli, such as sinusoidal AM and
naturalistic stimuli.
The auditory system we selected to study was the response of single units in the
central nucleus of the inferior colliculus (IC) of an anaesthetised cat. Single unit
responses were recorded extraceUularly. Action potentials were amplified and stored
on DAT tape, and were discriminated offline using a commercial computer-based
spike sorter (Brainwave). 20 units were recorded, of which 10 yielded sufficiently
stable responses to be analyzed.
2
M-Sequences and Linear Systems
A binary m-sequence is a two-level pseudo-random sequence of +1's and -1's. The
sequence length is L = 2n - 1, where n is the order of the sequence. Typically, a
binary m-sequence can be generated by a shift register with n bits and feedback
connections derived from an irreducible polynomial over the multiplicative group Z2
[Golomb1982]. For linear systems identification, m-sequences have two important
properties. The first is that m-sequences have nearly zero mean: ~~':OI m[t] = -l.
The second is that the autocorrelation function takes on the impulse-like form
L-l
Smm(T)
= ~ m[t]m[t + T] =
{L-1
if T = 0
otherwise
(1)
Impulse stimuli also have a 8-function autocorrelation function. In the context
of perturbative stimuli, the advantage of an m-sequence stimulus over an impulse
stimulus is that for a given signal to noise ratio, an m-sequence perturbation stays
much closer to the original signal (in the least squares sense) than an impulse perturbation. Thus the perturbed signal does not stray as far from the operating point
and measurement of linear response about that operating point is more accurate.
We model the IC response with a system F through which a scalar stimulus s(t) is
passed to give a response r(t):
r(t) = F[s(t)].
(2)
M. Kvale and C. E. Schreiner
182
For the purposes of this section, the functional F is taken to be a linear functional
plus a DC component. In real experiments, the input and output signal are sampled
into discrete sequences with t becoming an integer indexing the sequence. Then the
system can be written as the discrete convolution
L-1
r[t] = ho
+L
(3)
h[ttls[t - t1]
with kernels ho and h[tt] to be determined. We assume that the system has a finite
memory of M time steps (with perhaps a delay) so that at most M of the h[t]
coefficients are nonzero. To determine the kernels perturbatively, we add a small
amount of m-sequence to a base stimulus so:
s[t] = so[t]
+ am[t].
(4)
Cross-correlating the response with the original m-sequence yields
L-l
L-l
t=o
t=o
L m[t]r[t + r] = L
L-l L- l
m[t]ho +
LL
h[ttlm[t]so[t + r - ttl
t=o tl =0
L-1 L-l
+L
L
ah[tdm[t]m[t + r - tl]'
(5)
t=o tl=O
Using the sum formula for am -sequence above, the first sum in Eq. (5) can be
simplified to -ho. Using the autocorrelation Eq. (1), the third sum in Eq. (5)
simplifies, and we find
L-l
Rrm(r)
= a(L + l)h[r] -
ho - a
L
L-l L-l
h[tl]
+L
tl =0
L
h[tt]m[t]so[t + r -
ttl
(6)
t=o tl =0
Although the values for the kernels h(t) are set implicitly by this equation, the
terms on the right hand side of Eq. (6) are widely different in size for large Land
the equation can be simplified. As is customary in auditory systems, we assume
the DC response ho is small. To estimate the size of the other terms, we compute
statistical estimates of their sizes and look at their scaling with the parameters.
The term a L:~-==~ h[tt] is a sum of M kernel elements; they may be correlated or
uncorrelated, so a conservative estimate of their size is on the order of O( aM).
The last term in (6) is more subtle. We rewrite it as
L-l L-l
LL
L-l
h[tdm[t]so[t + r -
h=O t=o
ttl
=
L
h[tt]p[r, ttl
tl=O
L-l
L m[t]so[t + r - ttl
(7)
t=o
The time series of the ambient stimulus so[t] and m-sequence m[t] are assumed to
be uncorrelated. By the central limit theorem, the sum p[r, tl] will then have an
average of zero with a standard deviation of 0(L 1 / 2 ). If in turn, the terms p[r, ttl
are un correlated with the kernels h[tl], we have that
L-l L-l
L L h[tt]m[t]so[t + r - ttl '" 0(MI/2 Ll/2)
tl=O t=o
(8)
Perturbative M-Sequences for Auditory Systems Identification
183
If N cycles of the m-sequence are performed, in which sort] is different for each
cycle, all the terms in Eq. (6) scale with N as O(N), except for the double sum.
By the same central limits arguments above, the double sum scales as O(Nl/2).
Putting all these results together into Eq. (6) and solving for the kernels yields
h(r)
a(L1+ 1) Rrm(r) - 0 (
~) + 0 (aN~~~1/2 )
M
1
a(L + 1) Rrm(r) - C L
1
Ml/2
2
+ C aNI/2?1/2'
.
(9)
with the constants C1 , C2 '" O(h[r]) depending neural firing rate, statistics, etc.,
determined from experiment. If we take the kernel element h(r) to be the first term
in Eq. 9, then the last two terms in Eq. (9) contribute errors in determining the
kernel and can be thought of as noise. Both error terms vanish as L -+ 00 and the
procedure is asymptotically exact for arbitrary uncorrelated stimuli sort]. In order
for the cross-correlation Ram (r) to yield a good estimate, the inequalities
(10)
must hold. In practice, the kernel memory is much smaller than the sequence length,
and the second inequality is the stricter bound. The second inequality represents a
tradeoff among sequence length, number of trials and the size of the perturbation for
a given level of systematic noise in the kernel estimate. For instance, if L = 215 - 1,
N = 10, M = 30, and noise floor at 10%, the perturbation should be larger than
a = 0.095. If no signal sort] is present, then the O(Ml/2a- 1(NL)-1/2) term drops
out and the usual m-sequence cross-correlation result is recovered.
3
M-Sequences for Modulation Response
Previous work, e.g., [M011er and Rees1986, Langner and Schreinerl988] has shown
that many of the cells in the inferior colliculus are tuned not only to a characteristic
frequency, but are also tuned to a best frequency of modulation of the carrier. A
highly simplified model of the IC unit response to sound stimuli is the Ll- N - L2
cascade filter, with L1 a linear tank circuit with a transfer function matching that
of the frequency tuning curve, N a nonlinear rectifying unit, and L2 a linear circuit with a transfer function matching that of the modulation transfer function.
Detecting this modulation is an inherently nonlinear operation and N is not well
approximated by a linear kernel. Thus IC modulation responses will not be well
characterized by ordinary m-sequence stimuli using the methods described in Section 2.
A better approach is to bypass the Ll - N demodulation step entirely and concentrate on measuring L2. This can be accomplished by creating a modulation
m-sequence:
(11)
s[t] = a (so[t] + bm[t]) sin[wet],
where Iso[t]1 :::; 1 is the ambient signal, i.e., the operating point, m[t] E [-1,1] is an
m-sequence added with amplitude b, and We is the carrier frequency. Demodulation
gives the effective input stimulus
sm[t] = a (so[t] + bm[t]) .
(12)
Note that there is little physiological evidence for a purely linear rectifier N. In
fact, both the work of [M011er and Rees1986, Rees and M011er1987] and ours below
show that there is a nonlinear modulation response. Taking a modulation transfer
184
M. Kvale and C. E. Schreiner
function seriously, however, implies that one assumes that modulation response
is linear, which implies that the static nonlinearity used is something like a halfwave rectifier. Linearity is used here as a convenient assumption for organizing the
stimulus and asking whether nonlinearities exist.
For full m-sequence modulation (so[t] = 1 and b = 1) the stimulus Sm and the
neural response can be used to compute, via the Lee--Schetzen cross-correlation,
the modulation transfer function for the L2 system. Alternatively, for b ? 1, the
m-sequence is a perturbation on the underlying modulation envelope sort]. The
derivation above shows that the linear modulation kernel can also be calculated
using a Lee--Schetzen cross-correlation. M-sequences at full modulation depth were
first used by [M0ller and Rees1986, Rees and M011erI987] to calculate white-noise
kernels. Here, we are using m-sequence in a different way-we are calculating the
small-signal properties around the stimulus sort].
The m-sequences used in this experiment were of length 215 -1 = 32,767. For each
unit, 10 cycles of the m-sequence were presented back-to-back. After determining
the characteristic frequency of a unit, stimuli were presented which never differed
from the characteristic frequency by more than 500 Hz. Figure 1 depicts the sinusoidal and m-sequence components and their combined result. The stimuli were
presented in random order so as to mitigate adaptation effects.
Figure 1: A depiction of stimuli used in the experiment. The top
a pure sine wave modulation at modulation depth 0.8. The middle
an m-sequence modulation at depth 1.0. The bottom graph shows a
m-sequence modulation at depth 0.2 added to a sinusoidal modulation
4
graph shows
graph shows
perturbative
at depth 0.8.
Results
Figure 2 shows the spike rates for both the pure sinusoid and the combined sinusoid
and m-sequence stimuli. Note that the rates are nearly the same, indicating that
the perturbation did not have a large effect on the average response of the unit.
The unit shows an adaptation in firing rate over the 10 trials, but we did not find
Perturbative M-Sequences for Auditory Systems Identification
185
a statistically significant change in the kernels of different trials in any of the units.
G----e sinusoid
~ sinusoid + m-sequence
.-...
o
Q)
~
80.0
C/)
Q)
~
'5.
~
60.0
Q)
?i
"40.0
100.0
200.0
300.0
400.0
500.0
Time (sec)
Figure 2: A plot of the unit firing rates for both the pure sinusoid and the sinusoid +
m-sequence stimuli. The carrier frequency is 9 kHz and is close to the characteristic
frequency of the neuron. The sinusoidal modulation has a frequency of 20 Hz and
the m-sequence modulation has a frequency of 800 sec-I .
Figure 3 shows modulation response kernels at several different values of the modulation depth. Note that if the system was a linear, superposition would cause all
the kernels to be equivalent; in fact it is seen that the nonlinearities are of the same
magnitude as the linear response. In this particular unit, the triphasic behavior
at small modulation depths gives way to monophasic behavior at high modulation
depths and an FFT of the kernel shows that the bandwidth of the modulation
transfer function also broadens with increasing depth.
5
Discussion
In this paper, we have introduced a new type of stimulus, perturbative m-sequences,
for the study of auditory systems and derived their properties. We then applied
perturbative m-sequences to the analysis of the modulation response of units in the
Ie, and found the linear response at a few different operation point. We demonstrated that the nonlinear response in the presence of sinusoidal modulations are
nearly as large as the linear response and thus a description of unit response with
only an MTF is incomplete. We believe that perturbative stimuli can be an effective
tool for the analysis of many systems whose units phase lock to a stimulus.
The main limiting factor is the systematic noise discussed in section 2, but it is
possible to trade off duration of measurement and size of the perturbation to achieve
good results. The m-sequence stimuli also make it possible to derive higher order
information [Sutter1987] and with a suitable noise floor, it may be possible to derive
second-order kernels as well.
This work was supported by The Sloan foundation and ONR grant number N0001494-1-0547.
M. Kvale and C. E. Schreiner
186
Response vs. modulation depth
sine wave @40Hz + pert. m-sequence
90.0
- 0.2
- 0.4
-0.6
-0.8
-1.0
70.0
50.0
CD
:::>
"C
""
a.
30.0
E
ns
10.0
-10.0
-30.0
-50.0
0.0
5.0
10.0
15.0
20.0
time from spike (milliseconds)
Figure 3: A plot of the temporal kernels derived from perturbative m-sequence
stimuli in conjunction with sinusoidal modulations at various modulation depth.
The y-axis units are amplitude per spike and the x-axis is in milliseconds before the
spike.
References
[Golomb1982] S. W. Golomb. Shift Register Sequences. Aegean Park Press, Laguna
Hills, CA, 1982.
[Langner and Schreiner1988] G. Langner and C. E. Schreiner. Periodicity coding
in the inferior colliculus of the cat: 1. neuronal mechanisms. Journal of Neurophysiology, 60: 1799-1822, 1988.
[Marmarelis and Marmarelis1978] Panos Z. Marmarelis and Vasilis Z. Marmarelis.
Analysis of Physiological Systems. Plenum Press, New York, NY, 10011, 1978.
[M011er and Rees1986] Aage R. M011er and Adrian Rees. Dynamic properties of
single neurons in the inferior colliculus of the rat. Hearing Research, 24:203-215,
1986.
[Pinter and Nabet1987] Robert B. Pinter and Bahram Nabet. Nonlinear Vision.
CRC Press, Boca Raton, FL, 1987.
[Rees and M011er1987] Adrian Rees and Aage R. M01ler. Stimulus properties influencing the responses of inferior colliculus neurons to amplitude-modulated
sounds. Hearing Research, 27:129-143, 1987.
[Rieke et al.1995] F. Rieke, D. A. Bodnar, and W. Bialek. Naturalistic stimuli
increase the rate and efficiency of information transmission by primary auditory
afi"erents. Proceedings of the Royal Society of London. Series B, 262:259-265,
1995.
[Sutter1987] E. E. Sutter. A practical non-stochastic approach to nonlinear timedomain analysis. In Vasilis Z. Marmarelis, editor, Advanced Methods of Physiological Modeling, Vol. 1, pages 303-315. Biomedical Simulations Resource, University of Southern California, Los Angeles, CA 90089-1451, 1987.
| 1430 |@word mild:1 neurophysiology:2 trial:3 middle:1 polynomial:2 adrian:2 simulation:1 phy:2 series:3 tuned:2 ours:1 seriously:1 recovered:1 z2:1 perturbative:17 written:1 must:1 drop:1 plot:2 v:1 selected:1 iso:1 anaesthetised:1 sutter:1 detecting:1 contribute:1 along:1 c2:1 autocorrelation:3 manner:1 behavior:6 little:2 increasing:1 becomes:1 linearity:2 underlying:1 circuit:3 golomb:1 ttl:7 pseudo:1 mitigate:1 temporal:1 collecting:1 stricter:1 unit:18 grant:1 carrier:4 t1:1 engineering:2 before:1 influencing:1 depended:1 limit:2 laguna:1 firing:3 modulation:33 becoming:1 mtf:1 plus:1 christoph:1 range:1 statistically:1 practical:1 practice:1 procedure:2 physiology:1 thought:1 cascade:1 matching:2 convenient:1 naturalistic:3 close:1 context:1 equivalent:1 demonstrated:1 center:1 duration:1 identifying:1 pure:3 schreiner:5 deriving:1 rieke:3 limiting:1 plenum:1 commercial:1 exact:1 element:2 approximated:1 bottom:1 boca:1 calculate:3 cycle:3 trade:1 colli:2 dynamic:1 rewrite:1 solving:1 purely:1 efficiency:1 langner:3 various:3 cat:4 derivation:1 effective:4 london:1 broadens:1 whose:1 widely:1 larger:1 otherwise:1 statistic:1 sequence:53 advantage:1 analytical:1 adaptation:2 vasilis:2 organizing:1 achieve:1 amplified:1 description:1 los:1 double:2 transmission:1 depending:1 derive:2 eq:8 implies:2 concentrate:1 filter:1 stochastic:1 crc:1 biological:1 hold:1 sufficiently:1 around:1 ic:4 purpose:1 condensed:1 wet:1 superposition:1 vibration:1 tool:1 sensitively:1 conjunction:1 derived:3 ave:1 am:6 sense:1 dependent:1 typically:1 tank:1 among:1 field:1 construct:1 never:1 represents:1 park:1 look:1 nearly:3 stimulus:38 piecewise:1 serious:1 primarily:1 irreducible:1 few:1 phase:1 highly:2 analyzed:1 nl:2 diagnostics:1 accurate:2 ambient:2 closer:1 elasticity:1 orthogonal:1 afi:1 incomplete:1 taylor:1 theoretical:1 instance:2 modeling:1 asking:1 measuring:1 ordinary:1 hearing:2 deviation:1 delay:1 characterize:2 stored:1 perturbed:1 combined:2 rees:5 ie:1 stay:1 systematic:2 physic:1 lee:2 off:1 together:1 central:3 recorded:2 marmarelis:6 sorter:1 creating:1 potential:1 sinusoidal:7 nonlinearities:4 rrm:3 sec:2 coding:1 coefficient:1 matter:1 sloan:2 register:2 multiplicative:1 performed:1 sine:2 wave:2 sort:5 rectifying:1 perturbatively:2 oi:1 square:1 wiener:1 characteristic:4 ensemble:3 yield:3 identification:7 history:1 ah:1 email:1 frequency:10 destructive:1 mi:1 static:1 sampled:1 auditory:10 popular:1 subtle:1 amplitude:4 back:2 nerve:1 higher:1 response:34 box:1 biomedical:1 correlation:4 hand:1 nonlinear:9 perhaps:1 impulse:4 believe:1 effect:2 sinusoid:6 nonzero:1 white:2 ll:5 sin:1 inferior:7 nabet:1 rat:1 hill:1 tt:5 l1:2 ranging:1 functional:3 physical:1 discriminated:1 khz:1 discussed:1 numerically:1 measurement:2 significant:1 tuning:1 nonlinearity:1 stable:1 operating:9 depiction:1 etc:1 add:2 base:1 something:1 tdm:2 binary:2 inequality:3 onr:1 kvale:5 accomplished:1 seen:1 floor:2 determine:1 signal:9 full:3 sound:2 characterized:1 cross:5 long:1 demodulation:2 panos:1 vision:1 kernel:21 cell:1 c1:1 parnassus:1 extra:1 envelope:1 hz:3 integer:1 structural:1 near:2 presence:3 fft:1 fit:2 bandwidth:1 simplifies:1 tradeoff:1 shift:2 schetzen:2 angeles:1 whether:2 passed:1 pinter:3 effort:1 resistance:1 speech:1 york:1 cause:1 tape:1 action:1 amount:1 viscosity:1 halfwave:1 exist:1 millisecond:2 per:1 discrete:2 timedomain:1 vol:1 group:1 putting:1 changing:1 ani:1 ram:1 asymptotically:1 graph:3 sum:7 colliculus:5 scaling:1 bit:1 entirely:1 bound:1 fl:1 yielded:1 argument:1 combination:2 cui:2 smaller:1 indexing:1 taken:1 equation:2 resource:1 discus:1 turn:1 mechanism:1 studying:1 operation:2 ho:6 customary:1 original:2 assumes:1 top:1 lock:1 calculating:2 especially:1 society:1 dat:1 added:2 spike:5 volterra:1 receptive:1 primary:1 usual:1 bialek:1 southern:1 chris:1 unstable:1 cellular:1 length:4 ratio:1 robert:1 unknown:1 convolution:1 neuron:3 sm:2 finite:1 neurobiology:1 communication:1 dc:2 perturbation:7 arbitrary:1 raton:1 introduced:1 connection:1 california:2 below:1 royal:1 memory:2 power:1 suitable:1 difficulty:1 advanced:1 axis:3 l2:4 determining:2 foundation:1 nucleus:1 editor:1 uncorrelated:3 bypass:1 land:1 cd:1 periodicity:1 supported:1 last:2 offline:1 side:1 allow:2 wide:2 taking:1 feedback:1 curve:1 calculated:1 depth:11 pert:1 adaptive:2 san:2 simplified:3 bm:2 far:1 smm:1 implicitly:1 ml:2 correlating:1 assumed:1 francisco:2 alternatively:1 un:1 transfer:7 ca:3 correlated:2 inherently:1 expansion:2 did:2 main:1 noise:8 cryptography:1 neuronal:1 tl:10 depicts:1 differed:1 ny:1 n:1 stray:1 bullfrog:1 vanish:1 third:1 formula:1 theorem:1 rectifier:2 physiological:3 evidence:1 magnitude:1 simply:1 visual:1 scalar:1 change:1 determined:2 except:1 conservative:1 forgo:1 aegean:1 indicating:1 mark:1 modulated:2 ucsf:2 bodnar:1 handling:1 |
473 | 1,431 | Learning to Order Things
William W. Cohen Robert E. Schapire Yoram Singer
AT&T Labs, 180 Park Ave., Florham Park, NJ 07932
{wcohen,schapire,singer} @research.att.com
Abstract
There are many applications in which it is desirable to order rather than classify
instances. Here we consider the problem of learning how to order, given feedback
in the form of preference judgments, i.e., statements to the effect that one instance
should be ranked ahead of another. We outline a two-stage approach in which one
first learns by conventional means a preference Junction, of the form PREF( u, v),
which indicates whether it is advisable to rank u before v. New instances are
then ordered so as to maximize agreements with the learned preference function. We show that the problem of finding the ordering that agrees best with
a preference function is NP-complete, even under very restrictive assumptions.
Nevertheless, we describe a simple greedy algorithm that is guaranteed to find a
good approximation. We then discuss an on-line learning algorithm, based on the
"Hedge" algorithm, for finding a good linear combination of ranking "experts."
We use the ordering algorithm combined with the on-line learning algorithm to
find a combination of "search experts," each of which is a domain-specific query
expansion strategy for a WWW search engine, and present experimental results
that demonstrate the merits of our approach.
1 Introduction
Most previous work in inductive learning has concentrated on learning to classify. However,
there are many applications in which it is desirable to order rather than classify instances.
An example might be a personalized email filter that gives a priority ordering to unread
mail. Here we will consider the problem of learning how to construct such orderings, given
feedback in the form of preference judgments, i.e., statements that one instance should be
ranked ahead of another.
Such orderings could be constructed based on a learned classifier or regression model,
and in fact often are. For instance, it is common practice in information retrieval to rank
documents according to their estimated probability of relevance to a query based on a
learned classifier for the concept "relevant document." An advantage of learning orderings
directly is that preference judgments can be much easier to obtain than the labels required
for classification learning.
For instance, in the email application mentioned above, one approach might be to rank
messages according to their estimated probability of membership in the class of "urgent"
messages, or by some numerical estimate of urgency obtained by regression. Suppose,
however, that a user is presented with an ordered list of email messages, and elects to read
the third message first. Given this election, it is not necessarily the case that message three
is urgent, nor is there sufficient information to estimate any numerical urgency measures;
however, it seems quite reasonable to infer that message three should have been ranked
ahead of the others. Thus, in this setting, obtaining preference information may be easier
and more natural than obtaining the information needed for classification or regression.
w. W. Cohen, R. E. Schapire and Y. Singer
452
In the remainder of this paper, we will investigate the following two-stage approach to
learning how to order. In stage one, we learn a preference junction, a two-argument
function PREF( u, v) which returns a numerical measure of how certain it is that u should
be ranked before v. In stage two, we use the learned preference function to order a set of
new instances U; to accomplish this, we evaluate the learned function PREF( u, v) on all
pairs of instances u, v E U, and choose an ordering of U that agrees, as much as possible,
with these pairwise preference judgments. This general approach is novel; for related work
in various fields see, for instance, references [2, 3, 1, 7, 10].
As we will see, given an appropriate feature set, learning a preference function can be
reduced to a fairly conventional classification learning problem. On the other hand, finding
a total order that agrees best with a preference function is NP-complete. Nevertheless, we
show that there is an efficient greedy algorithm that always finds a good approximation to
the best ordering. After presenting these results on the complexity of ordering instances
using a preference function, we then describe a specific algorithm for learning a preference
function. The algorithm is an on-line weight allocation algorithm, much like the weighted
majority algorithm [9] and Winnow [8], and, more directly, Freund and Schapire's [4]
"Hedge" algorithm. We then present some experimental results in which this algorithm is
used to combine the results of several "search experts," each of which is a domain-specific
query expansion strategy for a WWW search engine.
2 Preliminaries
Let X be a set of instances (possibly infinite). A preference junction PREF is a binary
function PREF : X x X ~ [0,1]. A value of PREF(u, v) which is close to 1 or a is
interpreted as a strong recommendation that u should be ranked before v. A value close to
1/2 is interpreted as an abstention from making a recommendation. As noted above, the
hypothesis of our learning system will be a preference function, and new instances will be
ranked so as to agree as much as possible with the preferences predicted by this hypothesis.
In standard classification learning, a hypothesis is constructed by combining primitive
features. Similarly, in this paper, a preference function will be a combination of other
preference functions. In particular, we will typically assume the availability of a set of N
primitive preference functions RI , ... , RN. These can then be combined in the usual ways,
e.g., with a boolean or linear combination of their values; we will be especially interested
in the latter combination method.
It is convenient to assume that the Ri'S are well-formed in certain ways. To this end, we
introduce a special kind of preference function called a rank ordering. Let S be a totally
ordered set l with' >' as the comparison operator. An ordering function into S is a function
f : X ~ S. The function f induces the preference function Rj, defined as
I
Rj(u,v) ~ { 0
21
if f (u) > f (v)
if f(u) < f(v)
otherwise.
We call Rf a rank ordering for X into S. If Rf(u, v)
to v, or u is ranked higher than v.
= I, then we say that u is preferred
It is sometimes convenient to allow an ordering function to "abstain" and not give a
preference for a pair u, v. Let ?> be a special symbol not in S, and let f be a function into
S U {?>}. We will interpret the mapping f (u)
?> to mean that u is "unranked," and let
Rf (u, v) = if either u or v is unranked.
=
!
To give concrete examples of rank ordering, imagine learning to order documents based on
the words that they contain. To model this, let X be the set of all documents in a repository,
)That is, for all pairs of distinct elements
8J, 82
E S, either 8)
<
82
or 8)
> 82 .
Learning to Order Things
453
and for N words WI, ... , W N. let Ii (u) be the number of occurrences of Wi in u. Then
Rf; will prefer u to v whenever Wi occurs more often in u than v. As a second example.
consider a meta-search application in which the goal is to combine the rankings of several
WWW search engines. For N search engines el, ... , eN. one might define h so that R'i
prefers u to v whenever u is ranked ahead of v in the list Li produced by the corresponding
search engine. To do this, one could let Ii(u) = -k for the document u appearing in the
k-th position in the list L i ? and let Ii( u) = </> for any document not appearing in L i .
3
Ordering instances with a preference function
We now consider the complexity of finding the total order that agrees best with a learned
preference function. To analyze this. we must first quantify the notion of agreement between
a preference function PREF and an ordering. One natural notion is the following: Let X
be a set. PREF be a preference function. and let p be a total ordering of X. expressed
again as an ordering function (i.e .? p( u) > p( v) iff u precedes v in the order). We define
AGREE(p, PREF) to be the sum of PREF( u, v) over all pairs u, v such that u is ranked
ahead of v by p:
AGREE(p, PREF) =
(1)
PREF(u, v).
u.v:p{ul>p{v)
Ideally. one would like to find a p that maximizes AGREE(p, PREF). This general optimization problem is of little interest since in practice, there are many constraints imposed
by learning: for instance PREF must be in some restricted class of functions. and will
generally be a combination of relatively well-behaved preference functions R i . A more
interesting question is whether the problem remains hard under such constraints.
The theorem below gives such a result. showing that the problem is NP-complete even if
PREF is restricted to be a linear combination of rank orderings. This holds even if all the
rank orderings map into a set S with only three elements. one of which mayor may not be
</>. (Clearly. if S consists of more than three elements then the problem is still hard.)
Theorem 1 The following decision problem is NP-complete:
Input: A rational number 1\,; a set X; a set S with lSI ~ 3; a collection of
N ordering functions Ii : X -t S; and a preference function PREF defined as
PREF(u, v) = L~I wiR'i (u, v) where w = (WI, ... ,WN) is a weight vector in [0, l]N
with L~I Wi = 1.
Question: Does there exist a total order p such that AGREE(p, PREF)
~ I\,?
The proof (omitted) is by reduction from CYCLIC-ORDERING [5. 6].
Although this problem is hard when lSI ~ 3. it becomes tractable for linear combinations
of rank orderings into a set S of size two. In brief. suppose one is given X, Sand PREF as
in Theorem 1, save that S is a two-element set. which we assume without loss of generality
to be S = {O, I}. Now define p(u) = Li Wdi(U). It can be shown that the total order
defined by p maximizes AGREE(p, PREF). (In case of a tie, p( u) = p( v) for distinct u
and v. p defines only a partial order. The claim still holds in this case for any total order
which is consistent with this partial order.) Of course, when lSI = 2, the rank orderings
are really only binary classifiers. The fact that this special case is tractable underscores the
fact that manipulating orderings can be computationally more difficult than performing the
corresponding operations on binary classifiers.
Theorem 1 implies that we are unlikely to find an efficient algorithm that finds the optimal
total order for a weighted combination of rank orderings. Fortunately. there do exist efficient
algorithms for finding an approximately optimal total order. Figure 1 summarizes a greedy
w. W. Cohen, R. E. Schapire and Y. Singer
454
Algorithm Order- By- Preferences
Inputs: an instance set X; a preference function PREF
Output: an approximately optimal ordering function p
let V = X
for each v E V do7l'(v) = LUEVPREF(v,u) - LUEVPREF(u,v)
while V is non-empty do
let t = argmaxuEv 71'(u)
let pet) IVI
=
V=V-{t}
for each v E V do 71'(v) = 71'(v)
+ PREF(t, v) -
PREF(v, t)
endwhile
Figure 1: A greedy ordering algorithm
algorithm that produces a good approximation to the best total order, as we will shortly
demonstrate.
The algorithm is easiest to describe by thinking of PREF as a directed
weighted graph where, initially, the set of vertices V is equal to the set of instances X,
and each edge u -t v has weight PREF( u, v). We assign to each vertex v E V a potential
value 71'( v), which is the weighted sum of the outgoing edges minus the weighted sum of
the ingoing edges. That is, 71'(v) = LUEV PREF(v,u) - LUEV PREF(u, v) . The greedy
algorithm then picks some node t that has maximum potential, and assigns it a rank by
setting pet) = lVI, effectively ordering it ahead of all the remaining nodes. This node,
together with all incident edges, is then deleted from the graph, and the potential values
71' of the remaining vertices are updated appropriately: This process is repeated until the
graph is empty; notice that nodes removed in subsequent iterations will have progressively
smaller and smaller ranks.
The next theorem shows that this greedy algorithm comes within a factor of two of optimal.
Furthermore, it is relatively simple to show that the approximation factor of 2 is tight.
Theorem 2 Let OPT(PREF) be the weighted agreement achieved by an optimal total
orderfor the preference junction PREF and let APPROX(PREF) be the weighted agreement
achieved by the greedy algorithm. Then APPROX(PREF) ;::: !OPT(PREF).
4 Learning a good weight vector
In this section, we look at the problem of learning a good linear combination of a set of
preference functions. Specifically, we assume access to a set of ranking experts which
provide us with preference functions Ri of a set of instances. The problem, then, is to learn
a preference function of the form PREF(u,v)
L~I wiRi(U,V). We adopt the on-line
learning framework first studied by Littlestone [8J in which the weight Wi assigned to each
ranking expert Ri is updated incrementally.
=
Learning is assumed to take place in a sequence of rounds. On the t-th round, the learning
algorithm is provided with a set X t of instances to be ranked and to a set of N preference
functions R~ of these instances. The learner may compute R!( u, v) for any and all preference
functions R~ and pairs u, v E X t before producing a final ordering Pt of xt. Finally, the
learner receives feedback from the environment. We assume that the feedback is an arbitrary
set of assertions of the form "u should be preferred to v." That is, formally we regard the
feedback on the t-th round as a set Ft of pairs (u, v) indicating such preferences.
The algorithm we propose for this problem is based on the "weighted majority algorithm" [9J
and, more directly, on the "Hedge" algorithm [4]. We define the loss of a preference function
Learning to Order Things
455
Allocate Weights for Ranking Experts
Parameters:
(3 E [0,1] , initial weight vector WI E [0, I]N with l:~1
N ranking experts, number of rounds T
Do fort = 1,2, ... ,T
wl =
1
1. Receive a set of elements X t and preference functions R~, ... , R'N.
2. Use algorithm Order-By-Preferences to compute ordering function
proximatesPREFt(u,v) = E~I wiRHu,v).
3. Order X t using
Pt
which ap-
Pt .
4. Receive feedback Ft from the user.
5. Evaluate losses Loss(RL Ft) as defined in Eq. (2).
6. Set the new weight vector w!+ 1 = w! . (3Loss(R: ,Ft) / Zt where Zt is a normalization
constant, chosen so that E~I w!+1 = 1.
Figure 2: The on-line weight allocation algorithm.
R with respect to the user's feedback F as
L
oss
(R F) ~
E(U,V)EF(1 -
R(u,v))
IFI'
,
(2)
This loss has a natural probabilistic interpretation. If R is viewed as a randomized prediction
algorithm that predicts that u will precede v with probability R(u, v), then Loss(R, F) is
the probability of R disagreeing with the feedback on a pair (u, v) chosen uniformly at
random from F.
We now can use the Hedge algorithm almost verbatim, as shown in Figure 2. The algorithm
maintains a positive weight vector whose value at time t is denoted by w t
(wf, . . . , w'N).
If there is no prior knowledge about the ranking experts, we set all initial weights to be
equal so that
= 1/N. The weight vector w t is used to combine the preference functions
of the different experts to obtain the preference function PREFt = E~ I w~ R~. This, in
tum, is converted into an ordering Pt on the current set of elements Xl using the method
described in Section 3. After receiving feedback pt, the loss for each preference function
Loss(RL Ft) is evaluated as in Eq. (2) and the weight vector w t is updated using the
mUltiplicative rule W!+I = w~ . (3LQss(R: ,Ft) / Zt where (3 E [0, 1] is a parameter, and Zt is
a normalization constant, chosen so that the weights sum to one after the update. Thus, based
on the feedback, the weights of the ranking experts are adjusted so that experts producing
preference functions with relatively large agreement with the feedback are promoted.
=
wI
We will briefly sketch the theoretical rationale behind this algorithm. Freund and
Schapire [4] prove general results about Hedge which can be applied directly to this loss
function. Their results imply almost immediately a bound on the cumulative loss of the
preference function PREFt in terms of the loss of the best ranking expert, specifically
T
LLoss(PREFt,F t )
T
~ a,Bm~n LLoss(RLFt) +c,BlnN
l
t=1
t=1
where a,B = InO / (3) / (1 - (3) and C,B = 1/( I - (3). Thus, if one of the ranking experts has
low loss, then so will the combined preference function PREFt .
However, we are not interested in the loss ofPREFt (since it is not an ordering), but rather in
the performance of the actual ordering Pt computed by the learning algorithm. Fortunately,
w. W. Cohen, R. E. Schapire and y. Singer
456
the losses of these can be related using a kind of triangle inequality. It can be shown that,
for any PREF, F and p:
Loss(Rp, F) ~
OISAGREE(p PREF)
IFI '
+ Loss(PREF, F)
(3)
where, similar to Eq. (1), OISAGREE(p, PREF) = Lu,v :p(u?p(v)(l - PREF(u, v)). Not
surprisingly, maximizing AGREE is equivalent to minimizing DISAGREE.
So, in sum, we use the greedy algorithm of Section 3 to minimize (approximately) the first
term on the right hand side ofEq. (3), and we use the learning algorithm Hedge to minimize
the second term.
5
Experimental results for metasearch
We now present some experiments in learning to combine the results of several WWW
searches. We note that this problem exhibits many facets that require a general approach
such as ours. For instance, approaches that learn to combine similarity scores are not
applicable since the similarity scores of WWW search engines are often unavailable.
We chose to simulate the problem of learning a domain-specific search engine. As test
cases we picked two fairly narrow classes of queries-retrieving the home pages of machine learning researchers (ML), and retrieving the home pages of universities (UNIV).
We obtained a listing of machine learning researchers, identified by name and affiliated
institution, together with their home pages, and a similar list for universities, identified by
name and (sometimes) geographical location. Each entry on a list was viewed as a query,
with the associated URL the sole relevant document.
We then constructed a series of special-purpose "search experts" for each domain. These
were implemented as query expansion methods which converted a name, affiliation pair
(or a name, location pair) to a likely-seeming Altavista query. For example, one expert
for the ML domain was to search for all the words in the person's name plus the words
"machine" and "learning," and to further enforce a strict requirement that the person's last
name appear. Overall we defined 16 search experts for the ML domain and 22 for the UN IV
domain. Each search expert returned the top 30 ranked documents. In the ML domain there
were 210 searches for which at least one search expert returned the named home page; for
the UNIV domain, there were 290 such searches.
For each query t, we first constructed the set X t consisting of all documents returned by all
of the expanded queries defined by the search experts. Next, each search expert i computed
a preference function R~. We chose these to be rank orderings defined with respect to an
ordering function If in the natural way: We assigned a rank of if = 30 to the first listed
document, Ii = 29 to the second-listed document, and so on, finally assigning a rank of
Ii = 0 to every document not retrieved by the expanded query associated with expert i.
To encode feedback, we considered two schemes. In the first we simulated complete
relevance feedback-that is, for each query, we constructed feedback in which the sole
relevant document was preferred to all other documents. In the second, we simulated the
sort of feedback that could be collected from "click data," i.e., from observing a user's
interactions with a metasearch system. For each query, after presenting a ranked list of
documents, we noted the rank of the one relevant document. We then constructed a feedback
ranking in which the relevant document is preferred to all preceding documents. This would
correspond to observing which link the user actually followed, and making the assumption
that this link was preferred to previous links.
To evaluate the expected performance of a fully-trained system on novel queries in this
domain, we employed leave-one-out testing. For each query q, we removed q from the
Learning to Order Things
Learned System (Full Feedback)
Learned System ("Click Data")
Naive
Best (Top 1)
Best (Top 10)
Best (Top 30)
Best (Av. Rank)
457
ML Domain
University Domain
Top 1 Top lO Top 30 Av. rank Top 1 Top lO Top 30 Av. rank
225
253
7.8
114
185
198
4.9
111
4.9
93
185
198
87
229
259
7.8
165
176
7.7
79
157
14.4
89
191
184
119
170
112
221
247
8.2
6.7
114
182
190
5.3
249
8.0
111
223
194
223
8.0
97
181
5.6
111
249
190
5.3
111
223
249
8.0
114
182
Table 1: Comparison of learned systems and individual search queries
query set, and recorded the rank of q after training (with (3 = 0.5) on the remaining
queries. For click data feedback, we recorded the median rank over 100 randomly chosen
permutations of the training queries.
We the computed an approximation to average rank by artificially assigning a rank of 31
to every document that was either unranked, or ranked above rank 30. (The latter case is
to be fair to the learned system, which is the only one for which a rank greater than 30 is
possible.) A summary of these results is given in Table 1, together with some additional
data on "top-k performance"-the number of times the correct homepage appears at rank
no higher than k. In the table we give the top-k performance (for three values of k) and
average rank for several ranking systems: the two learned systems, the naive query (the
person or university's name), and the single search expert that performed best with respect
to each performance measure. The table illustrates the robustness of the learned systems,
which are nearly always competitive with the best expert for every performance measure
listed; the only exception is that the system trained on click data trails the best expert in
top-k performance for small values of k. It is also worth noting that in both domains, the
naive query (simply the person or university's name) is not very effective. Even with the
weaker click data feedback, the learned system achieves a 36% decrease in average rank
over the naive query in the ML domain, and a 46% decrease in the UNIV domain.
To summarize the experiments, on these domains, the learned system not only performs
much better than naive search strategies; it also consistently performs at least as well as,
and perhaps slightly better than, any single domain-specific search expert. Furthermore, the
performance of the learned system is almost as good with the weaker "click data" training
as with complete relevance feedback.
References
[1] D.S. Hochbaum (Ed.). Approximation Algorithms for NP-hard problems. PWS Publishing
Company, 1997.
[2] O. Etzioni, S. Hanks, T. Jiang, R M. Karp, O. Madani, and O. Waarts. Efficient information
gathering on the internet. In 37th Ann. Symp. on Foundations of Computer Science, 1996.
[3] P.C Fishburn. The Theory of Social Choice. Princeton University Press, Princeton, NJ, 1973.
[4] Y. Freund and RE. Schapire. A decision-theoretic generalization of on-line learning and an
application to boosting. Journal of Computer and System Sciences, 1997.
[5] Z. Galil and N. Megido. Cyclic ordering is NP-complete. Theor. Compo Sci. , 5:179-182, 1977.
[6] M.R Gary and D.S. Johnson. Computers and Intractibility: A Guide to the Theory of NPcompleteness. W. H. Freeman and Company, New York, 1979.
[7j P.B. Kantor. Decision level data fusion for routing of documents in the TREC3 context: a best
case analysis of worste case results. In TREC-3, 1994.
[8] N. Littlestone. Learning quickly when irrelevant attributes abound: A new linear-threshold
algorithm. Machine Learning, 2(4), 1988.
[9) N. Littlestone and M.K. Warmuth. The weighted majority algorithm. Infonnation and Computation, 108(2):212-261, 1994.
[10] K.E. Lochbaum and L.A. Streeter. Comparing and combining the effectiveness of latent semantic
indexing and the ordinary vector space model for information retrieval. Infonnation processing
and management, 25(6):665-676, 1989.
| 1431 |@word repository:1 briefly:1 seems:1 pick:1 minus:1 reduction:1 initial:2 cyclic:2 series:1 att:1 score:2 document:20 ours:1 current:1 com:1 comparing:1 assigning:2 must:2 subsequent:1 numerical:3 progressively:1 update:1 greedy:8 warmuth:1 compo:1 institution:1 boosting:1 node:4 location:2 preference:48 wir:1 constructed:6 retrieving:2 consists:1 prove:1 combine:5 symp:1 introduce:1 pairwise:1 expected:1 os:1 nor:1 freeman:1 company:2 election:1 little:1 actual:1 totally:1 becomes:1 provided:1 abound:1 maximizes:2 homepage:1 easiest:1 kind:2 interpreted:2 finding:5 nj:2 every:3 tie:1 classifier:4 appear:1 producing:2 before:4 positive:1 jiang:1 approximately:3 ap:1 might:3 chose:2 plus:1 studied:1 directed:1 testing:1 practice:2 convenient:2 word:4 close:2 operator:1 context:1 www:5 equivalent:1 conventional:2 imposed:1 map:1 maximizing:1 primitive:2 assigns:1 immediately:1 rule:1 notion:2 updated:3 imagine:1 suppose:2 pt:6 user:5 trail:1 hypothesis:3 agreement:5 element:6 predicts:1 disagreeing:1 ft:6 ordering:37 decrease:2 removed:2 mentioned:1 environment:1 complexity:2 ideally:1 trained:2 unread:1 tight:1 learner:2 triangle:1 various:1 univ:3 distinct:2 describe:3 effective:1 query:21 precedes:1 pref:38 quite:1 whose:1 say:1 otherwise:1 florham:1 final:1 advantage:1 sequence:1 propose:1 interaction:1 remainder:1 relevant:5 combining:2 iff:1 empty:2 requirement:1 produce:1 leave:1 advisable:1 sole:2 eq:3 strong:1 implemented:1 predicted:1 implies:1 come:1 quantify:1 correct:1 attribute:1 filter:1 routing:1 sand:1 require:1 assign:1 generalization:1 really:1 preliminary:1 opt:2 theor:1 adjusted:1 hold:2 considered:1 wdi:1 mapping:1 claim:1 achieves:1 adopt:1 omitted:1 purpose:1 precede:1 applicable:1 label:1 infonnation:2 agrees:4 wl:1 weighted:9 clearly:1 always:2 rather:3 karp:1 encode:1 consistently:1 rank:29 indicates:1 underscore:1 ave:1 wf:1 el:1 membership:1 typically:1 unlikely:1 initially:1 manipulating:1 interested:2 overall:1 classification:4 denoted:1 special:4 fairly:2 field:1 construct:1 equal:2 park:2 look:1 nearly:1 thinking:1 np:6 others:1 randomly:1 verbatim:1 individual:1 madani:1 consisting:1 william:1 interest:1 message:6 investigate:1 behind:1 edge:4 partial:2 iv:1 littlestone:3 re:1 theoretical:1 instance:21 classify:3 boolean:1 facet:1 assertion:1 ordinary:1 vertex:3 entry:1 johnson:1 accomplish:1 combined:3 person:4 geographical:1 randomized:1 probabilistic:1 receiving:1 together:3 quickly:1 concrete:1 again:1 recorded:2 management:1 choose:1 possibly:1 fishburn:1 priority:1 expert:25 return:1 li:2 potential:3 converted:2 seeming:1 availability:1 wcohen:1 ranking:12 multiplicative:1 performed:1 picked:1 lab:1 analyze:1 observing:2 competitive:1 sort:1 maintains:1 ofeq:1 minimize:2 formed:1 listing:1 judgment:4 correspond:1 produced:1 lu:1 worth:1 researcher:2 whenever:2 ed:1 email:3 proof:1 associated:2 rational:1 knowledge:1 actually:1 appears:1 ingoing:1 higher:2 tum:1 evaluated:1 generality:1 furthermore:2 hank:1 stage:4 until:1 hand:2 receives:1 sketch:1 incrementally:1 defines:1 perhaps:1 behaved:1 name:8 effect:1 concept:1 contain:1 inductive:1 assigned:2 read:1 semantic:1 round:4 elect:1 noted:2 presenting:2 outline:1 theoretic:1 demonstrate:2 complete:7 performs:2 novel:2 abstain:1 ef:1 common:1 urgency:2 rl:2 cohen:4 interpretation:1 interpret:1 approx:2 similarly:1 access:1 similarity:2 retrieved:1 winnow:1 irrelevant:1 certain:2 meta:1 binary:3 inequality:1 affiliation:1 abstention:1 fortunately:2 greater:1 promoted:1 preceding:1 employed:1 additional:1 maximize:1 ii:6 full:1 desirable:2 rj:2 infer:1 retrieval:2 metasearch:2 prediction:1 regression:3 mayor:1 iteration:1 sometimes:2 normalization:2 hochbaum:1 achieved:2 receive:2 median:1 appropriately:1 ivi:1 strict:1 thing:4 effectiveness:1 call:1 noting:1 wn:1 ifi:2 identified:2 click:6 whether:2 allocate:1 url:1 ul:1 returned:3 york:1 prefers:1 generally:1 listed:3 concentrated:1 induces:1 reduced:1 schapire:8 exist:2 lsi:3 notice:1 estimated:2 nevertheless:2 threshold:1 deleted:1 graph:3 sum:5 named:1 place:1 almost:3 reasonable:1 home:4 decision:3 prefer:1 summarizes:1 bound:1 internet:1 guaranteed:1 followed:1 ahead:6 constraint:2 kantor:1 ri:4 personalized:1 simulate:1 argument:1 performing:1 expanded:2 relatively:3 according:2 combination:10 smaller:2 slightly:1 wi:8 urgent:2 making:2 restricted:2 indexing:1 gathering:1 computationally:1 agree:7 remains:1 discus:1 singer:5 needed:1 merit:1 tractable:2 end:1 junction:4 operation:1 appropriate:1 enforce:1 occurrence:1 appearing:2 save:1 robustness:1 shortly:1 rp:1 top:13 remaining:3 publishing:1 yoram:1 restrictive:1 especially:1 question:2 occurs:1 strategy:3 usual:1 exhibit:1 link:3 simulated:2 sci:1 majority:3 mail:1 collected:1 pet:2 minimizing:1 difficult:1 robert:1 statement:2 zt:4 affiliated:1 disagree:1 av:3 ino:1 trec:1 rn:1 arbitrary:1 pair:9 required:1 fort:1 engine:7 learned:15 narrow:1 below:1 summarize:1 rf:4 ranked:13 natural:4 scheme:1 brief:1 imply:1 naive:5 prior:1 freund:3 loss:17 fully:1 permutation:1 rationale:1 interesting:1 allocation:2 foundation:1 etzioni:1 incident:1 sufficient:1 consistent:1 lo:2 course:1 summary:1 surprisingly:1 last:1 side:1 allow:1 weaker:2 guide:1 regard:1 feedback:20 cumulative:1 collection:1 bm:1 social:1 preferred:5 ml:6 assumed:1 search:24 un:1 latent:1 streeter:1 table:4 learn:3 obtaining:2 unavailable:1 expansion:3 necessarily:1 artificially:1 domain:17 lvi:1 repeated:1 fair:1 en:1 position:1 xl:1 third:1 learns:1 theorem:6 pws:1 specific:5 xt:1 showing:1 symbol:1 list:6 npcompleteness:1 fusion:1 effectively:1 illustrates:1 easier:2 simply:1 likely:1 expressed:1 ordered:3 recommendation:2 gary:1 hedge:6 goal:1 viewed:2 ann:1 hard:4 infinite:1 specifically:2 uniformly:1 total:10 called:1 experimental:3 indicating:1 formally:1 exception:1 latter:2 relevance:3 evaluate:3 outgoing:1 princeton:2 |
474 | 1,432 | The Storage Capacity
of a Fully-Connected Committee Machine
Yuansheng Xiong
Department of Physics, Pohang Institute of Science and Technology,
Hyoja San 31, Pohang , Kyongbuk, Korea
xiongOgalaxy.postech.ac.kr
Chulan Kwon
Department of Physics, Myong Ji University,
Yongin, Kyonggi, Korea
ckwonOwh.myongji.ac.kr
Jong-Hoon Oh
Lucent Technologies, Bell Laboratories,
600 Mountain Ave., Murray Hill , NJ07974, U. S. A.
jhohOphysics.bell-labs.com
Abstract
We study the storage capacity of a fully-connected committee machine with a large number K of hidden nodes. The storage capacity is obtained by analyzing the geometrical structure of the weight
space related to the internal representation . By examining the asymptotic behavior of order parameters in the limit of large K, the
storage capacity Q c is found to be proportional to ]{ Jln ]{ up to the
leading order. This result satisfies the mathematical bound given
by Mitchison and Durbin , whereas the replica-symmetric solution
in a conventional Gardner 's approach violates this bound.
1
INTRODUCTION
Since Gardner's pioneering work on the storage capacity of a single layer
perceptron[1], there have been numerous efforts to use the statistical mechanics
formulation to study feed-forward neural networks. The storage capacity of multilayer neural networks has been of particular interest, together with the generalization problem. Barkai, Hansel and Kanter[2] studied a parity machine with a
The Storage Capacity of a Fully-Connected Committee Machine
379
non-overlapping receptive field of continuous weights within a one-step replica symmetry breaking (RSB) scheme, and their result agrees with a mathematical bound
previously found by Mitchison and Durbin (MD)[3] . Subsequently Barkai, Hansel
and Sompolinsky[4] and Engel et al.[5] have studied the committee machine , which
is closer to the multi-layer perceptron architecture and is most frequently used in
real-world applications. Though they have derived many interesting results , particularly for the case of a finite number of hidden units, it was found that their the
replica-symmetric (RS) result violates the MD bound in the limit where the number
of hidden units K is large.
Recently, Monasson and O'Kane[6] proposed a new statistical mechanics formalism
which can analyze the weight-space structure related to the internal representations
of hidden units . It was applied to single layer perceptrons[7 , 8, 9] as well as multilayer networks[10 , 11, 12]. Monasson and Zecchina[lO] have successfully applied this
formalism to the case of both committee and parity machines with non-overlapping
receptive fields (NRF)[lO] . They suggested that analysis of the RS solution under
this new statistical mechanics formalism can yield results just as good as the onestep RSB solution in the conventional Gardner's method .
In this letter, we apply this formalism for a derivation of the storage capacity of a
fully-connected committee machine, which is also called a committee machine with
overlapping receptive field (ORF) and is believed to be a more relevant architecture.
In particular, we obtain the value of the critical storage capacity in the limit of
large K , which satisfies the MD bound. It also agrees with a recent one-step RSB
calculation, using the conventional Gardner method, to within a small difference of
a numerical prefactor[13]. Finally we will briefly discuss the fully-connected parity
machine.
2
WEIGHT SPACE STRUCTURE OF THE
COMMITTEE MACHINE
We consider a fully-connected committee machine with N input units , K hidden
units and one output unit , where weights between the hidden units and the output
unit are set to 1. The network maps input vectors {xf}, where J1. = 1, ... , P , to
output yPo as:
(1)
where Wji is the weight between the ith input node and the jth hidden unit.
hj ~ sgn(E~l WjiXn is the jth component of the internal representation for
input pattern {xn . We consider continuous weights with spherical constraint ,
EfWji=N .
Given P = aN patterns, the learning process in a layered neural network can be
interpreted as the selection of cells in the weight space corresponding to a set of
suitable internal representations h = {hj}, each of which has a non-zero elementary
Y. Xiong, C. Kwon and J-H. Oh
380
volume defined by:
(2)
where 8(x) is the Heaviside step function. The Gardner's volume VG, that is, the
volume of the weight space which satisfies the given input-output relations, can be
written as the sum of the cells over all internal representations:
(3)
VG=LVh.
h
The method developed by Monasson and his collaborators [6, 10] is based on analysis
of the detailed internal structure, that is, how the Gardner's volume VG is decomposed into elementary volumes Vb associated with a possible internal representation.
The distribution of the elementary volumes can be derived from the free energy,
(4)
where ((- . -?) denotes the average over patterns. The entropy N[w(r)] of the volumes
whose average sizes are equal to w (r) = -1/ N In (( Vb)), can be given by the Legendre
relations
og(r)
N[w(r)]
= - o(l/r)'
w(r)
=
=
= o[rg(r)]
or
(5)
respecti vely.
=
=
The entropies ND
N[w(r
1)] and NR
N[w(r
0)) are of most importance, and will be discussed below. In the thermodynamic limit, 1/ N ((In(VG)))
-g(r = 1) is dominated by elementary volumes of size w(r = 1), of which there
are exp(N ND). Furthermore, the most numerous elementary volumes have the size
w(r = 0) and number exp(NNR). The vanishing condition for the entropies is related to the zero volume condition for VG and thus gives the storage capacity. We
focus on the entropy N D of elementary volumes dominating the weight space VG.
3
=
ORDER PARAMETERS AND PHASE TRANSITION
For a fully-connected machine, the overlaps between different hidden units should
be taken into account, which makes this problem much more difficult than the treelike (NRF) architecture studied in Ref. [10] . The replicated partition function for
the fully-connected committee machine reads:
(( (~ Vh)")) = ((
Thh;?ThWj?
IT e (It hr") ,n. e (hr" ~ Wj~.X;) )),
(6)
with a = 1,???, rand 0' = 1, ? ??, n. Unlike Gardner's conventional approach, we
need two sets of replica indices for the weights. We introduce the order parameters,
(7)
where the indices a, b originate from the integer power r of elementary volumes,
and 0', {3 are the standard replica indices. The replica symmetry ansatz leads to five
The Storage Capacity of a Fully-Connected Committee Machine
order parameters as:
I
q'"
QOI{3ab
'k
_
-
J
q
C
d'"
d
=
381
= (3, a =1= b),
(j k, 0:
(j = k, 0: =1=
(j =1= k,o: =
(j =1= k,o:=
(j =1= k, 0: =1=
(3),
(3,a = b),
(3,a =1= b),
(3),
(8)
where q'" and q are, respectively, the overlaps between the weight vectors connected
to the same hidden unit of the same (0: = (3) and different (0: =1= (3) replicas corresponding to the two different internal representations. The order parameters c, d'"
and d describe the overlaps between weights that are connected to different hidden units, of which c and d'" are the overlaps within the same replica whereas d
correlates different replicas.
Using a standard replica trick, we obtain the explicit form of g( r). One may notice
that the free energy evaluated at r 1 is reduced to the RS results obtained by the
conventional method on the committee machine[4, 5], which is independent of q'"
and d"'. This means that the internal structure of the weight space is overlooked by
conventional calculation of the Gardner's volume. When we take the limit r ~ 1,
the free energy can be expanded as:
=
d'" d)
(
d)
( '"
gr,q,q,c,
, =gl,q,c,
Ir=l'
+ (r-l )og(r,q"',q,c,d"',d)
or
(9)
As noticed, g(r, q"', q, c, d"', d) is the same as the RS free energy in the Gardner's
method. From the relation:
N: _
D -
-
og(r)
o(l/r)
I _og(r) I
r=l -
-a;:-
r=l '
(10)
we obtain the explicit form of N D .
In the case of the NRF committee machine, where each of the hidden units is
connected to different input units, we do not have a phase transition. Instead, a
single solution is applicable for the whole range of 0:. In contrast, the phase-space
structure of the fully-connected committee machine is more complicated than that
of the NRF committee machine. When a small number of input patterns are given,
the system is in the permutation-symmetry (PS) phase[4, 5, 14), where the role of
each hidden unit is not specialized. In the PS phase, the Gardner's volume is a single
connected region. The order parameters associated with different hidden units are
equal to the corresponding ones associated with the same hidden unit. When a
critical number of patterns is given, the Gardner's volume is divided into many
islands, each one of which can be transformed into other ones by permutation of
hidden units. This phenomenon is called permutation symmetry breaking (PSB),
and is usually accompanied by a first-order phase transition. In the PSB phase,
the role of each hidden unit is specialized to store a larger number of patterns
effectively. A similar breaking of symmetry has been observed in the study of
generalization[14, 15], where the first-order phase transition induces discontinuity of
the learning curve. It was pointed out that the critical storage capacity is attained in
the PSB phase[4, 5), and our recent one-step replica symmetry breaking calculation
confirmed this picture[13) . Therefore, we will focus on the analysis of the PSB
solution near the storage capacity, in which q*, q ~ 1, and c, d"', d are of order
11K.
Y. Xiong. C. Kwon and J-H. Oh
382
4
STORAGE CAPACITY
When we analyze the results for free energy, the case with q( r = 1), c( r = 1) and
d(r = 1) is reduced to the usual saddle-point solutions of the replica symmetric
expression of the Gardner's volume g(r = 1)[4, 5). When K is large, the trace
over all allowed internal representations can be evaluated similarly to Ref.[4]. The
saddle-point equations for q* and d* are derived from the derivative of the free
energy in the limit r -+ 1, as in Eq. (9). The details of the self-consistent equations
are not shown for space consideration. In the following, we only summarize the
asymptotic behavior of the order parameters for large a:
128
K2
1 - q + d - c '" (1r _ 2)2 0'2 '
"
32 ]{
(11)
1 - q + (Ii - 1)( c - d) '" 1r _ 2 a 2 '
( 12)
1r-2
q + (K - l)d '" - a '
(13)
(14)
1 - q*
where
+ (I{ -
r = -[..j1r J duH(u)lnH(u)]-l
1r 2
r2
l)(c - d*) '" 20'2 '
~
(15)
0.62.
It is found that all the overlaps between weights connecting different hidden units
have scaling of -1/ K, whereas the typical overlaps between weights connecting the
same hidden unit approach one. The order parameters c, d and d* are negative,
showing antiferromagnetic correlations between different hidden units, which implies that each hidden unit attempts to store patterns different from those of the
others[4, 5).
Finally, the asymptotic behavior of the entropy N D in the large K limit can be
derived using the scaling given above. Near the storage capacity, ND can be written,
up to the leading order, as:
"I r
N D '"
- K n '\
-
(1r - 2)20'2
256K
(16)
Being the entropy of a discrete system, ND cannot be negative. Therefore,
a gives an indication of the upper bound of storage capacity, that is,
a c '" 7r~2 K Vln K. The storage capacity per synapse, 7r~2 Vln K, satisfies the
rigorous bound", In K derived by Mitchison and Durbin (MD)[3], whereas the
conventional RS result[4, 5], which scales as .JR, violates the MD bound.
ND =
5
DISCUSSIONS
Recently, we have studied this problem using a conventional Gardner approach in
the one-step RSB scheme[13]. The result yields the same scaling with respect to K,
but a coefficient smaller by a factor v'2. In the present paper, we are dealing with
the fine structure of version space related to internal representations. On the other
hand, the RSB calculation seems to handle this fine structure in association with
symmetry breaking between replicas. Although the physics of the two approaches
seems to be somehow related, it is not clear which of the two can yield a better
The Storage Capacity of a Fully-Connected Committee Machine
383
estimate of the storage capacity. It is possible that the present RS calculation does
not properly handle the RSB picture of the system. Monasson and his co-workers
reported that the Almeida-Thouless instability of the RS solutions decreases with
increasing K, in the NRF case[1O, 11] . A similar analysis for the fully-connected case
certainly deserves further research. On the other hand, the one-step RSB scheme
also introduces approximation, and possibly it cannot fully explain the weight-space
structure associated with internal representations.
It is interesting to compare our result with that of the NRF committee machine
along the same lines[10]. Based on the conventional RS calculation, Angel et al.
suggested that the same storage capacity per synapse for both fully-connected and
NRF committee machines will be similar, as the overlap between the hidden nodes
approaches zero.[5] . While the asymptotic scaling with respect to K is the same,
the storage capacity in the fully-connected committee machine is larger than in the
NRF one. It is also consistent with our result from one-step RSB calculation[13].
This implies that the small, but nonzero negative correlation between the weights
associated with different hidden units, enhances the storage capacity. This may
be good news for those people using a fully connected multi-layer perceptron in
applications.
From the fact that the storage capacity of the NRF parity machine is In Kj In 2[2,
10], which saturates the MD bound, one may guess that the storage capacity of a
fully-connected parity machine is also proportional to Kin K. It will be interesting
to check whether the storage capacity per synapse of the fully-connected parity
machine is also enhanced compared to the NRF machine[16].
Acknowledgements
This work was partially supported by the Basic Science Special Program of
POSTECH and the Korea Ministry of Education through the POSTECH Basic
Science Research Institute(Grant No. BSRI-96-2438). It was also supported by
non-directed fund from Korea Research Foundation, 1995, and by KOSEF grant
971-0202-010-2.
References
[1] E. Gardner, Europhys, Lett. 4(4), 481 (1987); E. Gardner, J. Phys. A21, 257
(1988); E. Gardner and B. Derrida, J . Phys. A21, 271 (1988).
[2] E. Barkai, D. Hansel and 1. Kanter, Phys. Rev. Lett. V 65, N18, 2312 (1990).
[3] G. J. Mitchison and R. M. Durbin, Boil. Cybern. 60,345 (1989).
[4] E. Barkai, D. Hansel and H. Sompolinsky, Phys. Rev. E45, 4146 (1992) .
[5] A. Engel, H. M. Kohler, F. Tschepke, H. Vollmayr, and A. Zippeelius, Phys .
Rev. E45, 7590 (1992).
[6] R. Monasson and D. O'Kane, Europhys. Lett. 27,85(1994).
[7] B. Derrida, R. B. Griffiths and A Prugel-Bennett, J. Phys. A 24,4907 (1991).
[8] M. Biehl and M. Opper, Neural Networks: The Statistical Mechanics Perspective, Jong-Hoon Oh, Chulan Kwon, and Sungzoon Cho (eds.) (World Scientific,
Singapore, 1995).
[9] A. Engel and M. Weigt, Phys. Rev. E53, R2064 (1996).
[10] R. Monasson and R. Zecchina, Phys. Rev. Lett. 75, 2432 (1995); 76, 2205
(1996).
384
Y. Xiong, C. Kwon and J-H. Oh
R. Monasson and R. Zecchina, Mod. Phys. B, Vol. 9 , 1887-1897 (1996) .
S. Cocco, R. Monasson and R. Zecchina, Phys. Rev . E54, 717 (1996) .
C. Kwon and J. H. Oh, J . Phys. A, in press.
K. Kang, J. H. Oh, C. Kwon and Y. Park, Phys. Rev. E48, 4805 (1993).
[15] H. Schwarze and J. Hertz, Europhys. Lett. 21, 785 (1993) .
[16] Y. Xiong, C. Kwon and J .-H. Oh, to be published (1997).
[11]
[12]
[13]
[14]
| 1432 |@word version:1 briefly:1 seems:2 nd:5 r:8 lnh:1 orf:1 com:1 written:2 numerical:1 partition:1 j1:1 fund:1 guess:1 ith:1 vanishing:1 node:3 five:1 mathematical:2 along:1 introduce:1 angel:1 behavior:3 frequently:1 mechanic:4 multi:2 spherical:1 decomposed:1 increasing:1 mountain:1 interpreted:1 developed:1 zecchina:4 k2:1 unit:24 grant:2 limit:7 analyzing:1 studied:4 kane:2 co:1 range:1 directed:1 bell:2 griffith:1 cannot:2 layered:1 selection:1 storage:25 cybern:1 instability:1 conventional:9 map:1 oh:8 his:2 handle:2 nnr:1 enhanced:1 trick:1 particularly:1 hyoja:1 observed:1 role:2 prefactor:1 wj:1 region:1 connected:21 news:1 sompolinsky:2 decrease:1 derivation:1 describe:1 cocco:1 europhys:3 kanter:2 whose:1 larger:2 dominating:1 biehl:1 indication:1 relevant:1 p:2 ac:2 derrida:2 eq:1 implies:2 subsequently:1 sgn:1 violates:3 education:1 generalization:2 collaborator:1 elementary:7 exp:2 applicable:1 hansel:4 e45:2 agrees:2 engel:3 successfully:1 hj:2 og:4 derived:5 focus:2 properly:1 check:1 contrast:1 ave:1 rigorous:1 treelike:1 hidden:22 relation:3 jln:1 transformed:1 respecti:1 special:1 field:3 equal:2 nrf:10 park:1 others:1 kwon:8 thouless:1 phase:9 ab:1 attempt:1 interest:1 certainly:1 introduces:1 closer:1 worker:1 hoon:2 korea:4 vely:1 formalism:4 deserves:1 examining:1 gr:1 reported:1 cho:1 physic:3 ansatz:1 together:1 connecting:2 possibly:1 derivative:1 leading:2 account:1 accompanied:1 coefficient:1 lab:1 analyze:2 complicated:1 ir:1 yield:3 confirmed:1 weigt:1 published:1 explain:1 phys:12 ed:1 energy:6 associated:5 boil:1 feed:1 attained:1 rand:1 synapse:3 formulation:1 evaluated:2 though:1 furthermore:1 just:1 correlation:2 hand:2 prugel:1 overlapping:3 somehow:1 schwarze:1 scientific:1 barkai:4 lvh:1 read:1 symmetric:3 laboratory:1 nonzero:1 postech:3 self:1 hill:1 geometrical:1 consideration:1 recently:2 specialized:2 ji:1 volume:16 discussed:1 association:1 similarly:1 pointed:1 recent:2 perspective:1 store:2 wji:1 ministry:1 ii:1 thermodynamic:1 xf:1 calculation:7 believed:1 divided:1 basic:2 multilayer:2 cell:2 whereas:4 fine:2 unlike:1 mod:1 integer:1 near:2 architecture:3 whether:1 expression:1 effort:1 detailed:1 clear:1 induces:1 reduced:2 psb:4 singapore:1 notice:1 per:3 discrete:1 vol:1 pohang:2 replica:13 sum:1 letter:1 j1r:1 scaling:4 vb:2 bound:9 layer:4 durbin:4 constraint:1 dominated:1 expanded:1 department:2 legendre:1 hertz:1 jr:1 smaller:1 island:1 rev:7 taken:1 equation:2 previously:1 discus:1 committee:19 apply:1 xiong:5 monasson:8 denotes:1 murray:1 noticed:1 receptive:3 md:6 nr:1 usual:1 enhances:1 duh:1 capacity:25 originate:1 index:3 difficult:1 trace:1 negative:3 upper:1 finite:1 saturates:1 overlooked:1 kang:1 discontinuity:1 suggested:2 below:1 pattern:7 usually:1 summarize:1 pioneering:1 program:1 rsb:8 power:1 critical:3 suitable:1 overlap:7 hr:2 scheme:3 technology:2 numerous:2 picture:2 gardner:16 kj:1 vh:1 acknowledgement:1 asymptotic:4 fully:18 permutation:3 interesting:3 proportional:2 vg:6 foundation:1 consistent:2 lo:2 gl:1 parity:6 free:6 supported:2 jth:2 perceptron:3 institute:2 curve:1 lett:5 xn:1 world:2 transition:4 opper:1 forward:1 san:1 replicated:1 correlate:1 dealing:1 mitchison:4 continuous:2 symmetry:7 whole:1 allowed:1 ref:2 vln:2 explicit:2 a21:2 breaking:5 kin:1 lucent:1 showing:1 r2:1 effectively:1 kr:2 importance:1 entropy:6 rg:1 saddle:2 partially:1 antiferromagnetic:1 satisfies:4 bennett:1 onestep:1 typical:1 called:2 perceptrons:1 jong:2 internal:12 almeida:1 people:1 kohler:1 heaviside:1 phenomenon:1 |
475 | 1,433 | A Generic Approach for Identification of
Event Related Brain Potentials via a
Competitive Neural Network Structure
Daniel H. Lange
Department of Electrical Engineering
Technion - liT
Haifa 32000
Israel
e-mail: lange@turbo.technion.ac.il
Hillel Pratt
Evoked Potential Laboratory
Technion - liT
Haifa 32000
Israel
e-mail: hillel@tx.technion.ac.il
Hava T. Siegelmann
Department of Industrial Engineering
Technion - liT
Haifa 32000
Israel
e-mail: iehava@ie.technion.ac.il
Gideon F. Inbar
Department of Electrical Engineering
Technion - liT
Haifa 32000
Israel
e-mail: inbar@ee.technion.ac.il
Abstract
We present a novel generic approach to the problem of Event Related
Potential identification and classification, based on a competitive Neural Net architecture. The network weights converge to the embedded
signal patterns, resulting in the formation of a matched filter bank.
The network performance is analyzed via a simulation study, exploring
identification robustness under low SNR conditions and compared to
the expected performance from an information theoretic perspective.
The classifier is applied to real event-related potential data recorded
during a classic odd-ball type paradigm; for the first time, withinsession variable signal patterns are automatically identified, dismissing the strong and limiting requirement of a-priori stimulus-related
selective grouping of the recorded data.
D. H. Lange, H. T. Siegelmann, H. Pratt and G. F. Inbar
902
1
1.1
INTRODUCTION
EVENT RELATED POTENTIALS
Ever since Hans Berger's discovery that the electrical activity of the brain can be
measured and recorded via surface electrodes mounted on the scalp, there has been
major interest in the relationship between such recordings and brain function. The
first recordings were concerned with the spontaneous electrical activity of the brain,
appearing in the form of rhythmic voltage oscillations, which later received the term
electroencephalogram or EEG. Subsequently, more recent research has concentrated on
time-locked brain activity, related to specific events, external or internal to the subject.
This time-locked activity, referred to also as Event Related Potentials (ERP's), is
regarded as a manifestation of brain processes related to preparation for or in response
to discrete events meaningful to the subject.
The ongoing electrical activity of the brain, the EEG, is comprised of relatively slow
fluctuations, in the range of 0.1 - 100 Hz, with magnitudes of 10 - 100 uV. ERP's
are characterized by overlapping spectra with the EEG, but with significantly lower
magnitudes of 0.1 - 10 uV. The unfavorable Signal to Noise Ratio (SNR) requires
filtering of the raw signals to enable analysis of the time-locked signals. The common
method used for this purpose is signal averaging, synchronized to repeated occurrences
of a specific event. Averaging-based techniques assume a deterministic signal within
the averaged session, and thus signal variability can not be modeled unless a-priori
stimulus- or response-based categorization is available; it is the purpose of this paper to
provide an alternative working method to enhance conventional averaging techniques,
and thus facilitating identification and analysis of variable brain responses.
1.2
COMPETITIVE LEARNING
Competitive learning is a well-known branch of the general unsupervised learning
theme. The elementary principles of competitive learning are (Rumelhart & Zipser,
1985): (a) start with a set of units that are all the same except for some randomly
distributed parameter which makes each of them respond slightly differently to a set
of input patterns, (b) limit the strength of each unit, and (c) allow the units to compete in some way for the right to respond to a given subset of inputs. Applying these
three principles yields a learning paradigm where individual units learn to specialize
on sets of similar patterns and thus become feature detectors. Competitive learning
is a mechanism well-suited for regularity detection (H aykin , 1994), where there is a
popUlation of input patterns each of which is presented with some probability. The
detector is supposed to discover statistically salient features of the input population,
without a-priori categorization into which the patterns are to be classified. Thus the
detector needs to develop its own featural representation of the population of input
patterns capturing its most salient features.
1.3
PROBLEM STATEMENT
The complicated, generally unknown relationships between the stimulus and its associated brain response, and the extremely low SNR of the brain responses which are
practically masked by the background brain activity, make the choice of a self organizing structure for post-stimulus epoch analysis most appropriate. The competitive
network, having the property that its weights converge to the actual embedded signal patterns while inherently averaging out the additive background EEG, is thus an
evident choice.
A Generic Approach for Identification of Event Related Brain Potentials
2
2.1
903
THE COMPETITIVE NEURAL NETWORK
THEORY
The common architecture of a competitive learning system appears in Fig. 1. The
system consists of a set of hierarchically layered neurons in which each layer is connected
via excitatory connections with the following layer. Within a layer, the neurons are
divided into sets of inhibitory clusters in which all neurons within a cluster inhibit
all other neurons in the cluster, which results in a competition among the neurons to
respond to the pattern appearing on the previous layer.
Let Wji denote the synaptic weight connecting input node i to neuron j. A neuron
learns by shifting synaptic weights from its inactive to active input nodes. If a neuron
does not respond to some input pattern, no learning occurs in that neuron. When a
single neuron wins the competition, each of its input nodes gives up some proportion of
its synaptic weight, which is distributed equally among the active input nodes, fulfilling:
2:i Wji = 1. According to the standard competitive learning rule, for a winning neuron
to an input vector Xi, the change llWji is defined by: llWji = 7J(Xi - Wji), where
7J is a learning rate coefficient. The effect of this rule is that the synaptic weights
of a winning neuron are shifted towards the input pattern; thus assuming zero-mean
additive background EEG, once converged, the network operates as a matched filter
bank classifier.
2.2
MATCHED FILTERING
From an information theoretic perspective, once the network has converged, our classification problem coincides with the general detection problem of known signals in
additive noise. For simplicity, we shall limit the discussion to the binary decision
problem of a known signal in additive white Gaussian noise, expandable to the M-ary
detection in colored noise (Van Trees, 1968).
Adopting the common assumption of EEG and ERP additivity (Gevins, 1984), and
distinct signal categories, the competitive NN weights inherently converge to the general
signal patterns embedded within the background brain activity; therefore the converged
network operates as a matched filter bank. Assuming the simplest binary decision
problem, the received signal under one hypothesis consists of a completely known signal,
VEs(t), representing the EP, corrupted by an additive zero-mean Gaussian noise w(t)
with variance (72; the received signal under the other hypothesis consists of the noise
w(t) alone. Thus:
= wet),
0 $ t $ T
HI: ret) = ../Es(t) + wet),
0$ t $ T
For convenience we assume that JoT s2(t)dt = 1, so that E represents the signal energy.
Ho: ret)
The problem is to observe r(t) over the interval [0, T] and decide whether Ho or Hl
is true. It can be shown that the matched filter is the optimal detector, its impulse
response being simply the signal reversed in time and shifted:
her) = s(T - r)
(1)
Assuming that there is no a-priori knowledge of the probability of signal presence, the
total probability of error depends only on the SNR and is given by (Van Trees, 1968):
Pe
1 fOO
= Vr.c
exp( 21r
IJi:
V-;;2
~2 )dz
2
(2)
Fig. 2 presents the probability of true detection: (a) as a function of SNR, for minimized
error probability, and (b) as a function of the probability of false detection. These
D. H. Lange, H. T. Siegelmann, H. Pratt and G. F. Inbar
904
results are applicable to our detection problem assuming approximate Gaussian EEG
characteristics (Gersch, 1970), or optimally by using a pre-whitening approach (Lange
et. al., 1997).
Probability of True Detection with Minimum Error
J~ r
L ll yer 1
Illhlb," n ry C lu l le rs
1.xIllIlOry
("" "nUn OIl S
Layer :!
Inhibi to ry CJ u~ lcn.
).K llillory
('Ollonec.u ons
? 0 00
0
0 00 0 0 .
o ? oooo ~
0
.0
0
.
1
}!
20dB
~O.B
0
~0.6
0 ?
tt
"0
~04
~
0.2
?
0.1
INPUT PATffiRN
Figure 1: The architecture of a competitive learning structure: learning takes
place in hierarchically layered units,
presented as filled (active) and empty
(inactive) dots .
2.3
SNR In dB
Detection performance : SNA "" +2 0 ? ? ,0. 0 , - 10 and -20 dB
~ F,~~~;;~::~====~~~~~~~~;:~
9
1??lyel I
I" pull lnils.
: ? : :I
~?~0----~3~
0 ----~
20~---~
'0~--~
0 --~1~
0 --~2~
0 --~3~
0 --~40
0 .2
0 .3
0.4
05
O.B
0 .7
Probability of False Detection
0 ,8
Figure 2: Detection performance. Top:
probability of detection as a function of
the SNR. Bottom: detection characteristics.
NETWORK TRAINING AND CONVERGENCE
Our net includes a 300-node input layer and a competitive layer consisting of singlelayered competing neurons. The network weights are initialized with random values
and trained with the standard competitive learning rule, applied to the normalized
input vectors:
z?
AWji = 77( ~ - Wji)
(3)
L.Ji
Xi
The training is applied to the winning neuron of each epoch, while increasing the bias of
the frequently winning neuron to gradually reduce its chance of winning consecutively
(eliminating the dead neuron effect (Freeman & Skapura, 1992)). Symmetrically, its
bias is reduced with the winnings of other neurons.
In order to evaluate the network performance, we explore its convergence by analyzing
the learning process via the continuously adapting weights:
pj(n)
~ J~!lWl.
; ;~
1,2, .??,C
(4)
where C represents the pre-defined number of categories. We define a set of classification confidence coefficients of the converged network:
(5)
Assuming existence of a null category, in which the measurements include only background noise (EEG), maxj{pj(N)} corresponds to the noise variance. Thus the values
of r j, the confidence coefficients, ranging from 0 to 1 (random classification to completely separated categories), indicate the reliability of classification, which breaks down
with the fall of SNR. Finally, it should be noted that an explicit statistical evaluation
of the network convergence properties can be found in (Lange, 1997).
0 .9
A Generic Approachfor Identification of Event Related Brain Potentials
2.4
905
SIMULATION STUDY
A simulation study was carried out to assess the performance of the competitive network classification system. A moving average (MA) process of order 8 (selected according to Akaike's condition applied to ongoing EEG (Gersch, 1970)), driven by a
deterministic realization of a Gaussian white noise series, simulated the ongoing background activity x(n). An average of 40 single-trials from a cognitive odd-ball type
experiment (to be explained in the Experimental Study), was used as the signal s(n).
Then, five 100-trial ensembles were synthesized, to study the classification performance
under variable SNR conditions. A sample realization and its constituents, at an SNR
of dB, is shown in Fig. 3. The simulation included embedding the signal s(n) in
the synthesized background activity x(n) at five SNR levels (-20,-10,0,+10, and +20
dB), and training the network with 750 sweeps (per SNR level). Fig. 4 shows the
convergence patterns and classification confidences of the two neurons, where it can be
seen that for SNR's lower than -10dB the classification confidence declines sharply.
?
..
Templat.,
"2100
~d
cotalu:lanoe..o ~7
i,o'z . . . ~- -~Il __ ..r..l...:.. ___ _
NOIU de5hed
RNllzallon dolted
SNR..od8
C 100
ito'
com.cMn-o g
z ... _--. __ ._----_ .. _---100
200
300
'0' .---"'''''"''"'''''''''''''''"''P''''-="----,
SNR __ 20 dB
oon""'_006
100
200
300
10-0
100
200
300
i?j):~??--20
- ,0
0
I
20
SNR
Figure 3: A sample single realization
(dotted) and its constituents (signal solid, noise - dashed). SNR = 0 dB .
Figure 4: Convergence patterns and
classification confidence values for
varying SNR levels.
The classification results, tested on 100 input vectors, 50 of each category, for each
SNR, are presented in the table below; due to the competitive scheme, Positives and
False Negatives as well as Negatives and False Positives are complementary. These
empirical results are in agreement with the analytical results presented in the above
Matched Filtering sectioll.
Table 1: Classification Results
I Pos I Neg I FP
3
3.1
I
FN
EXPERIMENTAL STUDY
MOTIVATION
An important task in ERP research is to identify effects related to cognitive processes
triggered by meaningful versus non-relevant stimuli. A common procedure to study
these effects is the classic odd-ball paradigm, where the subject is exposed to a random
906
D. H. Lange, H. T. Siegelmann, H. Pratt and G. F Inbar
sequence of stimuli and is instructed to respond only to the task-relevant (Target) ones.
Typically, the brain responses are extracted via selective averaging of the recorded data,
ensembled according to the types of related stimuli. This method of analysis assumes
that the brain responds equally to the members of each type of stimulus; however
the validity of this assumption is unknown in this case where cognition itself is being
studied. Using our proposed approach, a-priori grouping of the recorded data is not
required, thus overcoming the above severe assumption on cognitive brain function.
The results of applying our method are described below.
3.2
EXPERIMENTAL PARADIGM
Cognitive event-related potential data was acquired during an odd-ball type paradigm
from pz referenced to the mid-lower jaw, with a sample frequency of 250 Hz (Lange et.
al., 1995). The subject was exposed to repeated visual stimuli, consisting of the digits
'3' and '5', appearing on a PC screen. The subject was instructed to press a pushbutton upon the appearance of '5' - the Target stimulus, and ignore the appearances
of the digit '3'.
With odd-ball type paradigms, the Target stimulus is known to elicit a prominent
positive component in the ongoing brain activity, related to the identification of a
meaningful stimulus. This component has been labeled P300, indicating its polarity
(positive) and timing of appearance (300 ms after stimulus presentation). The parameters of the P 300 component (latency and amplitude) are used by neurophysiologists to
assess effects related to the relevance of stimulus and level of attention (Lange et. al.,
1995).
3.3
IDENTIFICATION RESULTS
The competitive network was trained with 80 input vectors, half of which were Target
ERP's and the other half were Non Target. The network converged after approximately
300 iterations (per neuron), yielding a reasonable confidence coefficient of 0.7.
A sample of two single-trial post-stimulus sweeps, of the Target and Non-Target averaged ERP templates and of the NN identified signal categories, are presented in Fig.
5. The convergence pattern is shown in Fig. 6. The automatic identification procedure
has provided two signal categories, with almost perfect matches to the stimulus-related
selective averaged signals. The obtained categorization confirms the usage of averaging
methods for this classic experiment, and thus presents an important result in itself.
4
DISCUSSION AND CONCLUSION
A generic system for identification and classification of single-trial ERP's was presented.
The simulation study demonstrated the powerful capabilities of the competitive neural
net in classifying the low amplitude signals embedded within the large background
noise. The detection performance declined rapidly for SNR's lower than -10dB, which
is in general agreement with the theoretical statistical results, where loss of significance
in detection probability is evident for SNR's lower than -20dB. Empirically, high
classification performance was maintained with SNR's of down to -10dB, yielding
confidences in the order of 0.7 or higher.
The experimental study presented an unsupervised identification and classification of
the raw data into Target and Non-Target responses, dismissing the requirement of
stimulus-related selective data grouping. The presented results indicate that the noisy
brain responses may be identified and classified objectively in cases where relevance of
907
A Generic Approachfor Identification of Event Related Brain Potentials
the stimuli is unknown or needs to be determined, e.g. in lie-detection scenarios (Lange
& Inbar, 1996), and thus open new possibilities in ERP research.
J;:s::1
~F?'S2
-20~
-20L~;:--_-;:-;;o.'-----:,
,:F?tl ,:F=l
S NR _ _ 6 dB
'0'
.~g,~g
o
05
Soc
I
0
06
Soc
,
Figure 5: Top row: sample raw Target
and Non- Target sweeps. Middle row:
Target and Non- Target ERP templates.
Bottom row: the NN categorized patterns.
10-40!;-----;:;;,--------;
200:;;----==""';;---=----;;:;;--~
~ of~.
Figure 6: Convergence pattern of the
ERP categorization process; convergence is achieved after 300 iterations
per neuron.
References
[1] Freeman J.A. and Skapura D.M. Neural Network6: Algorithm6, Application6, and Programming Technique6: Addison-Wesley Publishing Company, USA, 1992.
[2] Gersch W., "Spectral Analysis of EEG's by Autoregressive Decomposition of Time Series," Math. Bi06c., vol. 7, pp. 205-222, 1970.
[3] Gevins A.S., "Analysis of the Electromagnetic Signals of the Human Brain: Milestones,
Obstacles, and Goals," IEEE Tran6. Biomed. Eng., vol. BME-31, pp. 833-850, 1984.
[4] Haykin S. Neural Network6: A Comprehen6ive Foundation. Macmillan College Publishing
Company, Inc., USA, 1994.
[5] Lange D. H. Modeling and E6timation of Tran6ient, Evoked Brain Potential6. D.Sc. dissertation, Techion - Israel Institute of Technology, 1997.
[6] Lange D.H. and Inbar G.F., "Brain Wave Based Polygraphy," Proceeding6 of the IEEE
EMBS96 - the 18th Annual International Conference of the IEEE Engineering on
Medicine and Biology Society, Amsterdam, October 1996.
[7] Lange D.H., Pratt H. and Inbar G.F., "Modeling and Estimation of Single Evoked Brain
Potential Components", IEEE. Tran6. Biomed. Eng., vol. BME-44, pp. 791-799, 1997.
[8] Lange D.H., Pratt H., and Inbar G.F., "Segmented Matched Filtering of Single Event
Related Evoked Potentials," IEEE. Tran6. Biomed. Eng., vol. BME-42, pp. 317-321,
1995.
[9] Rumelhart D.E. and Zipser D., "Feature Discovery by Competitive Learning," Cognitive
Science, vol. 9, pp. 75-112, 1985.
[10] Van Trees H.L. Detection, E6timation, and Modulation Theory: Part 1: John Wiley and
Sons, Inc., USA, 1968.
| 1433 |@word trial:4 middle:1 eliminating:1 proportion:1 open:1 confirms:1 simulation:5 r:1 decomposition:1 eng:3 solid:1 series:2 daniel:1 com:1 john:1 fn:1 additive:5 alone:1 half:2 selected:1 dissertation:1 haykin:1 colored:1 math:1 node:5 five:2 become:1 specialize:1 consists:3 acquired:1 expected:1 frequently:1 ry:2 brain:24 freeman:2 automatically:1 company:2 actual:1 increasing:1 provided:1 discover:1 matched:7 proceeding6:1 null:1 israel:5 ret:2 classifier:2 milestone:1 unit:5 positive:4 engineering:4 referenced:1 timing:1 limit:2 analyzing:1 gevins:2 fluctuation:1 modulation:1 approximately:1 studied:1 evoked:4 locked:3 range:1 averaged:3 statistically:1 digit:2 procedure:2 empirical:1 elicit:1 significantly:1 adapting:1 network6:2 pre:2 confidence:7 convenience:1 layered:2 applying:2 conventional:1 deterministic:2 demonstrated:1 dz:1 attention:1 simplicity:1 rule:3 regarded:1 pull:1 population:3 classic:3 embedding:1 limiting:1 spontaneous:1 target:13 programming:1 akaike:1 hypothesis:2 agreement:2 rumelhart:2 labeled:1 ep:1 bottom:2 electrical:5 connected:1 inhibit:1 tran6:3 trained:2 exposed:2 upon:1 completely:2 po:1 differently:1 tx:1 additivity:1 separated:1 distinct:1 sc:1 formation:1 hillel:2 objectively:1 itself:2 noisy:1 triggered:1 sequence:1 net:3 analytical:1 relevant:2 realization:3 rapidly:1 p300:1 organizing:1 supposed:1 competition:2 inhibi:1 constituent:2 convergence:8 cluster:3 regularity:1 electrode:1 requirement:2 empty:1 categorization:4 perfect:1 sna:1 develop:1 ac:4 bme:3 measured:1 odd:5 received:3 strong:1 soc:2 indicate:2 synchronized:1 filter:4 subsequently:1 consecutively:1 human:1 enable:1 ensembled:1 electromagnetic:1 elementary:1 exploring:1 practically:1 exp:1 cognition:1 major:1 purpose:2 estimation:1 applicable:1 wet:2 gaussian:4 varying:1 voltage:1 industrial:1 nn:3 typically:1 her:1 selective:4 biomed:3 classification:15 among:2 priori:5 once:2 having:1 biology:1 represents:2 lit:4 unsupervised:2 minimized:1 stimulus:18 randomly:1 individual:1 maxj:1 consisting:2 detection:16 interest:1 possibility:1 evaluation:1 severe:1 analyzed:1 yielding:2 pc:1 unless:1 tree:3 filled:1 initialized:1 haifa:4 theoretical:1 modeling:2 obstacle:1 subset:1 snr:22 masked:1 technion:8 comprised:1 optimally:1 corrupted:1 international:1 ie:1 enhance:1 connecting:1 continuously:1 recorded:5 external:1 dead:1 cognitive:5 potential:12 includes:1 coefficient:4 inc:2 depends:1 later:1 break:1 competitive:19 start:1 wave:1 complicated:1 capability:1 ass:2 il:5 variance:2 characteristic:2 ensemble:1 yield:1 expandable:1 identify:1 identification:12 raw:3 lu:1 ary:1 classified:2 converged:5 detector:4 synaptic:4 hava:1 energy:1 frequency:1 pp:5 associated:1 iji:1 knowledge:1 oooo:1 cj:1 amplitude:2 appears:1 wesley:1 higher:1 dt:1 response:9 working:1 overlapping:1 impulse:1 usage:1 effect:5 validity:1 normalized:1 true:3 oil:1 usa:3 laboratory:1 white:2 ll:1 during:2 self:1 maintained:1 noted:1 coincides:1 m:1 manifestation:1 prominent:1 evident:2 theoretic:2 tt:1 electroencephalogram:1 ranging:1 lwl:1 novel:1 common:4 ji:1 empirically:1 synthesized:2 measurement:1 automatic:1 uv:2 session:1 dot:1 reliability:1 moving:1 han:1 surface:1 whitening:1 own:1 recent:1 perspective:2 driven:1 scenario:1 binary:2 wji:4 neg:1 seen:1 minimum:1 converge:3 paradigm:6 dashed:1 signal:27 branch:1 segmented:1 match:1 characterized:1 divided:1 post:2 equally:2 iteration:2 adopting:1 achieved:1 background:8 interval:1 lcn:1 inbar:9 recording:2 subject:5 hz:2 db:12 member:1 ee:1 zipser:2 presence:1 symmetrically:1 pratt:6 concerned:1 architecture:3 identified:3 competing:1 lange:14 reduce:1 decline:1 inactive:2 whether:1 f:1 generally:1 latency:1 mid:1 concentrated:1 category:7 simplest:1 reduced:1 inhibitory:1 shifted:2 dotted:1 per:3 discrete:1 shall:1 vol:5 salient:2 erp:10 pj:2 compete:1 powerful:1 respond:5 place:1 almost:1 reasonable:1 decide:1 oscillation:1 decision:2 capturing:1 layer:7 hi:1 turbo:1 activity:10 scalp:1 strength:1 annual:1 dismissing:2 sharply:1 extremely:1 relatively:1 jaw:1 department:3 according:3 ball:5 gersch:3 slightly:1 son:1 hl:1 explained:1 gradually:1 fulfilling:1 mechanism:1 addison:1 available:1 observe:1 appropriate:1 generic:6 spectral:1 appearing:3 occurrence:1 alternative:1 robustness:1 ho:2 existence:1 top:2 assumes:1 include:1 publishing:2 medicine:1 siegelmann:4 society:1 sweep:3 occurs:1 responds:1 nr:1 win:1 reversed:1 simulated:1 mail:4 assuming:5 modeled:1 berger:1 relationship:2 ratio:1 polarity:1 october:1 statement:1 negative:2 unknown:3 neuron:20 ever:1 variability:1 overcoming:1 required:1 connection:1 below:2 pattern:17 fp:1 gideon:1 shifting:1 event:13 representing:1 scheme:1 technology:1 declined:1 carried:1 featural:1 epoch:2 discovery:2 embedded:4 loss:1 mounted:1 filtering:4 versus:1 foundation:1 principle:2 bank:3 classifying:1 row:3 excitatory:1 bias:2 allow:1 institute:1 fall:1 template:2 rhythmic:1 jot:1 distributed:2 van:3 autoregressive:1 instructed:2 approximate:1 ignore:1 ons:1 active:3 xi:3 spectrum:1 table:2 learn:1 inherently:2 eeg:10 cmn:1 significance:1 hierarchically:2 s2:2 noise:11 motivation:1 repeated:2 facilitating:1 complementary:1 categorized:1 fig:6 referred:1 tl:1 screen:1 slow:1 vr:1 foo:1 wiley:1 theme:1 explicit:1 winning:6 lie:1 pe:1 ito:1 learns:1 down:2 specific:2 pz:1 grouping:3 false:4 magnitude:2 yer:1 suited:1 simply:1 explore:1 appearance:3 visual:1 amsterdam:1 macmillan:1 corresponds:1 chance:1 extracted:1 ma:1 goal:1 presentation:1 towards:1 change:1 included:1 neurophysiologists:1 except:1 operates:2 determined:1 averaging:6 oon:1 total:1 e:1 experimental:4 unfavorable:1 meaningful:3 indicating:1 college:1 internal:1 relevance:2 preparation:1 ongoing:4 evaluate:1 tested:1 |
476 | 1,434 | Hybrid reinforcement learning and its
application to biped robot control
Satoshi Yamada, Akira Watanabe, M:ichio Nakashima
{yamada, watanabe, naka}~bio.crl.melco.co.jp
Advanced Technology R&D Center
Mitsubishi Electric Corporation
Amagasaki, Hyogo 661-0001, Japan
Abstract
A learning system composed of linear control modules, reinforcement learning modules and selection modules (a hybrid reinforcement learning system) is proposed for the fast learning of real-world
control problems. The selection modules choose one appropriate
control module dependent on the state. This hybrid learning system was applied to the control of a stilt-type biped robot. It learned
the control on a sloped floor more quickly than the usual reinforcement learning because it did not need to learn the control on a
flat floor, where the linear control module can control the robot.
When it was trained by a 2-step learning (during the first learning
step, the selection module was trained by a training procedure controlled only by the linear controller), it learned the control more
quickly. The average number of trials (about 50) is so small that
the learning system is applicable to real robot control.
1
Introduction
Reinforcement learning has the ability to solve general control problems because it
learns behavior through trial-and-error interactions with a dynamic environment.
It has been applied to many problems, e.g., pole-balance [1], back-gammon [2],
manipulator [3], and biped robot [4]. However, reinforcement learning has rarely
been applied to real robot control because it requires too many trials to learn the
control even for simple problems.
For the fast learning of real-world control problems, we propose a new learning system which is a combination of a known controller and reinforcement learning. It is
called the hybrid reinforcement learning system. One example of a known controller
is a linear controller obtained by linear approximation. The hybrid learning system
S. Yamada, A. Watanabe and M. Nakashima
1072
will learn the control more quickly than usual reinforcement learning because it does
not need to learn the control in the state where the known controller can control
the object.
A stilt-type biped walking robot was used to test the hybrid reinforcement learning
system. A real robot walked stably on a flat floor when controlled by a linear
controller [5]. Robot motions could be approximated by linear differential equations.
In this study, we will describe hybrid reinforcement learning of the control of the
biped robot model on a sloped floor, where the linear controller cannot control the
robot.
2
Biped Robot
a)
b)
pitch axis
Figure 1: Stilt-type biped robot. a) a photograph of a real biped robot, b) a model
structure of the biped robot. Ul, U2, U3 denote torques.
Figure I-a shows a stilt-type biped robot [5J. It has no knee or ankle, has 1 m
legs and weighs 33 kg. It is modeled by 3 rigid bodies as shown in Figure I-b.
By assuming that motions around a roll axis and those around a pitch axis are
independent, 5-dimensional differential equations in a single supporting phase were
obtained. Motions of the real biped robot were simulated by the combination of
these equations and conditions at a leg exchange period. If angles are approximately
zero, these equations can be approximated by linear equations. The following linear
controller is obtained from the linear equations. The biped robot will walk if the
angles of the free leg are controlled by a position-derivative (PD) controller whose
desired angles are calculated as follows:
r{J
(J+~+{3
if; -
(J
(
A
=
+ 2~
-A7) + 6
If
(1)
where~, {3, 6, and 9 are a desired angle between the body and the leg (7?), a constant
to make up a loss caused by a leg exchange (1.3?), a constant corresponding to
walking speed, and gravitational acceleration (9.8 ms- 2 ), respectively.
The linear controller controlled walking of the real biped robot on a flat floor [5].
However, it failed to control walking on a slope (Figure 2). In this study, the
objective of the learning system was to control walking on the sloped floor shown
in Figure 2-a.
Hybrid Reinforcement Learning for Biped Robot Control
a)
1073
lOem]
Oem
i
I
1m
b)
45 ?'s , -Angular
Velocity
-
-
I
2m
3m
- ---------,-,- ----,
I fall down
iJ -45 .'sO~---------------:'
IOcm b
Time(s)
1.0
Height of
Free Leg's Tip
. ______ fall down
VVy';.:?!Pft'~',....~'t:,
-2cmO
??~
Robo'
10
Time(s)
Po".: I =
-lm O
I
10
Time(s)
Figure 2: Biped robot motion on a sloped floor controlled by the linear controller.
a) a shape of a floor, b) changes in angular velocity, height of free leg's tip, and
robot position
3
Hybrid Reinforcement Learning
reinforcement: r(t)
state
inputs
71,~,iJ
linear
control
module
1-----10( decision
of k
Figure 3: Hybrid reinforcement learning system.
We propose a hybrid reinforcement learning system to learn control quickly, The
hybrid reinforcement learning system shown in Figure 3 is composed of a linear
control module, a reinforcement learning module, and a selection module. The
reinforcement learning module and the selection module select an action and a
module dependent on their respective Q-values. This learning system is similar to
the modular reinforcement learning system proposed by Tham [6] which was based
on hierarchical mixtures of the experts (HME) [7]. In the hybrid learning system ,
the selection module is trained by Q-Iearning.
To combine the reinforcement learning with the linear controller described in (1),
the ~u!put of the reinforcement learning module is set to k in the adaptable equation
for (, ( = -kiJ + 6. The angle and the angular velocity of the supporting leg at the
leg exchange period ('T], iJJi) are used as inputs. The k values are kept constant until
the next leg exchange. The reinforcement learning module is trained by "Q-sarsa"
learning [8]. Q values are calculated by CMAC neural networks [9], [10].
The Q values for action k (Q c (x, k)) and those for module s selection (Q s (x, s)) are
s. Yamada, A Watanabe and M. Nakashima
1074
calculated as follows:
L we(k, m, i, t)y(m, i, t)
Tn ,i
Q.. (x, s)
=
L w.. (s , m, i, t)y(m, i, t),
(2)
m,i
where we{k,m,i,t) and w.. {s,m,i,t) denote synaptic strengths and y{m,i,t) represents neurons' outputs in CMAC networks at time t .
Modules were selected and actions performed according to the ?-greedy policy [8]
with ? = O.
The temporal difference (TD) error for the reinforcement learning module (fe(t)) is
calculated by
sel(t)
1
0
= lin
+ Qe(x(t + l),per(t + 1)) - Qe{x{t),per{t)) sel{t) = rein
sel(t + 1) = rein
r{t) + Q.. {x(t + 1), sel(t + 1)) - Qe(x(t),per{t)), sel{t) = rein
sel{t + 1) = lin
r(t)
fe(t) =
(3)
where r{t), per{t), sel(t), lin and rein denote reinforcement signals (r{t) = -1 if
the robot falls down, 0 otherwise), performed actions, selected modules, the linear
control module and the reinforcement learning module, respectively.
TD error (ft (t)) calculated by Q.. (x, s) is considered to be a sum of TD error caused
by the reinforcement learning module and that by the selection module. TD error
(f .. {t)) used in the selection-module's learning is calculated as follows:
f .. {t)
=
=
ft(t) - fe(t)
r{t) + ,Q .. {x(t + 1), sel(t + 1)) - Q.. (x{t), sel{t)) - fe(t),
(4)
where, denotes a discount factor.
The reinforcement learning module used replacing eligibility traces {e c (k, m, i, t))
[11]. Synaptic strengths are updated as follows:
we(k, m, i, t + 1)
w.. (s,m,i ,t + 1)
ee(k, m, i, t)
=
wc{k, m, i, t) + Qefc{t)ee{k, m, i, t)/nt
{ w.. (s , m, i, t) + Q.. f .. (t)y(m , i , t)/nt s = sel(t)
otherwise
w.. (s,m,i,t)
k=per(t),y(m,i,t)=l
1
{ o
k ::f: per(t), y(m, i, t) = 1
>.ec(k, m, i, t - 1) otherwise
(5)
where Qe, Q .. , >. and nt are a learning constant for the reinforcement learning module,
that for the selection module, decay rates and the number of tHings, respectively.
In this study, the CMAC used 10 tHings. Each of the three dimensions was divided into 12 intervals. The reinforcement learning module had 5 actions (k =
0, A/2, A, 3A/2, 2A). The parameter values were Q .. = 0.2, Q e = 0.4, >. = 0.3,
, = 0.9 and 6 = 0.05. Each run consisted of a sequence of trials, where each
trial began with robot state of position=O, _5? < () < -2.5?,1 .5? < "I < 3?, cp =
()+~, 'I/J = cp+~, ( = "1+ 2?,9 = cp = "j; = iJ = ( = 0, and ended with a failure signals
indicating robot's falling down. Runs were terminated if the number of walking
steps of three consecutive trials exceeded 100. All results reported are an average
of 50 runs.
Hybrid Reinforcement Learning for Biped Robot Control
ft''"
1075
80
'"
bll
~
60
C;
~
40
o
o
Z
20
'-
-??
.
.... ???... ----...................................
.
..
..
?
..
.
o~~~~~~~~~~
o
jO
100
IjO
200
Trials
Figure 4: Learning profiles for control of walking on the sloped floor. (0) hybrid
reinforcement learning, (0) 2-step hybrid reinforcement learning, (\7) reinforcement
learning and (6) HME-type modular reinforcement learning
4
Results
Walking control on the sloped floor (Figure 2-a) was first trained by the usual reinforcement learning. The usual reinforcement learning system needed many trials
for successful termination (about 800, see Figure 4(\7)). Because the usual reinforcement learning system must learn the control for each input, it requires many
trials.
Figure 4(0) also shows the learning curve for the hybrid reinforcement learning.
The hybrid system learned the control more quickly than the usual reinforcement
learning (about 190 trials). Because it has a higher probability of succeeding on the
flat floor, it learned the control quickly. On the other hand, HME-type modular
reinforcement learning [6] required many trials to learn the control (Figure 4(6)).
45?'s ,--- - - - - - - - - - -- -------,
Angular
Vel~ity
~A AAAAAA A hAA Ab AAA AAA AAAAAAAAAA~
I
VYVV vvrW{VYvvvVYVVlJm
rv YVvVVl
() -45?,s ' - - - - - - - - - - - - - - - - - - - ' o
Time(s)
20
0:
"hfrl}f~iM?f11tfWfNWV\{\/YVVVy1
Height Oem \
.
Free Leg's Tip ~
-2cm
RObotP~i::[ ~
-1m
o
Time(s)
2
1?
20
Figure 5: Biped robot motion controlled by the network trained by the 2-step hybrid
reinforcement learning.
S. Yamada, A. Watanabe and M Nakashima
1076
In order to improve the learning rate, a 2-step learning was examined. The
2-step learning is proposed to separate the selection-module learning from the
reinforcement-learning-module learning. In the 2-step hybrid reinforcement learning, the selection module was first trained by a special training procedure in which
the robot was controlled only by the linear control module. And then the network
was trained by the hybrid reinforcement learning. The 2-step hybrid reinforcement
learning learned the control more quickly than the I-step hybrid reinforcement
learning (Figure 4(0)). The average number of trials were about 50. The hybrid
learning system may be applicable to the real biped robot.
Figure 5 shows the biped robot motion controlled by the trained network. On the
slope, the free leg's lifting was magnified irregularly (see changes in the height of
the free leg's tip of Figure 5) in order to prevent the reduction of an amplitude of
walking rhythm. On the upper flat floor, the robot was again controlled stably by
the linear control module.
a)
b)
004
~
"""
bI)
.!!
~
EO."
"2
00]
t:
.~
j
15
8
002
0.4
...t;
.5
001
:: 02
o
.2
eOc.........................~.....................~...L.....~
o
-0.03
-002
-001
0
0.01
0 .02
003
initial synaptic strength values
?003
-002
-0.01
0
001
002
003
initial synaptic strength values
Figure 6: Dependence of (a) the learning rate and (b) the selection ratio of the
linear control module on the initial synaptic strength values (wa(rein, m, i, 0)). (a)
learning rate of (0) the hybrid reinforcement learning, and (0) the 2-step hybrid
reinforcement learning. The learning rate is defined as the inverse of the number of
trials where the average walking steps exceed 70. (b) the ratio of the linear-controlmodule selection. Circles represent the selection ratio of the linear control module
when controlled by the network trained by the hybrid reinforcement learning, rectangles represent that by the 2-step hybrid reinforcement learning. Open symbols
represent the selection ratio on the flat floor, closed symbols represent that on the
slope.
The dependence of learning characteristics on initial synaptic strengths for
the reinforcement-learning-module selection (W3 (rein, m, i, 0)) was considered
(other initial synaptic strengths were 0). If initial values of ws(rein, m, i, t)
(ws(rein, m, i, 0)) are negative, the Q-values for the reinforcement-learning-module
selection (Q8(x,rein)) are smaller than Q8(x,lin) and then the linear control module is selected for all states at the beginning of the learning. In the case of the
2-step learning, if Ws (rein, m, i, 0) are given appropriate negative values, the reinforcement learning module is selected only around failure states, where Qa(x, lin) is
trained in the first learning step, and the linear control module is selected otherwise
at the beginning of the second learning step. Because the reinforcement learning
module only requires training around failure states in the above condition, the 2-
Hybrid Reinforcement Learning fo r Biped Robot Control
1077
step hybrid system is expected to learn the control quickly. Figure 6-a shows the
dependence of the learning rate on the initial synaptic strength values. The 2-step
hybrid reinforcement learning had a higher learning rate when Ws (rein, m, i, 0) were
appropriate negative values (-0.01 '" -0.005). The trained system selected the linear
control module on the flat floor (more than 80%), and selected both modules on
the slope (see Figure 6-b), when ws(rein, m, i, 0) were negative.
Three trials were required in the first learning step of the 2-step hybrid reinforcement
learning. In order to learn the Q-value function around failure states, the learning
system requires 3 trials.
5
Conclusion
We proposed the hybrid reinforcement learning which learned the biped robot control quickly. The number of trials for successful termination in the 2-step hybrid
reinforcement learning was so small that the hybrid system is applicable to the real
biped robot. Although the control of real biped robot was not learned in this study,
it is expected to be learned quickly by the 2-step hybrid reinforcement learning.
The learning system for real robot control will be easily constructed and should be
trained quickly by the hybrid reinforcement learning system.
References
[1] Barto, A. G., Sutton, R. S. and Anderson, C. W.: Neuron like adaptive elements that can solve difficult learning control problems, IEEE Trans. Sys. Man
Cybern., Vol. SMC-13, pp. 834-846 (1983).
[2] Tesauro, G.: TD-gammon, a self-teaching backgammon program, achieves
master-level play, Neural Computation, Vol. 6, pp. 215-219 (1994).
[3] Gullapalli, V., Franklin, J. A. and Benbrahim, H.: Acquiring robot skills via
reinforcement learning, IEEE Control System, Vol. 14, No.1, pp. 13-24 (1994).
[4] Miller, W. T.: Real-time neural network control of a biped walking robot, IEEE
Control Systems, Vol. 14, pp. 41-48 (1994).
[5] Watanabe, A., Inoue, M. and Yamada, S.: Development of a stilts type biped
robot stabilized by inertial sensors (in Japanese), in Proceedings of 14th Annual
Conference of RSJ, pp. 195-196 (1996).
[6] Tham, C. K.: Reinforcement learning of multiple tasks using a hierarchical
CMAC architecture, Robotics and Autonomous Systems, Vol. 15, pp. 247-274
(1995).
[7] Jordan, M. I. and Jacobs, R. A.: Hierarchical mixtures of experts and the EM
algorithm, Neural Computation, Vol. 6, pp. 181-214 (1994).
[8] Sutton, R. S.: Generalization in reinforcement learning: successful examples
using sparse coarse coding, Advances in NIPS, Vol. 8, pp. 1038-1044 (1996).
[9] Albus, J. S.: A new approach to manipulator control: The cerebellar model
articulation controller (CMAC), Transaction on ASME J. Dynamical Systems,
Measurement, and Controls, pp. 220-227 (1975).
[10] Albus, J. S.: Data storage in the cerebellar articulation controller (CMAC),
Transaction on ASME J. Dynamical Systems, Measurement, and Controls, pp.
228-233 (1975).
[11] Singh, S. P. and Sutton, R. S.: Reinforcement learning with replacing eligibility
traces, Machine Learning, Vol. 22, pp. 123-158 (1996).
| 1434 |@word trial:16 consisted:1 rsj:1 objective:1 open:1 ankle:1 termination:2 mitsubishi:1 dependence:3 jacob:1 usual:6 during:1 self:1 eligibility:2 separate:1 exchange:4 qe:4 rhythm:1 reduction:1 initial:7 m:1 generalization:1 simulated:1 asme:2 sarsa:1 franklin:1 im:1 tn:1 motion:6 cp:3 gravitational:1 sloped:6 nt:3 around:5 considered:2 hyogo:1 assuming:1 modeled:1 must:1 began:1 ratio:4 balance:1 lm:1 robo:1 difficult:1 fe:4 u3:1 shape:1 achieves:1 consecutive:1 trace:2 jp:1 succeeding:1 negative:4 greedy:1 selected:7 applicable:3 policy:1 measurement:2 upper:1 neuron:2 beginning:2 sys:1 yamada:6 supporting:2 teaching:1 coarse:1 biped:25 sensor:1 bll:1 had:2 robot:39 height:4 sel:10 constructed:1 barto:1 differential:2 required:2 combine:1 backgammon:1 tesauro:1 learned:8 expected:2 behavior:1 nip:1 qa:1 trans:1 dependent:2 rigid:1 torque:1 aaaaaaaaaa:1 dynamical:2 akira:1 floor:14 td:5 eo:1 w:5 articulation:2 period:2 program:1 signal:2 rv:1 multiple:1 benbrahim:1 kg:1 cm:1 development:1 hybrid:37 special:1 lin:5 advanced:1 divided:1 magnified:1 corporation:1 ended:1 improve:1 technology:1 temporal:1 represents:1 controlled:10 pitch:2 inoue:1 axis:3 controller:14 iearning:1 represent:4 bio:1 control:54 cerebellar:2 robotics:1 composed:2 interval:1 loss:1 phase:1 sutton:3 ab:1 approximately:1 thing:2 examined:1 jordan:1 co:1 mixture:2 ee:2 smc:1 exceed:1 bi:1 w3:1 architecture:1 free:6 loem:1 respective:1 procedure:2 fall:3 cmac:6 walk:1 desired:2 circle:1 gullapalli:1 sparse:1 weighs:1 curve:1 dimension:1 ul:1 calculated:6 gammon:2 kij:1 world:2 ijo:1 reinforcement:66 adaptive:1 cannot:1 selection:18 ec:1 action:5 put:1 aaaaaa:1 cybern:1 storage:1 pole:1 transaction:2 skill:1 center:1 successful:3 discount:1 too:1 reported:1 knee:1 stabilized:1 cmo:1 per:6 tham:2 ity:1 learn:9 autonomous:1 tip:4 vol:8 updated:1 quickly:11 play:1 jo:1 again:1 japanese:1 falling:1 electric:1 choose:1 prevent:1 did:1 velocity:3 element:1 approximated:2 kept:1 walking:11 expert:2 derivative:1 rectangle:1 terminated:1 profile:1 japan:1 sum:1 hme:3 ft:3 module:45 run:3 angle:5 coding:1 inverse:1 master:1 body:2 caused:2 amagasaki:1 position:3 performed:2 decision:1 watanabe:6 closed:1 environment:1 pd:1 walked:1 learns:1 dynamic:1 slope:4 down:4 trained:13 singh:1 annual:1 strength:8 roll:1 characteristic:1 symbol:2 miller:1 flat:7 decay:1 satoshi:1 po:1 easily:1 wc:1 speed:1 haa:1 lifting:1 fast:2 describe:1 according:1 combination:2 fo:1 photograph:1 rein:12 synaptic:8 whose:1 modular:3 failure:4 solve:2 smaller:1 em:1 pp:11 otherwise:4 aaa:2 failed:1 ability:1 naka:1 leg:13 nakashima:4 u2:1 acquiring:1 sequence:1 equation:7 propose:2 interaction:1 amplitude:1 inertial:1 needed:1 back:1 adaptable:1 irregularly:1 exceeded:1 acceleration:1 higher:2 crl:1 man:1 change:2 hierarchical:3 albus:2 vel:1 appropriate:3 called:1 anderson:1 angular:4 q8:2 until:1 hand:1 rarely:1 select:1 indicating:1 replacing:2 denotes:1 a7:1 object:1 oem:2 stably:2 ij:3 manipulator:2 |
477 | 1,435 | Incorporating Contextual Information in White
Blood Cell Identification
Xubo Song*
Department of Electrical Engineering
California Institute of Technology
Pasadena, CA 91125
xubosong@fire.work.caltech.edu
Yaser Abu-Mostafa
Dept. of Electrical Engineering
and Dept. of Computer Science
California Institute of Technology
Pasadena, CA 91125
Yaser@over. work.caltech.edu
Joseph Sill
Computation and Neural Systems Program
California Institute of Technology
Pasadena, CA 91125
joe@busy.work.caltech.edu
Harvey Kasdan
International Remote Imaging Systems
9162 Eton Ave.,
Chatsworth, CA 91311
Abstract
In this paper we propose a technique to incorporate contextual information into object classification. In the real world there are cases where the
identity of an object is ambiguous due to the noise in the measurements
based on which the classification should be made. It is helpful to reduce the ambiguity by utilizing extra information referred to as context,
which in our case is the identities of the accompanying objects. This
technique is applied to white blood cell classification. Comparisons are
made against "no context" approach, which demonstrates the superior
classification performance achieved by using context. In our particular
application, it significantly reduces false alarm rate and thus greatly reduces the cost due to expensive clinical tests.
? Author for correspondence.
Incorporating Contextual Information in White Blood Cell Identification
951
1 INTRODUCTION
One of the most common assumptions made in the study of machine learning is that the
examples are drawn independently from some joint input-output distribution. There are
cases, however, where this assumption is not valid. One application where the independence assumption does not hold is the identification of white blood cell images. Abnormal
cells are much more likely to appear in bunches than in isolation. Specifically, in a sample
of several hundred cells, it is more likely to find either no abnormal cells or many abnormal
cells than it is to find just a few.
In this paper, we present a framework for pattern classification in situations where the
independence assumption is not satisfied. In our case, the identity of an object is dependent
of the identities of the accompanying objects, which provides the contextual information.
Our method takes into consideration the joint distribution of all the classes, and uses it to
adjust the object-by-object classification.
In section 2, the framework for incorporating contextual information is presented, and an
efficient algorithm is developed. In section 3 we discuss the application area of white
blood cell classification, and address the importance of using context for this application.
Empirical testing results are shown in Section 4, followed by conclusions in Section 5.
2 INCORPORATING CONTEXTUAL INFORMATION INTO
CLASSIFICATION
2.1
THE FRAMEWORK
Let Xi be the feature vector of an object, and Ci = C(Xi} be the classification for Xi, i =
I, ... N, where N is the total number of objects. Ci E {I, ..., D}, where D is the number of
total classes.
According to Bayes rule,
( I ) --
pC X
p(xlc}p(c}
p(x}
It follows that the "with context" a posteriori probability of the class labels of all the objects
assuming values Cl, C2, "', CN, given all the feature vectors, is
It is reasonable to assume that the feature distribution given a class is independent of the
feature distributions of other classes, i.e.,
P(Xb X2, ... , xNlcl, C2, ... , CN} = p(xllcd???p(XNlcN}
Then Equation (1) can be rewritten as
-
p(cllxd",p(CNlxN }P(Xl}???P(XN )P(Cl ' C2, ... , CN)
P(Cl}? ..P(CN }P(Xl ' X2, ... , XN}
X Song, Y. Abu-Mostafa, 1. Sill and H. Kasdan
952
where p( cilxi) is the "no context" object-by-object Bayesian a posteriori probability, and
p( Ci) is the a priori probability of the classes, p( Xi) is the marginal probability of the
features, and P(Xl' X2, ... , XN) is the joint distribution of all the feature vectors.
Since the features
(Xl, X2, ... , XN)
are given, p(Xb X2, ... , XN) and p(xd are constant,
where
(3)
The quantity p( Cl, C2, ?.. , CN ), which we call context ratio and through which the context
plays its role, captures the dependence among the objects. In the case where all the objects
are independent, p( Cl, C2, ..? , CN) equals one - there will be no context. In the dependent
case, p( Cl, C2, ..? , CN) will not equal one, and the context has an effect on the classifications.
We deal with the application of object classification where it is the count in each class,
rather than the particular ordering or numbering of the objects, that matters. As a result,
p (Cl , C2, .?. , CN) is only a function of the count in each class. Let N d be the count in class
d, and Vd = !ft, d = 1..., D,
(4)
where Pd is the prior distribution of class d, for d = 1, ... D.
2:f=l Nd
Nand
L:f=l Vd = l.
The decision rule is to choose class labels
(Cl'
C2, ?.. , CN) =
Cl,
argmax
C2, .. . , CN such that
P(Cll C2, .?. , cNIXl, X2, ... , XN)
(5)
(Cl ,C2 , ... ,CN)
When implementing the decision rule, we need to compute and compare DN cases for
Equation 5. In the case of white blood cell recognition, D = 14 and N is typically around
600, which makes it virtually impossible to implement.
In many cases, additional constraints can be used to reduce computation, as is the case in
white blood cell identification, which will be demonstrated in the following section.
3 WmTE BLOOD CELL RECOGNITION
Leukocyte analysis is one of the major routine laboratory examinations. The utility of
leukocyte classification in clinical diagnosis relates to the fact that in various physiological
and pathological conditions the relative percentage composition of the blood leukocytes
Incorporating Contextual Infonnation in White Blood Cell Identification
953
changes. An estimate of the percentage of each class present in a blood sample conveys
information which is pertinent to the hematological diagnosis. Typical commercial differential WBC counting systems are designed to identify five major mature cell types. But
blood samples may also contain other types of cells, i.e. immature cells. These cells occur
infrequently in normal specimen, and most commercial systems will simply indicate the
presence of these cells because they can't be individually identified by the systems. But it
is precisely these cell types that relate to the production rate and maturation of new cells
and thus are important indicators of hematological disorders. Our system is designed to
differentiate fourteen WBC types which includes the five major mature types: segmented
neutrophils, lymphocytes, monocytes, eosinophils, and basophils; and the immature types:
bands (unsegmented neutrophils), metamyelocytes, myelocytes, promyelocytes, blasts, and
variant lymphocytes; as well as nucleated red blood cells and artifacts. Differential counts
are made based on the cell classifications, which further leads to diagnosis or prognosis.
The data was provided by IRIS, Inc. Blood specimens are collected at Harbor UCLA
Medical Center from local patients, then dyed with Basic Orange 21 metachromatic dye
supravital stain. The specimen is then passed through a flow microscopic imaging and image processing instrument, where the blood cell images are captured. Each image contains
a single cell with full color. There are typically 600 images from each specimen. The task
of the cell recognition system is to categorize the cells based on the images.
3.1
PREPROCESSING AND FEATURE EXTRACTION
The size of cell images are automatically tailored according to the size of the cell in the
images. Images containing larger cells have bigger sizes than those with small cells. The
range varies from 20x20 to 40x40 pixels. The average size is around 25x25. See Figure
3.1. At the preprocessing stage, the images are segmented to set the cell interior apart from
the background. Features based on the interior of the cells are extracted from the images.
The features include size, shape, color 1 and texture. See Table 1 for the list of features. 2
Figure 1: Example of some of the cell images.
3.2 CELL?BY?CELL CLASSIFICATION
The features are fed into a nonlinear feed-forward neural network with 20 inputs, 15 hidden
units with sigmoid transfer functions, and 14 sigmoid output units. A cross-entropy error
1 A color image is decomposed into three intensity images - red, green and blue respectively
2The red-blue distribution is the pixel-by-pixel log(red)- log(blue) distribution for pixels in cell
interior. The red distribution is the distribution of the red intensity in cell interior.
X Song, Y. Abu-Mostafa, 1. Sill and H. Kasdan
954
feature number
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
feature description
cell area
number of pixels on cell edge
the 4th quantile of red-blue distribution
the 4th quantile of green-red distribution
the median of red-blue distribution
the median of green-red distribution
the median of blue-green distribution
the standard deviation of red-blue distribution
the standard deviation of green-red distribution
the standard deviation of blue-green distribution
the 4th quantile of red distribution
the 4th quantile of green distribution
the 4th quantile of blue distribution
the median of red distribution
the median of green distribution
the median of blue distribution
the standard deviation of red distribution
the standard deviation of green distribution
the standard deviation of blue distribution
the standard deviation of the distance from the edge to the mass center
function is used in order to give the output a probability interpretation. Denote the input
feature vector as x, the network outputs a D dimensional vector ( D = 14 in our case)
p = {p(dlx)}, d = 1, ... , D, where p(dlx) is
p(dlx)
= Prob( a cell belongs to class dl feature x)
The decision made at this stage is
d(x) = argmax p(dlx)
d
3.3
COMBINING CONTEXTUAL INFORMATION
The "no-context" cell-by-cell decision is only based on the features presented by a cell,
without looking at any other cells. When human experts make decisions, they always look
at the whole specimen, taking into consideration the identities of other cells and adjusting
the cell-by-cell decision on a single cell according to the company it keeps. On top of the
visual perception of the cell patterns, such as shape, color, size, texture, etc., comparisons
and associations, either mental or visual, with other cells in the same specimen are made to
infer the final decision. A cell is assigned a certain identity if the company it keeps supports
that identity. For instance, the difference between lymphocyte and blast can be very subtle
sometimes, especially when the cell is large. A large unusual mononuclear cell with the
characteristics of both blast and lymphocyte is more likely to be a blast if surrounded by or
accompanied by other abnormal cells or abnormal distribution of the cells.
955
Incorporating Contextual Infonnation in White Blood Cell Identification
This scenario fits in the framework we described in section 2. The Combining Contextual
Information algorithm was used as the post-precessing of the cell-by-cell decisions.
3.4 OBSERVATIONS AND SIMPLIFICATIONS
Direct implementation of the proposed algorithm is difficult due to the computational
complexity. In the application of WBC identification, simplification is possible. We observed the following: First, we are primarily concerned with one class blast, the presence
of which has clinical significance. Secondly, we only confuse blast with another class
lymphocyte. In other words, for a potential blast, p(blastlx) ? 0, p(lymphocytelx) ?
0, p(any other classlx) ~ O. Finally, we are fairly certain about the classification of all
other classes, i.e. p(a certain c1asslx) ~ 1, p(any other c1asslx) ~ O. Based on the above
observations, we can simplify the algorithm, instead of doing an exhaustive search.
Let pf = P(Ci = dlxi), i = 1, ... , N. More specifically, let pf = p(blastlxd, pf =
p(lymphocytelxi) and pi = p(c1ass * IXi) where * is neither a blast nor a lymphocyte.
Suppose there are K potential blasts. Order the
over i, such that
B
B
pf, pf, ... , pf 's in a descending manner
B
Pl ~ P2 ~ ... ~ PK
then the probability that there are k blasts is
PB(k)
= PP ???pfpr+l???pj( PK+1"' piv p(VB = t, VL = v~ + K;/, V3, ""
VD)
where v~ is the proportion of unambiguous lymphocytes and V3, "" VD are the proportions
of the other cell types,
We can compute the PB(k)'s recursively,
for k:::: 1, "" K-l, and
This way we only need to compute K terms to get PB(k)'s . Pick the optimal number of
blasts k* that maximizes PB (k), k = 1, "., K,
An important step is to calculate p(Vl, "" VD) which can be estimated from the database,
3.5 THE ALGORITHM
Step 1 Estimate P(Vl' ,." VD) from the database, for d = 1"", D,
Step 2 Compute the object-by-object "no context" a posteriori probability p(cilxi), i
1, "', N, and Ci E {I, ... , D},
Step 3 Compute PB (k) and find k* for k = 1, .'" K, and relabel the cells accordingly,
X Song, Y. Abu-Mostafa, 1. Sill and H. Kasdan
956
4 EMPIRICAL TESTING
The algorithm has been intensively tested at IRIS, Inc. on the specimens obtained at Harbor
UCLA medical center. We compared the performances with or without using contextual
information on blood samples from 220 specimens (consisting of 13,200 cells). In about
50% of the cases, a false alarm would have occurred had context not been used. Most cells
are correctly classified, but a few are incorrectly labelled as immature cells, which raises a
flag for the doctors. Change of the classification of the specimen to abnormal requires expert intervention before the false alarm is eliminated, and it may cause unnecessary worry.
When context is applied, the false alarms for most of the specimens were eliminated, and
no false negative was introduced.
methods
no context
with context
cell
classification
88%
89%
normality
identification
50%
90%
I"V
I"V
false
positive
1"V50%
,....10%
false
negative
0%
0%
Table 2: Comparison of with and without using contextual information
5 CONCLUSIONS
In this paper we presented a novel framework for incorporating contextual information into
object identification, developed an algorithm to implement it efficiently, and applied it to
white blood cell recognition. Empirical tests showed that the "with context" approach is
significantly superior than the "no context" approach. The technique described could be
generalized to a number of domains where contextual information plays an essential role,
such a speech recognition, character recognition and other medical diagnosis regimes.
Acknowledgments
The authors would like to thank the members of Learning Systems Group at Caltech for
helpful suggestions and advice: Dr. Amir Atiya, Zehra Cataltepe, Malik Magdon-Ismail,
and Alexander Nicholson.
References
Richard, M.D., & Lippmann, R.P., (1991) Neural network classifiers estimate Bayesian a
posteriori probabilities. Neural Computation 3. pp.461-483. Cambridge, MA: MIT Press.
Kasdan, H.K., Pelmulder, J.P., Spolter, L., Levitt, G.B., Lincir, M.R., Coward, G.N., Haiby,
S. 1., Lives, J., Sun, N.C.J., & Deindoerfer, F.H., (1994) The WhiteIRISTM Leukocyte
differential analyzer for rapid high-precision differentials based on images of cytoprobereacted cells. Clinical Chemistry. Vol. 40, No.9, pp.1850-1861.
Haralick, R.M., & Shapiro, L.G.,(l992),Computer and Robot Vision, Vol.1 , AddisonWelsley.
Aus, H. A., Harms, H., ter Meulen, v., & Gunzer, U. (1987) Statistical evaluation of computer extracted blood cell features for screening population to detect leukemias. In Pierre
A. Devijver and Josef Kittler (eds.) Pattern Recognition Theory and Applications, pp. 509518. Springer-Verlag.
Kittler, J., (1987) Relaxation labelling. In Pierre A. Devijver and Josef Kittler (eds.) Pattern
Recognition Theory and Applications, pp. 99-108. Springer-Verlag.
| 1435 |@word proportion:2 nd:1 nicholson:1 pick:1 recursively:1 c1ass:1 contains:1 contextual:14 shape:2 pertinent:1 designed:2 amir:1 accordingly:1 mental:1 provides:1 five:2 dn:1 c2:11 direct:1 differential:4 manner:1 blast:11 rapid:1 nor:1 decomposed:1 automatically:1 company:2 pf:6 provided:1 maximizes:1 mass:1 developed:2 xd:1 demonstrates:1 classifier:1 unit:2 medical:3 intervention:1 appear:1 before:1 positive:1 engineering:2 local:1 au:1 sill:4 range:1 acknowledgment:1 testing:2 implement:2 area:2 empirical:3 significantly:2 hematological:2 word:1 get:1 interior:4 context:18 impossible:1 descending:1 demonstrated:1 center:3 independently:1 disorder:1 rule:3 utilizing:1 population:1 play:2 commercial:2 suppose:1 us:1 ixi:1 stain:1 infrequently:1 expensive:1 recognition:8 database:2 observed:1 role:2 ft:1 electrical:2 capture:1 calculate:1 kittler:3 sun:1 remote:1 ordering:1 zehra:1 pd:1 complexity:1 raise:1 v50:1 joint:3 various:1 exhaustive:1 larger:1 final:1 differentiate:1 propose:1 combining:2 ismail:1 description:1 object:19 p2:1 indicate:1 human:1 implementing:1 secondly:1 neutrophil:2 pl:1 accompanying:2 hold:1 around:2 normal:1 mostafa:4 major:3 precessing:1 label:2 infonnation:2 individually:1 mit:1 always:1 rather:1 haralick:1 greatly:1 xubo:1 ave:1 detect:1 helpful:2 posteriori:4 dependent:2 vl:3 typically:2 nand:1 pasadena:3 hidden:1 josef:2 pixel:5 classification:17 among:1 priori:1 orange:1 fairly:1 marginal:1 equal:2 extraction:1 eliminated:2 look:1 leukemia:1 simplify:1 richard:1 few:2 primarily:1 pathological:1 argmax:2 consisting:1 fire:1 screening:1 evaluation:1 adjust:1 pc:1 xb:2 edge:2 instance:1 cost:1 deviation:7 lymphocyte:7 hundred:1 varies:1 international:1 cll:1 ambiguity:1 satisfied:1 containing:1 choose:1 dr:1 expert:2 busy:1 potential:2 accompanied:1 chemistry:1 includes:1 matter:1 inc:2 xlc:1 doing:1 red:15 doctor:1 bayes:1 characteristic:1 efficiently:1 identify:1 identification:9 bayesian:2 bunch:1 classified:1 ed:2 against:1 pp:5 conveys:1 adjusting:1 intensively:1 color:4 subtle:1 routine:1 worry:1 feed:1 maturation:1 just:1 stage:2 dyed:1 unsegmented:1 nonlinear:1 artifact:1 effect:1 contain:1 assigned:1 laboratory:1 white:10 deal:1 ambiguous:1 unambiguous:1 iris:2 generalized:1 image:15 consideration:2 novel:1 superior:2 common:1 sigmoid:2 fourteen:1 association:1 interpretation:1 occurred:1 measurement:1 composition:1 cambridge:1 analyzer:1 had:1 robot:1 etc:1 showed:1 dye:1 belongs:1 apart:1 scenario:1 certain:3 verlag:2 harvey:1 immature:3 life:1 caltech:4 captured:1 additional:1 specimen:10 v3:2 relates:1 full:1 reduces:2 infer:1 segmented:2 clinical:4 cross:1 post:1 bigger:1 variant:1 basic:1 relabel:1 patient:1 vision:1 sometimes:1 tailored:1 achieved:1 cell:67 background:1 median:6 extra:1 virtually:1 mature:2 member:1 flow:1 call:1 counting:1 presence:2 ter:1 concerned:1 independence:2 isolation:1 harbor:2 fit:1 identified:1 prognosis:1 reduce:2 cn:11 x40:1 utility:1 passed:1 song:4 yaser:2 speech:1 cause:1 band:1 atiya:1 shapiro:1 percentage:2 estimated:1 correctly:1 blue:11 diagnosis:4 vol:2 abu:4 group:1 cataltepe:1 pb:5 blood:19 drawn:1 neither:1 pj:1 imaging:2 relaxation:1 prob:1 reasonable:1 decision:8 vb:1 abnormal:6 followed:1 simplification:2 correspondence:1 occur:1 constraint:1 precisely:1 x2:6 ucla:2 wbc:3 department:1 numbering:1 according:3 character:1 joseph:1 equation:2 discus:1 count:4 fed:1 instrument:1 unusual:1 magdon:1 rewritten:1 pierre:2 top:1 kasdan:5 include:1 quantile:5 especially:1 malik:1 quantity:1 dependence:1 microscopic:1 distance:1 thank:1 vd:6 collected:1 assuming:1 ratio:1 x20:1 difficult:1 cilxi:2 relate:1 negative:2 implementation:1 observation:2 incorrectly:1 situation:1 looking:1 intensity:2 introduced:1 california:3 leukocyte:4 address:1 pattern:4 perception:1 regime:1 program:1 green:9 examination:1 indicator:1 normality:1 technology:3 meulen:1 l992:1 prior:1 relative:1 suggestion:1 x25:1 surrounded:1 pi:1 production:1 institute:3 taking:1 xn:6 world:1 valid:1 author:2 made:6 forward:1 preprocessing:2 lippmann:1 keep:2 harm:1 unnecessary:1 xi:4 search:1 monocyte:1 table:2 transfer:1 ca:4 cl:10 domain:1 significance:1 pk:2 whole:1 noise:1 alarm:4 levitt:1 advice:1 referred:1 precision:1 dlx:4 xl:4 list:1 physiological:1 dl:1 incorporating:7 essential:1 joe:1 false:7 importance:1 ci:5 texture:2 labelling:1 confuse:1 xubosong:1 entropy:1 simply:1 likely:3 visual:2 springer:2 devijver:2 extracted:2 ma:1 identity:7 labelled:1 change:2 specifically:2 typical:1 flag:1 total:2 support:1 categorize:1 alexander:1 incorporate:1 dept:2 tested:1 |
478 | 1,436 | Ensemble and Modular Approaches for
Face Detection: a Comparison
Raphael Feraud ?and Olivier Bernier t
France-Telecom CNET DTLjDLI
Technopole Anticipa, 2 avenue Pierre Marzin, 22307 Lannion cedex, FRANCE
Abstract
A new learning model based on autoassociative neural networks
is developped and applied to face detection. To extend the detection ability in orientation and to decrease the number of false
alarms, different combinations of networks are tested: ensemble,
conditional ensemble and conditional mixture of networks. The
use of a conditional mixture of networks allows to obtain state of
the art results on different benchmark face databases.
1
A constrained generative model
Our purpose is to classify an extracted window x from an image as a face (x E
V) or non-face (x EN). The set of all possible windows is E = V uN, with
V n N = 0. Since collecting a representative set of non-face examples is impossible,
face detection by a statistical model is a difficult task. An autoassociative network,
using five layers of neurons, is able to perform a non-linear dimensionnality reduction
[Kramer, 1991]. However, its use as an estimator, to classify an extracted window
as face or non-face, raises two problems:
1. V', the obtained sub-manifold can contain non-face examples (V C V'),
2. owing to local minima, the obtained solution can be close to the linear
solution: the principal components analysis.
Our approach is to use counter-examples in order to find a sub-manifold as close as
possible to V and to constrain the algorithm to converge to a non-linear solution
[Feraud, R. et al., 1997]. Each non-face example is constrained to be reconstructed
as its projection on V. The projection P of a point x of the input space E on V, is
defined by:
?email: feraud@lannion.cnet.fr
t email: bernier@lannion.cnet.fr
473
Ensemble and Modular Approaches for Face Detection: A Comparison
? if x E V, then P{ x) = x,
? if x rJ. V: P{x) = argminYEv{d(x, V)), where
During the learning process, the projection
by: P(x) ,...., ~ 2:7=1 Vi, where VI, V2, ?.. , V n ,
in the training set of faces, of v, the nearest
d is the Euclidian distance.
P of x on V is approximated
are the n nearest neighbours,
face example of x.
The goal of the learning process is to approximate the distance V of an input space
element x to the set of faces V:
it
(x - ?)2, where M is the size of input image x
and ? the image reconstructed by the neural network,
? V{x, V) = Ilx - P(x)11 ,....,
? let x E ?, then x E V if and only if V{x, V) S T, with
threshold used to adjust the sensitivity of the model.
T
E IR, where
T
is a
15 x 20 outputs
50 neurons
35 neurons
15 x 20 inputs
Figure 1: The use of two hidden layers and counter-examples in a compression
neural network allows to realize a non-linear dimensionality reduction.
In the case of non-linear dimensionnality reduction, the reconstruction error is related to the position of a point to the non-linear principal components in the input
space. Nevertheless, a point can be near to a principal component and far from the
set of faces. With the algorithm proposed, the reconstruction error is related to the
distance between a point to the set of faces. As a consequence, if we assume that
the learning process is consistent [Vapnik, 1995], our algorithm is able to evaluate
the probability that a point belongs to the set of faces. Let y be a binary random
variable: y 1 corresponds to a face example and y 0 to a non-face example, we
use:
=
P(y
=
= llx) = e-
(,,_r):l
0'2
,where
(j
depends on the threshold
T
The size of the training windows is 15x20 pixels. The faces are normalized in
position and scale. The windows are enhanced by histogram equalization to obtain
a relative independence to lighting conditions, smoothed to remove the noise and
normalized by the average face, evaluated on the training set. Three face databases
are used : after vertical mirroring, B fl is composed of 3600 different faces with
orientation between 0 degree and 20 degree, Bf2 is composed of 1600 different faces
with orientation between 20 degree and 60 degree and B f3 is the concatenation of
Bfl and Bf2, giving a total of 5200 faces. All of the training faces are extracted
R. Feraud and O. Bernier
474
from the usenix face database(**), from the test set B of CMU(**), and from 100
images containing faces and complex backgrounds .
?
Figure 2: Left to right: the counter-examples successively chosen by the algorithm
are increasingly similar to real faces (iteration 1 to 8) .
The non-face databases (Bn!! ,B n !2,Bn!3), corresponding to each face database,
are collected by an iterative algorithm similar to the one used in
[Sung, K . and Poggio, T ., 1994] or in [Rowley, H. et al., 1995] :
?
?
?
?
1)
2)
3)
4)
Bn!
= 0, T = Tmin ,
the neural network is trained with B! + B n!,
the face detection system is tested on a set of background images,
a maximum of 100 su bimages
Xi
are collected with V
? 5) Bn! = Bn! + {xo, .. . , x n } , T = T + Jl , with Jl
? 6) while T < Tmax go back to step 2.
(Xi , V) ~ T ,
> 0,
After vertical mirroring, the size of the obtained set of non-face examples is respectively 1500 for B n!! , 600 for Bn!2 and 2600 for B n !3. Since the non-face set
(N) is too large, it is not possible to prove that this algorithm converge in a finite
time . Nevertheless, in only 8 iterations, collected counter-examples are close to the
set of faces (Figure 2) . Using this algorithm, three elementary face detectors are
constructed: the front view face detector trained on Bfl and Bnfl (CGM1), the
turned face detector trained on Bf2 and Bn!2 (CGM2) and the general face detector
trained on B!3 and Bn!3 (CGM3).
To obtain a non-linear dimensionnality reduction, five layers are necessary. However,
our experiments show that four layers are sufficient. Consequently, each CGM
has four layers (Figure 1) . The first and last layers consist each of 300 neurons,
corresponding to the image size 15x20. The first hidden layer has 35 neurons and
the second hidden layer 50 neurons. In order to reduce the false alarm rate and to
extend the face detection ability in orientation , different combinations of networks
are tested. The use of ensemble of networks to reduce false alarm rate was shown
by [Rowley, H. et al., 1995]. However, considering that to detect a face in an image,
there are two subproblems to solve, detection of front view faces and turned faces,
a modular architecture can also be used.
2
Ensemble of CGMs
Generalization error of an estimator can be decomposed in two terms: the bias and
the variance [Geman , S. et al., 1992] . The bias is reduced with prior knowledge.
The use of an ensemble of estimators can reduce the variance when these estimators
are independently and identically distributed [Raviv, Y. and Intrator, N., 1996].
Each face detector i produces :
Assuming that the three face detectors (CGM1,CGM2,CGM3) are independently
and identically distributed (iid), the ouput of the ensemble is:
Ensemble and Modular Approaches for Face Detection: A Comparison
3
475
Conditional mixture of CGMs
To extend the detection ability in orientation, a conditional mixture of CGMs is
tested. The training set is separated in two subsets: front view faces and the
corresponding counter-examples (B = 1) and turned faces and the corresponding
counter-examples (B = 0). The first subnetwork (CGMl) evaluates the probability
of the tested image to be a front view face, knowing the label equals 1 (P(y =
llx, B = 1)). The second (CGM2) evaluates the probability of the tested image to
be a turned face, knowing the label equals 0 (P(y = llx, B = 0)). A gating network
is trained to evaluate P(B llx), supposing that the partition B 1, B 0 can be
generalized to every input:
=
E[Ylx]
=
=
= E[yIB = 1, x]f(x) + E[yIB = 0, x](l- f(x))
Where f(x) is the estimated value of P{B = llx) .
This system is different from a mixture of experts introduced by
[Jacobs, R. A. et aI., 1991]: each module is trained separately on a subset of the
training set and then the gating network learns to combine the outputs.
4
Conditional ensemble of CGMs
To reduce the false alarm rate and to detect front view and turned faces, an original
combination, using (CGMl,CGM2) and a gate network, is proposed . Four sets are
defined:
? F is the front view face set,
? P is the turned face set, with F n p = 0,
? V = F U P is the face set,
? N is the non-face set, with V n N = 0,
Our goal is to evaluate P{x E Vlx). Each estimator computes respectively:
? P(x E Fix E FuN, x) (CGMl(x)),
? P(x E Pix E puN, x) (CGM2(x)),
Using the Bayes theorem, we obtain:
P(x E Fix)
P(x E Fix E FUN,x) = P(x E FUNlx)P(x E FuNlx E F,x)
Since x E F => x E FuN, then:
P(x E Fix)
P(x E Flx,x E FuN) = P(x E FUNlx)
R. Feraud and o. Bernier
476
~
~
P(x E Fix)
= P(x E Fix E FUN,x)P(x E FUNlx)
P(x E Fix) = P(x E Fix E FUN,x)[P(x E Fix)
+ P(x E Nix)]
In the same way, we have:
P(x E Pix)
= P(x E Pix E puN, x)[P(x E Pix) + P(x E Nix)]
Then:
P(x E Vlx) = P(x E Nlx)[P(x E pix E puN, x) + P(x E Fix E FuN, x)]
+P(x E Plx)P(x E Pix E puN, x) + P(x E Flx)P(x E Fix E FuN, x)
Rewriting the previous equation using the following notation, CGMl(x) for
P(x E Fix E FuN, x) and CGMl(x) for P(x E Pix E puN, x), we have:
P(x E Vlx) = P(x E Nlx)[CGMl(x) + CGM2(x)] (1)
+P(x E Plx)CGM2(x) + P(x E Flx)CGMl(x) (2)
Then, we can deduce the behaviour of the conditional ensemble:
? in N, if the output of the gate network is 0.5, as in the case of ensembles,
the conditional ensemble reduces the variance of the error (first term of the
right side of the equation (1)),
? in V, as in the case of the conditional mixture, the conditional ensemble
permits to combine two different tasks (second term of the right side of the
equation (2)) : detection of turned faces and detection of front view faces.
The gate network f(x) is trained to calculate the probability that the tested image
is a face (P(x E Vlx)), using the following cost function:
C = ~ ([f(xi)MGCl(x)
x.EV
5
+ (1- f(xd)]MGC2(x) - Yi)2 +
~ (J(xd - 0.5)2
xiEN
Discussion
Each 15x20 subimage is extracted and normalized by enhancing, smoothing and
substracting the average face, before being processed by the network. The detection
threshold T is fixed for all the tested images. To detect a face at different scales,
the image is subsampled.
The first test allows to evaluate the limits in orientation of the face detectors. The
sussex face database(**), containing different faces with ten orientations betwen 0
degree and 90 degrees, is used (Table 1). The general face detector (CGM3) uses
the same learning face database than the different mixtures of CGMs. Nevertheless,
CGM3 has a smaller orientation range than the conditional mixtures of CGMs,
and the conditional ensemble of CGMs. Since the performances of the ensemble
of CGMS are low, the corresponding hypothesis (the CGMs are iid) is invalid.
Moreover, this test shows that the combination by a gating neural network ofCGMs,
Ensemble and Modular Approaches for Face Detection: A Comparison
477
Table l :Results on Sussex face database
orientation
(degree)
CGM1
CGM2
0
10
20
30
40
50
60
70
100.0 0
62 .5 %
50.0 %
12.5 %
0.0 %
0.0 %
0.0 %
0.0 %
CGM3
Ensemble
(1,2 ,3)
Conditional
ensemble
Conditional
trained on different training set , allows to extend the detection ability to both front
view and turned faces. The conditional mixture of CGMs obtains results in term
of orientation and false alarm rate close to the best CGMs used to contruct it (see
Table 1 and Table 2).
The second test allows to evaluate the false alarms rate and to compare our results
with the best results published so far on the test set A [Rowley, H. et al., 1995] of
the CMU (**), containing 42 images of various quality. First, these results show that
the model , trained without counter-examples (GM) , overestimates the distribution
offaces and its false alarm rate is too important to use it as a face detector. Second,
the estimation of the probability distribution of the face performed by one CGM
(CGM3) is more precise than the one obtained by [Rowley, H. et al., 1995] with
one SWN (see Table 2). The conditional ensemble of CGMs and the conditional
mixture of CG Ms obtained a similar detection rate than an ensemble of SWN s
[Rowley, H. et al. , 1995], but with a false alarm rate two or three times lower. Since
the results of the conditional ensemble of CG Ms and the conditional mixture of
CGMs are close on this test , the detection rate versus the number of false alarms
is plotted (Figure 3), for different thresholds . The conditional mixture of CGMs
curve is above the one for the conditional ensemble of CG Ms.
l 00~--~----~----~----~----~--~
~
..... -
.............
........__ ........ ,,-
10
3D
Cl
so
Numt.rolt. . .wnw
110
70
10
Figure 3: Detection rate versus number of false alarms on the CMU test set A. In
dashed line conditional ensemble and in solid line conditional mixture.
The conditionnal mixture of CG Ms is used in an application called LISTEN
[Collobert, M. et al. , 1996] : a camera detects, tracks a face and controls a microphone array towards a speaker, in real time. The small size of the subimages (15x20)
processed allows to detect a face far from the camera (with an aperture of 60 degrees, the maximum distance to the camera is 6 meters). To detect a face in real
time, the number of tested hypothesis is reduced by motion and color analysis .
R. Feraud and o. Bernier
478
Table 2:Results on the CMU test set A
GM: the model trained without counter-examples, CGM1 : face detector, CGM2:
turned face detector, CGM3: general face detector. SWN: shared weight network.
(*) Considering that our goal is to detect human faces, non-human faces and rough
face drawings have not been taken into account.
Model
GM
CGMI
CGM2
CGM3
[Rowley, 1995]
(one SWN)
Ensemble
(CGM1,CGM2,CGM3)
Conditional ensemble
(CGM1,CGM2,gate)
[Rowley, 1995]
(three SWNs)
Conditional mixture
(CGM1,CGM2,gate)
Detection rate
84%
77
%
127/164
85 %
139/164
85 %
139/164
84 %
142/169*
74 %
121/164
82 %
134/164
85 %
144/169*
87 %
142/164
False alarms rate
1/1,000
+/- 5 %
+/- 5 % 5.43/1,000,000
47/33,700,000
6.3/1,000,000
+/- 5 %
212/33,700,000
1.36/1,000,000
+/- 5 %
46/33,700,000
8.13/1,000,000
+/- 5 %
179/22,000,000
0.71/1,000,000
+/- 5 %
24/33,700,000
0.77/1,000,000
+/- 5 %
26/33,700,000
2.13/1,000,000
+/- 5 %
47/22,000,000
1.15/1,000,000
+/- 5 %
39/33,700,000
+/- 0.1/1,000,00(
+/- 0.38/1,000,00
+/-
0.37/1,000,00
+/-
0.41/1,000,00
+/- 0.4/1,000,00(
+/-
0.43/1,000,00
+/- 0.38/1,000,00
+/- 0.42/1,000,00
+/-
0.35/1,000,001
(**) usenix face database, sussex face database and CMU test sets can be retrieved at
www.cs.rug.nl/peterkr/FACE/face.html.
References
[Collobert, M. et al., 1996] Collobert, M., Feraud, R, Le Tourneur, G ., Bernier, 0., Viallet, J.E, Mahieux, Y., and Collobert, D. (1996). Listen: a system for locating and
tracking individual speaker. In Second International Conference On Automatic Face
and Gesture Recognition.
[Feraud, R et al., 1997] Feraud, R, Bernier, 0., and Collobert, D. (1997). A constrained
generative model applied to face detection. Neural Processing Letters.
[Geman, S. et al., 1992] Geman, S., Bienenstock, E., and Doursat, R (1992).
networks and the bias-variance dilemma. Neural Computation, 4:1-58.
Neural
[Jacobs, R A. et al., 1991] Jacobs, R. A., Jordan, M. I., Nowlan, S. J., and Hinton, G. E.
(1991). Adaptative mixtures of local experts. Neural Computation, 3:79-87.
[Kramer, 1991] Kramer, M. (1991). Nonlinear principal component analysis using autoassociative neural networks. AIChE Journal, 37:233-243.
[Raviv, Y. and Intrator, N., 1996] Raviv, Y. and Intrator, N . (1996). Bootstrapping with
noise: An effective regularization technique. Connection Science, 8:355-372.
[Rowley, H. et al., 1995] Rowley, H., Baluja, S., and Kanade, T. (1995).
detection in visual scenes. In Neural Information Processing Systems 8.
Human face
[Sung, K. and Poggio, T ., 1994] Sung, K. and Poggio, T. (1994). Example-based learning
for view-based human face detection. Technical report, M.I.T.
[Vapnik, 1995] Vapnik, V. (1995). The Nature of Statistical Learning Theory. SpringerVerlag New York Heidelberg Berlin.
| 1436 |@word compression:1 bn:8 jacob:3 euclidian:1 solid:1 reduction:4 nowlan:1 realize:1 partition:1 remove:1 generative:2 pun:5 five:2 mahieux:1 constructed:1 ouput:1 prove:1 combine:2 detects:1 decomposed:1 window:5 considering:2 notation:1 moreover:1 bootstrapping:1 sung:3 every:1 collecting:1 fun:9 feraud:9 xd:2 control:1 overestimate:1 before:1 local:2 limit:1 consequence:1 tmax:1 range:1 camera:3 developped:1 projection:3 close:5 vlx:4 impossible:1 equalization:1 www:1 go:1 independently:2 estimator:5 array:1 enhanced:1 gm:3 olivier:1 us:1 hypothesis:2 element:1 approximated:1 recognition:1 geman:3 database:10 module:1 calculate:1 decrease:1 counter:8 rowley:9 trained:10 raise:1 dilemma:1 various:1 separated:1 effective:1 modular:5 solve:1 drawing:1 ability:4 reconstruction:2 raphael:1 fr:2 turned:9 bf2:3 produce:1 raviv:3 nearest:2 nlx:2 c:1 owing:1 human:4 behaviour:1 fix:12 generalization:1 elementary:1 purpose:1 estimation:1 label:2 rough:1 contruct:1 cgm:2 cg:4 detect:6 hidden:3 bienenstock:1 france:2 pixel:1 orientation:10 html:1 art:1 constrained:3 smoothing:1 equal:2 f3:1 betwen:1 report:1 neighbour:1 composed:2 individual:1 subsampled:1 detection:22 adjust:1 mixture:16 nl:1 necessary:1 poggio:3 plotted:1 classify:2 cost:1 subset:2 too:2 front:8 international:1 sensitivity:1 successively:1 containing:3 expert:2 account:1 vi:2 depends:1 collobert:5 performed:1 view:9 bayes:1 ir:1 offaces:1 variance:4 ensemble:26 iid:2 lighting:1 published:1 detector:12 email:2 evaluates:2 adaptative:1 knowledge:1 listen:2 dimensionality:1 color:1 back:1 evaluated:1 su:1 nonlinear:1 quality:1 contain:1 normalized:3 regularization:1 during:1 sussex:3 speaker:2 m:4 generalized:1 motion:1 image:13 jl:2 extend:4 llx:5 ai:1 automatic:1 deduce:1 aiche:1 retrieved:1 belongs:1 binary:1 yi:1 minimum:1 rug:1 converge:2 dashed:1 rj:1 reduces:1 technical:1 cgms:14 gesture:1 enhancing:1 cmu:5 histogram:1 iteration:2 background:2 separately:1 swn:4 doursat:1 cedex:1 supposing:1 bfl:2 jordan:1 near:1 identically:2 independence:1 architecture:1 reduce:4 avenue:1 knowing:2 plx:2 locating:1 york:1 autoassociative:3 mirroring:2 ylx:1 ten:1 numt:1 processed:2 reduced:2 estimated:1 track:1 four:3 threshold:4 nevertheless:3 rewriting:1 pix:7 letter:1 flx:3 layer:8 fl:1 constrain:1 scene:1 combination:4 smaller:1 increasingly:1 xo:1 taken:1 equation:3 permit:1 intrator:3 v2:1 pierre:1 gate:5 original:1 giving:1 subnetwork:1 distance:4 berlin:1 concatenation:1 manifold:2 collected:3 assuming:1 difficult:1 x20:4 subproblems:1 perform:1 vertical:2 neuron:6 dimensionnality:3 nix:2 benchmark:1 finite:1 hinton:1 tmin:1 cnet:3 precise:1 smoothed:1 usenix:2 lannion:3 introduced:1 connection:1 able:2 ev:1 wnw:1 prior:1 meter:1 relative:1 versus:2 substracting:1 degree:8 sufficient:1 consistent:1 last:1 bias:3 side:2 face:88 subimage:1 distributed:2 curve:1 computes:1 far:3 reconstructed:2 approximate:1 obtains:1 aperture:1 xi:3 un:1 iterative:1 table:6 kanade:1 nature:1 heidelberg:1 complex:1 cl:1 noise:2 alarm:11 telecom:1 representative:1 en:1 sub:2 position:2 yib:2 learns:1 theorem:1 gating:3 swns:1 consist:1 false:11 vapnik:3 subimages:1 ilx:1 visual:1 tracking:1 corresponds:1 extracted:4 conditional:25 goal:3 kramer:3 consequently:1 invalid:1 towards:1 shared:1 springerverlag:1 baluja:1 principal:4 microphone:1 total:1 called:1 evaluate:5 tested:9 |
479 | 1,437 | Wavelet Models for Video Time-Series
Sheng Ma and Chuanyi Ji
Department of Electrical, Computer, and Systems Engineering
Rensselaer Polytechnic Institute, Troy, NY 12180
e-mail: shengm@ecse.rpi.edu, chuanyi@ecse.rpi.edu
Abstract
In this work, we tackle the problem of time-series modeling of video
traffic. Different from the existing methods which model the timeseries in the time domain, we model the wavelet coefficients in the
wavelet domain. The strength of the wavelet model includes (1) a
unified approach to model both the long-range and the short-range
dependence in the video traffic simultaneously, (2) a computationally efficient method on developing the model and generating high
quality video traffic, and (3) feasibility of performance analysis using the model.
1
Introduction
As multi-media (compressed Variable Bit Rate (VBR) video, data and voice) traffic
is expected to be the main loading component in future communication networks,
accurate modeling of the multi-media traffic is crucial to many important applications such as video-conferencing and video-on-demand. From modeling standpoint, multi-media traffic can be regarded as a time-series, which can in principle
be modeled by techniques in time-seres modeling. Modeling such a time-series, however, turns out to be difficult, since it has been found recently that real-time video
and Ethernet traffic possesses the complicated temporal behavior which fails to be
modeled by conventional methods[3] [4]. One of the significant statistical properties
found recently on VBR video traffic is the co-existence of the long-range (LRD) and
the short-range (SRD) dependence (see for example [4][6] and references therein).
Intuitively, this property results from scene changes, and suggests a complex behavior of video traffic in the time domain[7]. This complex temporal behavior makes
accurate modeling of video traffic a challenging task. The goal of this work is to develop a unified and computationally efficient method to model both the long-range
and the short-range dependence in real video sources.
Ideally, a good traffic model needs to be (a) accurate enough to characterize pertinent statistical properties in the traffic, (b) computationally efficient, and (c) fea-
s. Ma and C. Jj
916
sible for the analysis needed for network design. The existing models developed
to capture both the long-range and the short-range dependence include Fractional
Auto-regressive Integrated Moving Average (FARIMA) models[4]' a model based
on Hosking's procedure[6], Transform-Expand-Sample (TES) model[9] and scenebased models[7]. All these methods model both LRD and SRD in the time domain .
The scene-based modeling[7] provides a physically interpretable model feasible for
analysis but difficult to be made very accurate. TES method is reasonably fast but
too complex for the analysis. The rest of the methods suffer from computational
complexity too high to be used to generate a large volume of synthesized video
traffic.
To circumvent these problems, we will model the video traffic in the wavelet domain
rather than in the time domain. Motivated by the previous work on wavelet representations of (the LRD alone) Fractional Gaussian Noise (FGN) process (see [2]
and references therein), we will show in this paper simple wavelet models can simultaneously capture the short-range and the long-rage dependence through modeling
two video traces. Intuitively, this is due to the fact that the (deterministic) similar
structure of wavelets provides a natural match to the (statistical) self-similarity of
the long-range dependence. Then wavelet coefficients at each time scale is modeled
based on simple statistics. Since wavelet transforms and inverse transforms is in
the order of O(N) , our approach will be able to attain the lowest computational
complexity to generate wavelet models. Furthermore, through our theoretical analysis on the buffer loss rate, we will also demonstrate the feasibility of using wavelet
models for theoretical analysis.
1.1
Wavelet Transforms
In L2(R) space, discrete wavelets ?j(t)'s are ortho-normal basis which can be represented as ?j(t) = 2- j / 2 ?(2- i t - m), for t E [0,2 K - 1] with K ~ 1 being an
integer. ?(t) is the so-called mother wavelet. 1 ~ j ~ K and 0 ~ m ~ 2K - j - 1
represent the time-scale and the time-shift, respectively. Since wavelets are the dilation and shift of a mother wavelet, they possess a deterministic similar structure
at different time scales. For simplicity, the mother wavelet in this work is chosen
to be the Haar wavelet, where ?(t) is 1 for 0 ~ t < 1/2, -1 for 1/2 ~ t < 1 and 0
otherwise.
Let dj's be wavelet coefficients of a discrete-time process x(t) (t E [0,2 K 1]) .
Then dj can be obtained through the wavelet transform dj =
K
L:;=O-l x(t)?j(t). x(t) can be represented through the inverse wavelet transform
K
2
1
X( t) =L:j=l L:m=O - dj?j(t) + ?o, where ?o is equal to the average of x(t).
K - ,
2
2.1
Wavelet Modeling of Video Traffic
The Video Sources
Two video sources are used to test our wavelet models: (1) "Star Wars" [4]' where
each frame is encoded by JPEG-like encoder, and (2) MPEG coded videos at Group
of Pictures (GOP) level[7][ll] called "MPEG GOP" in the rest of the paper. The
modeling is done at either the frame level or the GOP level.
Wavelet Models for Video Time-Series
917
31
&
..
, ,..
32
9
;30
21
>
?
i 21
u
22
20
..
? AMIA('.0.4.o)
?
- .'
34
i
20
? .+
31
?
?
. :. +
.1
dR{' )
~.O
~
? ?
':QOP
. :Si90Soutlt
.
0
"
..
...
.0
.2
~
.
.0
ijs
?
?
..
~ AR1IIA{O,Q.4.o)
? ?
. .? ..
, :.:
. .-
.
--
+.
..a
.
~
?
"0
12
10
-I
14
..
.,\
8
0
TWoScoIoI
Figure 1: Log 2 of Variance of dJ versus
Figure 2: Log 2 of Variance of dJ versus
the time scale j
the time scale j
0 .?
- : StarW.,.
.. :GOP
0.8
j
J
0 ??
0.2
0
-0.20
2
8
8
10
Lag
12
1.
18
18
Figure 3: The sample auto correlations of d
2.2
20
s.
The Variances and Auto-correlation of Wavelet Coefficients
As the first step to understand how wavelets capture the LRD and SRD, we plot in
Figure (1) the variance of the wavelet coefficients dj's at different time scales for
both sources. To understand what the curves mean, we also plot in Figure (2) the
variances of wavelet coefficients for three well-known processes: FARIMA(O, 0.4, 0),
FARIMA(l, 0.4, 0), and AR(l). FARIMA(O, 0.4,0) is a long-range dependent ptocess with Hurst parameter H = 0.9. AR(l) is a short-range dependent process, and
FARIMA(l, 0.4,0) is a mixture of the long-range and the short-range dependent
process.
As observed, for FARIMA(O, 0.4, 0) process (LRD alone), the variance increases
with j exponentially for all j. For AR(l) (SRD alone), the variance increases at
an even faster rate than that of FARIMA(O, 0.4, 0) when j is small but saturates
when j is large. For FARIMA(l, 0.4,0), the variance shows the mixed properties
from both AR(l) and FARIMA(O, 0.4, 0). The variance of the video sources behaves
similarly to that of FARIMA(l, 0.4,0), and thus demonstrate the co-existence of the
SRD and LRD in the video sources in the wavelet domain.
s
Figure 3 gives the sample auto-correlation of d in terms of m's. The autocorrelation function of the wavelet coefficients approaches zero very rapidly, and
s. Ma and C. Ji
918
.
,
~
~
.
;
I
m
-...
-2
0
so
...
Quantll._ of Stand.rd Norm ??
Figure 4: Quantile-Quantile of d';' for j
= 3.
Left: Star Wars. Right: GOP.
thus indicates the short-range dependence in the wavelet domain. This suggests
that although the autocorrelation of the video traffic is complex in the time-domain,
modeling wavelet coefficients may be done using simple statistics within each time
scale. Similar auto-correlations have been observed for the other j's.
2.3
Marginal Probability Density Functions
Is variance sufficient for modeling wavelet coefficients? Figure (4) plots the Q - Q
plots for the wavelet coefficients of the two sources at j
31 . The figure shows that
the sample marginal density functions of wavelet coefficients for both the "Star
Wars" and the MPEG GOP source at the given time scale have a much heavier tail
than that of the normal distribution. Therefore, the variance alone is only sufficient
when the marginal density function is normal, and in general a marginal density
function should be considered as another pertinent statistical property.
=
It should be noted that correlation among wavelet coefficients at different time
scales is neglected in this work for simplicity. We will show both empirically and
theoretically that good performance in terms of sample auto-correlation and sample
buffer loss probability can be obtained by a corresponding simple algorithm. More
careful treatment can be found in [8].
2.4
An Algorithm for Generating Wavelet Models
The algorithm we derive include three main steps: (a) obtain sample variances
of wavelet coefficients at each time scale, (b) generate wavelet coefficients independently from the normal marginal density function using the sample mean and
variance 2, and (c) perform a transformation on the wavelet coefficients so that the
ISimilar behaviors have been observed at the other time scales. A Q - Q plot is a
standard statistical tool to measure the deviation of a marginal density function from a
normal density. The Q - Q plots of a process with a normal marginal is a straight line.
The deviation from the line indicates the deviation from the normal density. See [4] and
references therein for more details.
2The mean of the wavelet coefficients can be shown to be zero for stationary processes.
Wavelet Models for Video Time-Series
919
resulting wavelet coefficients have a marginal density function required by the traffic. The obtained wavelet coefficients form a wavelet model from which synthesized
video traffic can be generated. The algorithm can be summarized as follows.
Let x(t) be the video trace oflength N.
Algorithm
1. Obtain wavelet coefficients from x(t) through the wavelet transform.
2. Compute the sample variance Uj of wavelet coefficients at each time scale
j.
3. Generate new wavelet coefficients dj's for all j and m independently
through Gaussian distributions with variances Uj 's obtained at the previous
step.
4. Perform a transformation on the wavelet coefficients so that the marginal
density function of wavelet coefficients is consistent with that determined
by the video traffic (see [6] for details on the transformation).
5. Do inverse wavelet transform using the wavelet coefficients obtained at the
previous step to get the synthesized video traffic in the time domain .
The computational complexity of both the wavelet transform (Step 1) and the
inverse transform (Step 5) is O(N). So is for Steps 2, 3 and 4. Then O(N) is
the computational cost of the algorithm, which is the lowest attainable for traffic
models.
2.5
Experimental Results
Video traces of length 171, 000 for "Star Wars" and 66369 for "MPEG GOP" are
used to obtain wavelet models. FARIMA models with 45 parameters are also obtained using the same data for comparison. The synthesized video traffic from
both models are generated and used to obtain sample auto-correlation functions in
the time-domain, and to estimate the buffer loss rate. The results3 are given in
Figure (6). Wavelet models have shown to outperform the FARIMA model.
For the computation time, it takes more than 5-hour CPU time 4 on a SunSPARC
5 workstation to develop the FARIMA model and to generate synthesized video
traffic of length 171, 000 5 . It only takes 3 minutes on the same machine for our
algorithm to complete the same tasks.
3
Theory
It has been demonstrated empirically in the previous section that the wavelet model,
which ignores the correlation among wavelet coefficients of a video trace, can match
well the sample auto-correlation function and the buffer loss probability. To further
evaluate the feasibility of the wavelet model, the buffer overflow probability has
been analyzed theoretically in [8]. Our result can be summarized in the following
theorem.
3Due to page limit, we only provide plots for JPEG. GOP has similar results and was
reported in [8].
4Computation time includes both parameter estimation and synthesized traffic
generation.
5The computational complexity to generate synthesized video traffic of length N is
O(N2) for an FARIMA model[5][4].
S. Ma and C. Ii
920
..-
-2
-2.5
OJ
01
-4.5
I
I
I
-5
~.5
0.2
4~1
0.1
Figure 5: "-": Autocorrelation of "Star
Wars"; "- -": ARIMA(25,d,20); " ".
Our Algorithm
O.li
0.4
0.45
o.s
0.&6
0.8
085
07
0.71
OJ
Figure 6: Loss rate attained via
simulation. Vertical axis: loglO (Loss
Rate); horizontal axis: work load.
"-": the single video source; "".
FARIMA(25,d,20); "-" Our algorithm.
The normalized buffer size: 0.1, 1, 10,30
and 100 from the top down.
Theorem Let BN and EN be the buffer sizes at the Nth time slot due to the syn-
thesized traffic by the our wavelet model, and by the FGN process, respectively. Let
C and B represent the capacity, and the maximum allowable buffer size respectively.
Then
InPr(BN > B)
InPr(EN > B)
(C - JL)2(i!:;?(1-H)e~7f-)2H
20- 2 (1- H)2
(1)
where ~ < H < 1 is the Hurst parameter. JL and 0- 2 is the mean and the variance
of the traffic, respectively. B is assume to be (C - It )2ko, where ko is a positive
integer.
This result demonstrates that using our simple wavelet model which neglects the
correlations among wavelet coefficients, buffer overflow probability obtained is similar to that of the original FGN process as given in[10]. In other words, it shows
that the wavelet model for a FG N process can have good modeling performance in
terms of the buffer overflow criterion.
We would like to point out that the above theorem is held for a FGN process.
Further work are needed to account for more general processes.
4
Concl usions
In this work, we have described an important application on time-series modeling: modeling video traffic. We have developed a wavelet model for the timeseries. Through analyzing statistical properties of the time-series and comparing
the wavelet model with FARIMA models, we show that one of the key factors to successfully model a time-series is to choose an appropriate model which naturally fits
the pertinant statistical properties of the time-series. We have shown wavelets are
particularly feasible for modeling the self-similar time-series due to the video traffic.
Wavelet Models for Video Time-Series
921
We have developed a simple algorithm for the wavelet models, and shown that the
models are accurate, computationally efficient and simple enough for analysis.
References
[1] I. Daubechies, Ten Lectures on Wavelets. Philadelphia: SIAM, 1992.
[2] Patrick Flandrin, "Wavelet Analysis and Synthesis of Fractional Brownian Motion", IEEE transactions on Information Theory, vol. 38, No.2, pp.910-917,
1992.
[3] W.E Leland, M.S . Taqqu, W. Willinger and D.V. Wilson, "On the Self-Similar
Nature of Ethernet Traffic (Extended Version)", IEEE/ACM Transactions on
Networking, vo1.2, 1-14, 1994.
[4] Mark W. Garrett and Walter Willinger. "Analysis, Modeling and Generation
of Self-Similar VBR Video Traffic.", in Proceedings of ACM SIGCOMM'94,
London, U.K, Aug., 1994
[5] J .R.M. Hosking, "Modeling Persistence in Hydrological Time Series Using Fractional Differencing", Water Resources Research, 20, pp . 1898-1908, 1984.
[6] C. Huang, M. Devetsikiotis, I. Lambadaris and A.R. Kaye, "Modeling and
Simulation of Self-Similar Variable Bit Rate Compressed Video: A Unified
Approach", in Proceedings of ACM SIGCOMM'95, pp. 114-125.
[7] Predrag R. Jelenlnovic, Aurel A. Lazar, and Nemo Semret. The effect of multiple time scales and subexponentiality in mpeg video streams on queuing behavior. IEEE Journal on Selected Area of Communications, 15, to appear in
May 1997.
[8] S. Ma and C. Ji, "Modeling Video Traffic in Wavelet Domain" , to appear IEEE
INFO COM, 1998.
[9] B. Melamed, D. Raychaudhuri, B. Sengupta, and J. Zdepski. Tes-based video
source modeling for performance evaluation of integrated networks. IEEE
Transactions on Communications, 10, 1994.
[10] Ilkka Norros, "A storage model with self-similar input," Queuing Systems,
vol.16, 387-396, 1994.
[11] O. Rose. "Statistical properties of mpeg video traffic and their impact on
traffic modeling in atm traffic engineering", Technical Report 101, University
of Wurzburg, 1995.
,
)
| 1437 |@word version:1 loading:1 norm:1 simulation:2 bn:2 attainable:1 series:13 existing:2 comparing:1 com:1 rpi:2 srd:5 willinger:2 pertinent:2 plot:7 interpretable:1 alone:4 stationary:1 selected:1 vbr:3 short:8 regressive:1 provides:2 leland:1 autocorrelation:3 theoretically:2 expected:1 behavior:5 multi:3 cpu:1 medium:3 lowest:2 what:1 developed:3 unified:3 transformation:3 temporal:2 tackle:1 demonstrates:1 appear:2 positive:1 engineering:2 limit:1 analyzing:1 amia:1 therein:3 suggests:2 challenging:1 co:2 range:15 procedure:1 area:1 attain:1 persistence:1 word:1 get:1 storage:1 conventional:1 deterministic:2 demonstrated:1 independently:2 simplicity:2 regarded:1 ortho:1 melamed:1 particularly:1 observed:3 electrical:1 capture:3 rose:1 complexity:4 ideally:1 neglected:1 taqqu:1 basis:1 represented:2 walter:1 fast:1 london:1 lazar:1 encoded:1 lag:1 otherwise:1 compressed:2 encoder:1 statistic:2 transform:7 predrag:1 fea:1 rapidly:1 generating:2 mpeg:6 derive:1 develop:2 aug:1 ethernet:2 considered:1 normal:7 estimation:1 successfully:1 tool:1 ecse:2 gaussian:2 loglo:1 hydrological:1 rather:1 wilson:1 indicates:2 dependent:3 integrated:2 wurzburg:1 expand:1 among:3 sengupta:1 marginal:9 equal:1 future:1 lrd:6 report:1 simultaneously:2 evaluation:1 mixture:1 analyzed:1 held:1 accurate:5 theoretical:2 modeling:22 jpeg:2 ar:4 cost:1 deviation:3 too:2 characterize:1 reported:1 hosking:2 density:10 siam:1 synthesis:1 daubechies:1 choose:1 huang:1 dr:1 li:1 sible:1 account:1 star:5 summarized:2 includes:2 coefficient:26 stream:1 queuing:2 traffic:35 complicated:1 atm:1 variance:16 kaye:1 flandrin:1 ijs:1 straight:1 networking:1 pp:3 naturally:1 workstation:1 treatment:1 fractional:4 syn:1 garrett:1 attained:1 done:2 furthermore:1 aurel:1 correlation:10 sheng:1 horizontal:1 quality:1 effect:1 normalized:1 ll:1 self:6 noted:1 criterion:1 allowable:1 complete:1 demonstrate:2 motion:1 recently:2 behaves:1 ji:3 empirically:2 exponentially:1 volume:1 jl:2 tail:1 synthesized:7 oflength:1 significant:1 mother:3 rd:1 similarly:1 dj:8 concl:1 moving:1 similarity:1 patrick:1 brownian:1 buffer:10 gop:8 ii:1 multiple:1 technical:1 match:2 faster:1 long:8 sigcomm:2 coded:1 feasibility:3 impact:1 ko:2 rage:1 physically:1 represent:2 source:10 crucial:1 standpoint:1 rest:2 posse:2 integer:2 hurst:2 enough:2 usions:1 fit:1 shift:2 motivated:1 war:5 heavier:1 suffer:1 jj:1 transforms:3 ten:1 generate:6 outperform:1 discrete:2 arima:1 vol:2 group:1 key:1 inverse:4 bit:2 strength:1 scene:2 conferencing:1 department:1 developing:1 intuitively:2 computationally:4 resource:1 turn:1 needed:2 chuanyi:2 polytechnic:1 appropriate:1 voice:1 existence:2 original:1 top:1 include:2 farima:16 neglect:1 quantile:2 uj:2 overflow:3 dependence:7 capacity:1 mail:1 water:1 length:3 modeled:3 differencing:1 difficult:2 troy:1 trace:4 info:1 design:1 perform:2 vertical:1 timeseries:2 saturates:1 communication:3 extended:1 frame:2 required:1 hour:1 able:1 oj:2 video:43 natural:1 circumvent:1 haar:1 nth:1 picture:1 axis:2 auto:8 philadelphia:1 l2:1 loss:6 lecture:1 mixed:1 generation:2 versus:2 sufficient:2 consistent:1 principle:1 understand:2 institute:1 fg:1 curve:1 stand:1 ignores:1 made:1 transaction:3 rensselaer:1 dilation:1 nature:1 reasonably:1 complex:4 domain:12 main:2 noise:1 n2:1 en:2 ny:1 fails:1 wavelet:72 minute:1 theorem:3 down:1 load:1 te:3 demand:1 fgn:4 acm:3 ma:5 slot:1 goal:1 careful:1 feasible:2 change:1 determined:1 vo1:1 called:2 experimental:1 mark:1 evaluate:1 |
480 | 1,438 | Learning Path Distributions using
Nonequilibrium Diffusion Networks
Paul Mineiro *
Javier Movellan
pmineiro~cogsci.ucsd.edu
movellan~cogsci.ucsd.edu
Department of Cognitive Science
University of California, San Diego
La Jolla, CA 92093-0515
Department of Cognitive Science
University of California, San Diego
La Jolla, CA 92093-0515
Ruth J. Williams
williams~math.ucsd.edu
Department of Mathematics
University of California, San Diego
La Jolla, CA 92093-0112
Abstract
We propose diffusion networks, a type of recurrent neural network
with probabilistic dynamics, as models for learning natural signals
that are continuous in time and space. We give a formula for the
gradient of the log-likelihood of a path with respect to the drift
parameters for a diffusion network. This gradient can be used to
optimize diffusion networks in the nonequilibrium regime for a wide
variety of problems paralleling techniques which have succeeded in
engineering fields such as system identification, state estimation
and signal filtering. An aspect of this work which is of particular interest to computational neuroscience and hardware design is
that with a suitable choice of activation function, e.g., quasi-linear
sigmoidal, the gradient formula is local in space and time.
1
Introduction
Many natural signals, like pixel gray-levels, line orientations, object position, velocity and shape parameters, are well described as continuous-time continuous-valued
stochastic processes; however, the neural network literature has seldom explored the
continuous stochastic case. Since the solutions to many decision theoretic problems
of interest are naturally formulated using probability distributions, it is desirable
to have a flexible framework for approximating probability distributions on continuous path spaces. Such a framework could prove as useful for problems involving
continuous-time continuous-valued processes as conventional hidden Markov models have proven for problems involving discrete-time sequences.
Diffusion networks are similar to recurrent neural networks, but have probabilistic
dynamics. Instead of a set of ordinary differential equations (ODEs), diffusion
networks are described by a set of stochastic differential equations (SDEs). SDEs
provide a rich language for expressing stochastic temporal dynamics and have proven
? To whom correspondence should be addressed.
Learning Path Distributions Using Nonequilibriwn Diffusion Networks
599
Figure 1: An example where the average of desirable paths yields an undesirable
path, namely one that collides with the tree.
useful in formulating continuous-time statistical inference problems, resulting in
such solutions as the continuous Kalman filter and generalizations of it like the
condensation algorithm (Isard & Blake, 1996).
A formula is given here for the gradient of the log-likelihood of a path with respect to the drift parameters for a diffusion network. Using this gradient we can
potentially optimize the model to approximate an entire probability distribution of
continuous paths, not just average paths or equilibrium points. Figure 1 illustrates
the importance of this kind of learning by showing a case in which learning average
paths would have undesirable results, namely collision with a tree. Experience has
shown that learning distributions of paths, not just averages, is crucial for dynamic
perceptual tasks in realistic environments, e.g., visual contour tracking (Isard &
Blake, 1996). Interestingly, with a suitable choice of activation function, e.g., quasilinear sigmoidal, the gradient formula depends only upon local computations, i.e.,
no time unfolding or explicit backpropagation of error is needed. The fact that
noise localizes the gradient is of potential interest for domains such as theoretical
neuroscience, cognitive modeling and hardware design.
2
Diffusion Networks
?
Hereafter Cn refers to the space of continuous Rn-valued functions over the time
interval [0, TJ, with T E R, T > fixed throughout this discussion.
A diffusion network with parameter A E RP is a random process defined via an
SD E of the form
dX(t) = JL(t, X(t), A)dt + adB(t),
X(O) '"
Ito
(1)
1I,
where X is a Cn-valued process that represents the temporal dynamics of the n
nodes in the network; JL: [0, T] x R n x RP -+ Rn is a deterministic function called
the drift; A E RP is the vector of drift parameters, e.g., synaptic weights, which
are to be optimized; B is a Brownian motion process which provides the random
driving term for the dynamics; 1I is the initial distribution of the solution; and
a E R, a > 0, is a fixed constant called the dispersion coefficient, which determines
the strength of the noise term. In this paper we do not address the problem of
optimizing the dispersion or the initial distribution of X. For the existence and
uniqueness in law of the solution to (1) JL(',', A) must satisfy some conditions. For
example, it is sufficient that it is Borel measurable and satisfies a linear growth
condition: IJL(t, x, A)I S; K-\(l + Ixl) for some K-\ > and all t E [0, T], x ERn; see
?
600
P. Mineiro, 1. Movellan and R. 1. Williams
(Karatzas & Shreve, 1991, page 303) for details.
It is typically the case that the n-dimensional diffusion network will be used to
model d-dimensional observations with n > d. In this case we divide X into hidden
and observable l components, denoted Hand 0 respectively, so that X = (H,O).
Note that with a = 0 in equation (1), the model becomes equivalent to a continuoustime deterministic recurrent neural network. Diffusion networks can therefore be
thought of as neural networks with "synaptic noise" represented by a Brownian motion process. In addition, diffusion networks have Markovian dynamics, and hidden
states if n > d; therefore, they are also continuous-time continuous-state hidden
Markov models. As with conventional hidden Markov models, the probability density of an observable state sequence plays an important role in the optimization of
diffusion networks. However, because X is a continuous-time process, care must
taken in defining a probability density.
2.1
Density of a continuous observable path
Let (XA, B>') defined on some filtered probability space (0, F, {Fd, p) be a (weak)
solution of (1) with fixed parameter A. Here X>' = (HA,O>') represents the states
of the network and is adapted to the filtration {Fd, B>' is an n-dimensional {Fdmartingale Brownian motion and the filtration {Ft } satisfies the usual conditions
(Karatzas and Shreve, 1991, page 300). Let Q>' be the unique probability law
generated by any weak solution of (1) with fixed parameter A
Q>'(A) = p(X>' E A) for all A E F,
(2)
where F is the Borel sigma algebra generated by the open sets of Cn. Setting
n = Cn , nh = Cn- d, and no = Cd with associated Borel a-algebras F, Fh and Fo,
respectively, we have n = nh x no, F = Fh ? Fo, and we can define the marginal
laws for the hidden and observable components of the network by
Q~(Ah) = QA(Ah x Cd) ~ P(H A E A h ) for all Ah E Fh,
(3)
Q~(Ao) = Q>'(Cn-d
Ao) ~ p(O>' E Ao) for all Ao E Fo.
(4)
For our purposes the appropriate generalization of the notion of a probability density
on Rm to the general probability spaces considered here is the Radon-Nikodym
derivative with respect to a reference measure that dominates all members of the
family {Q>'} >'ERp (Poor, 1994, p.264ff). A suitable reference measure P is the law of
the solution to (1) with zero drift (IJ = 0). The measures induced by this reference
measure over Fh and Fo are denoted by Ph and Po, respectively. Since in the
reference model there are no couplings between any of the nodes in the network,
the hidden and observable processes are independent and it follows that
P(Ah x Ao) = Ph (Ah)Po(Ao) for all Ah E Fh,Ao E .1'0'
(5)
The conditions on IJ mentioned above are sufficient to ensure a Radon-Nikodym
derivative for each QA with respect to the reference measure. Using Girsanov's
Theorem (Karatzas & Shreve, 1991, p.190ff) its form can be shown to be
Z'(w) =
X
~~ (w) = exp { : '
lT
-
I'(t,w(t), A) . dw(t)
2~'
lT
(6)
II'(t, w(t), A)I'dt} , wE \1,
lIn our treatment we make no distinction between observables which are inputs and
those which are outputs. Inputs can be conceptualized as observables under "environmental control," i.e., whose drifts are independent of both A and the hidden and output
processes.
Learning Path Distributions Using Nonequilibriwn Diffusion Networks
601
where the first integral is an Ito stochastic integral. The random variable Z>. can
be interpreted as a likelihood or probability density with respect to the reference
mode1 2 . However equation (6) defines the density of Rn-valued paths of the entire
network, whereas our real concern is the density of Rd-valued observable paths.
Denoting wE 0 as w = (Wh,w o) where Wh E Oh and Wo E 0 0 , note that
(7)
(8)
and therefore the Radon-Nikodym derivative of Q~ with respect to Po, the density
of interest, is given by
Z;(wo) = ~~: (wo) = EPh[Z>'( ., wo?)' Wo E 0 0 ,
2.2
(9)
Gradient of the density of an observable path
Z;
The gradient of
with respect to A is an important quantity for iterative optimization of cost functionals corresponding to a variety of problems of interest, e.g.,
maximum likelihood estimation of diffusion parameters for continuous path density
estimation. Formal differentiation 3 of (9) yields
\7>.logZ~(wo) = EPh[Z~lo( ? ,wo)\7),logZ'\(.,wo )),
(10)
>.
6
Z)'(w)
Zh lo (w) = Z;(wo) '
(11)
where
r
a Jo
\7),logZ)'(w) = -\-
T
J(t,W(t),A)' dl(w,t),
J. (
A) ~ 8J.lk(t, x , A)
Jk t,x,
8A '
'
(12)
(13)
J
lew, t)
~ wet) -
w(O)
-lot
J.l(s,w(s), A)ds.
(14)
Equation (10) states that the gradient of the density of an observable path can be
found by clamping the observable nodes to that path and performing an average
of Z~lo \7.x log Z)' with respect to Ph, i.e., average with respect to the hidden paths
distributed as a scaled Brownian motion. This makes intuitive sense: the output
gradient of the log density is a weighted average of the total gradient of the log
density, where each hidden path contributes according to its likelihood Z~lo given
the output.
In practice to evaluate the gradient, equation (10) must be approximated. Here we
use Monte Carlo techniques, the efficiency of which can be improved by sampling
according to a density which reduces the variance of the integrand. Such a density
2To ease interpretation of (6) consider the simpler case of a one-dimensional Gaussian
random variable with mean JL and variance (J"2. The ratio of the density of such a model
with respect to an equivalent model with zero mean is exp(-f<JLx - ~JL2). Equation (6)
can be viewed as a generalization of this same idea to Brownian motion.
3S ee (Levanony et aL, 1990) for sufficient conditions for the differentiation in equation
(10) to be valid.
602
P. Mineiro, J. Movellan and R. 1. Williams
is available for models with hidden dynamics which do not explicitly depend upon
the observables, i.e., the observable nodes do not send feedback connections to the
hidden states. Models which obey this constraint are henceforth denoted factorial.
Denoting J-Lh and J-Lo as the hidden and observable components, respectively, of the
drift vector, and Bh and Bo as the hidden and observable components, respectively,
of the Brownian motion, for a factorial network we have
dH(t) = J-Lh(t, H(t), >')dt + adBh(t),
(15)
(16)
dO(t) = J-Lo(t, H(t), O(t), >')dt + adBo(t).
The drift for the hidden variables does not depend on the observables, and Girsanov's theorem gives us an explicit formula for the density of the hidden process.
iT
Z~(Wh) = dQ>'
dP: (Wh) = exp { a12
0
J-Lh(t, Wh(t), >.) . dwh(t)
2~2 /,T Il'h(t,wh(t), ,\)I'dt} .
-
(17)
Equations (9) and (10) can then be written in the form
Z;(wo) = EQ~ [Zolh(', wo)],
(18)
1
>.
v\ log Zo>. (wo) = E Q).h [Z;lh("WO)
Z;(wo) yo >.log Z (', wo) ,
where
Z;lh(W)
Z>'(w)
= Z~(Wh)
= exp {
II
1
a2
iT
0
-
(19)
J-Lo(t,w(t), >.) . dwo(t)
2~' /,T lI'o(t, w( t), ,\) I' dt }
(20)
.
Note the expectations are now performed using the measure Q~. We can easily
generate samples according to Q~ by numerically integrating equation (15), and in
practice this leads to more efficient Monte Carlo approximations of the likelihood
and gradient.
3
Example: Noisy Sinusoidal Detection
This problem is a simple example of using diffusion networks for signal detection.
The task was to detect a sinusoid in the presence of additive Gaussian noise. Stimuli
were generated according to the following process
Y(t,w) = 1A(w).!.sin(47l't)
7l'
+ B(t,w),
(21)
where t E [0,1/2]. Here Y is assumed anchored in a probability space (0, F, P)
large enough to accommodate the event A which indicates a signal or noise trial.
Note that under P, B is a Brownian motion on Cd independent of A.
A model was optimized using 100 samples of equation (21) given W E A, i.e., 100
stimuli containing a signal. The model had four hidden units and one observable
unit (n = 5, d = 1). The drift of the model was given by
J-L(t, x, >.) = () + W . g(x),
gj(x)
= 1+
1
e-Zj
,
j E {I, 2, 3, 4, 5},
(22)
Learning Path Distributions Using Nonequilibriwn Diffusion Networks
603
ROC Curve, Sinewave Detection Problem
...?.---A----~----~---?~?? ??-?T-.- ..-. d'=1.89
0.8
---.----
1" .......
l"
!
0.6
,I
t
0.4
rf
~
0.2
o
L -_ _ _ _
o
~
__
~~
0.2
__
~
0.4
____
0.6
~
____
~
0.8
hit rate
Figure 2: Receiver operating characteristic (ROC) curve for a diffusion network
performing a signal detection task involving noisy sinusoids, Dotted line: Detection
performance estimated numerically using 10000 novel stimuli. Solid line: Best fit
curve corresponding to d' = 1.89. This value of d' corresponds to performance
within 1.5% of the Bayesian limit.
where 0 E R5 and W is a 5x5 real-valued connection matrix. In this case ~ =
{{Oi}, {Wij }, i,j = 1, ... ,5}. The connections from output to hidden units were set
to zero, allowing use of the more efficient techniques for factorial networks described
above. The initial distribution for the model was a 8-function at (1, -1, 1, -1,0).
The model was numerically simulated with ilt = 0.01, and 100 hidden samples were
used to approximate the likelihood and gradient of the log- likelihood, according to
equations (18) and (19). The conjugate gradient algorithm was used for training,
with the log-likelihood of the data as the cost function.
Once training was complete, the parameter estimation was tested using 10000 novel
stimuli and the following procedure. Given a new stimuli y we used the model to
estimate the likelihood Zo(Y I A) ~ Z~(Y), where ~ is the parameter vector at the
end of training. The decision rule employed was
D(Y) = {sig.nal
nOIse
if Zo(~ I A)
otherwIse,
> b,
(23)
where b E R is a bias term representing assumptions about the apriori probability
of a signal trial. By sweeping across different values of b the receiver-operator
characteristic (ROC) curve is generated. This curve shows how the probability of
a hit, P(D = signal I A), and the probability of a false alarm, P(D = signal I AC),
are related. From this curve the parameter d', a measure of sensitivity independent
of apriori assumptions, can be estimated. Figure 2 shows the ROC curve as found
by numerical simulation, and the curve obtained by the best fit value d' = 1.89.
This value of d' corresponds to a 82.7% correct detection rate for equal prior signal
probabilities.
The theoretically ideal observer can be derived for this problem, since the profile of
the unperturbed signal is known exactly (Poor, 1994, p. 278ff). For this problem
the optimal observer achieves d'max = 2, which implies at equal probabilities for
signal and noise trials, the Bayesian limit corresponds to a 84.1 % correct detection
rate. The detection system based upon the diffusion network is therefore operating
close to the Bayesian limit, but was designed using only implicit information, i.e.,
100 training examples, about the structure of the signal to be detected, in contrast
to the explicit information required to design the optimal Bayesian classifier.
| 1438 |@word trial:3 open:1 simulation:1 dwh:1 solid:1 accommodate:1 initial:3 hereafter:1 denoting:2 interestingly:1 activation:2 dx:1 must:3 written:1 realistic:1 additive:1 numerical:1 shape:1 sdes:2 designed:1 isard:2 filtered:1 provides:1 math:1 node:4 sigmoidal:2 simpler:1 differential:2 prove:1 theoretically:1 karatzas:3 becomes:1 kind:1 interpreted:1 differentiation:2 temporal:2 growth:1 exactly:1 classifier:1 rm:1 scaled:1 hit:2 control:1 unit:3 engineering:1 local:2 sd:1 limit:3 path:22 ease:1 unique:1 practice:2 movellan:4 backpropagation:1 procedure:1 logz:3 thought:1 integrating:1 refers:1 undesirable:2 close:1 operator:1 bh:1 optimize:2 equivalent:2 measurable:1 conventional:2 deterministic:2 conceptualized:1 send:1 williams:4 rule:1 oh:1 dw:1 notion:1 diego:3 play:1 paralleling:1 sig:1 velocity:1 approximated:1 jk:1 role:1 ft:1 mentioned:1 environment:1 dynamic:8 depend:2 algebra:2 upon:3 efficiency:1 observables:4 po:3 easily:1 represented:1 zo:3 monte:2 cogsci:2 detected:1 whose:1 valued:7 otherwise:1 noisy:2 sequence:2 propose:1 intuitive:1 object:1 coupling:1 recurrent:3 ac:1 ij:2 eq:1 implies:1 correct:2 filter:1 stochastic:5 a12:1 ao:7 generalization:3 adb:1 considered:1 blake:2 exp:4 equilibrium:1 driving:1 achieves:1 a2:1 fh:5 purpose:1 uniqueness:1 estimation:4 wet:1 weighted:1 unfolding:1 gaussian:2 derived:1 yo:1 likelihood:10 indicates:1 contrast:1 sense:1 detect:1 inference:1 entire:2 typically:1 hidden:19 quasi:1 wij:1 pixel:1 orientation:1 flexible:1 denoted:3 marginal:1 field:1 once:1 apriori:2 equal:2 sampling:1 represents:2 r5:1 stimulus:5 continuoustime:1 detection:8 interest:5 fd:2 tj:1 succeeded:1 integral:2 experience:1 lh:5 tree:2 divide:1 theoretical:1 modeling:1 markovian:1 ordinary:1 cost:2 nonequilibrium:2 density:17 sensitivity:1 probabilistic:2 mode1:1 jo:1 containing:1 henceforth:1 cognitive:3 derivative:3 li:1 potential:1 sinusoidal:1 coefficient:1 satisfy:1 explicitly:1 depends:1 performed:1 lot:1 observer:2 il:1 oi:1 lew:1 variance:2 characteristic:2 yield:2 weak:2 identification:1 bayesian:4 carlo:2 ah:6 fo:4 girsanov:2 synaptic:2 naturally:1 associated:1 treatment:1 wh:7 javier:1 dt:6 improved:1 just:2 xa:1 shreve:3 implicit:1 d:1 hand:1 defines:1 gray:1 sinusoid:2 sin:1 x5:1 ijl:1 theoretic:1 complete:1 motion:7 novel:2 nh:2 jl:4 interpretation:1 numerically:3 expressing:1 rd:1 seldom:1 mathematics:1 language:1 had:1 operating:2 gj:1 brownian:7 optimizing:1 jolla:3 care:1 employed:1 signal:14 ii:2 desirable:2 reduces:1 lin:1 involving:3 expectation:1 condensation:1 addition:1 whereas:1 ode:1 addressed:1 interval:1 crucial:1 collides:1 induced:1 member:1 ee:1 presence:1 ideal:1 enough:1 variety:2 fit:2 idea:1 cn:6 wo:15 quasilinear:1 useful:2 collision:1 factorial:3 ph:3 hardware:2 generate:1 zj:1 dotted:1 neuroscience:2 estimated:2 discrete:1 four:1 erp:1 nal:1 diffusion:20 ilt:1 throughout:1 family:1 decision:2 radon:3 correspondence:1 adapted:1 strength:1 constraint:1 aspect:1 integrand:1 formulating:1 performing:2 ern:1 department:3 according:5 eph:2 poor:2 conjugate:1 across:1 taken:1 equation:12 needed:1 end:1 available:1 obey:1 appropriate:1 jl2:1 rp:3 existence:1 ensure:1 approximating:1 quantity:1 usual:1 gradient:16 dp:1 simulated:1 whom:1 ruth:1 kalman:1 ratio:1 potentially:1 sigma:1 filtration:2 design:3 allowing:1 observation:1 dispersion:2 markov:3 defining:1 rn:3 ucsd:3 sweeping:1 drift:9 namely:2 required:1 optimized:2 connection:3 california:3 distinction:1 qa:2 address:1 regime:1 rf:1 max:1 suitable:3 event:1 natural:2 localizes:1 representing:1 lk:1 prior:1 literature:1 zh:1 law:4 filtering:1 proven:2 ixl:1 sufficient:3 dq:1 nikodym:3 cd:3 lo:7 formal:1 bias:1 wide:1 distributed:1 feedback:1 curve:8 valid:1 rich:1 contour:1 san:3 functionals:1 approximate:2 observable:13 receiver:2 assumed:1 continuous:16 iterative:1 mineiro:3 anchored:1 jlx:1 ca:3 contributes:1 domain:1 noise:7 paul:1 alarm:1 profile:1 borel:3 ff:3 roc:4 position:1 explicit:3 perceptual:1 ito:2 formula:5 theorem:2 showing:1 sinewave:1 unperturbed:1 explored:1 dominates:1 concern:1 dl:1 false:1 importance:1 illustrates:1 clamping:1 lt:2 visual:1 tracking:1 bo:1 corresponds:3 determines:1 satisfies:2 environmental:1 dh:1 viewed:1 formulated:1 called:2 total:1 la:3 evaluate:1 tested:1 |
481 | 1,439 | Neural Basis of Object-Centered
Representations
Sophie Deneve and Alexandre Pouget
Georgetown Institute for Computational and Cognitive Sciences
Georgetown University
Washington, DC 20007-2197
sophie, alex@giccs.georgetown.edu
Abstract
We present a neural model that can perform eye movements to a
particular side of an object regardless of the position and orientation of the object in space, a generalization of a task which has
been recently used by Olson and Gettner [4] to investigate the neural structure of object-centered representations. Our model uses an
intermediate representation in which units have oculocentric receptive fields- just like collicular neurons- whose gain is modulated by
the side of the object to which the movement is directed, as well as
the orientation of the object. We show that these gain modulations
are consistent with Olson and Gettner's single cell recordings in the
supplementary eye field. This demonstrates that it is possible to
perform an object-centered task without a representation involving an object-centered map, viz., without neurons whose receptive
fields are defined in object-centered coordinates. We also show that
the same approach can account for object-centered neglect, a situation in which patients with a right parietal lesion neglect the left
side of objects regardless of the orientation of the objects.
Several authors have argued that tasks such as object recognition [3] and manipulation [4] are easier to perform if the object is represented in object-centered coordinates, a representation in which the subparts of the object are encoded with respect
to a frame of reference centered on the object. Compelling evidence for the existence
of such representations in the cortex comes from experiments on hemineglect- a
neurological syndrome resulting from unilateral lesions of the parietal cortex such
that a right lesion, for example, leads patients to ignore stimuli located on the left
side of their egocentric space. Recently, Driver et al. (1994) showed that the deficit
can also be object-centered. Hence, hemineglect patients can detect a gap in the
upper edge of a triangle when this gap is associated with the right side of the object
25
Neural Basis of Object-Centered Representations
A.
B.
Obje~tnt.recl
cu?"'S
./
Spotiol
cueing
I
I
!1
e
?
:
1
' I
,, !'2
e1
2
C.
I
Three possible locations:
"
eI
;.~
,
""~
;
1
I
e
I,
.1
I
I
3
1
[4
I
I
!
eI!
1!
? !
i
1
l5/
.
?
Time
Figure 1: A- Driver et al. (1994) experiment demonstrating object-centered neglect.
Subjects were asked to detect a gap in the upper part of the middle triangle, while
fixating at the cross, when the overall figure is tilted clockwise (top) or counterclockwise (bottom). Patients perform worse for the clockwise condition, when the gap
is perceived to be on the left of the overall figure. B- Sequence of screens presented
on each trial in Olson and Gettner experiment (1995). 1- Fixation, 2- apparition of
a cue indicating where the saccade should go, either in object-centered coordinates
(object-centered cueing), or in screen coordinates (spatial cueing), 3- delay period,
4- apparition of the bar in one of three possible locations (dotted lines), and 5- saccade to the cued location. C- Schematic response of an SEF neuron for 4 different
conditions. Adapted from [4].
but not when it belongs to the left side (figure I-A).
What could be the neural basis of these object-centered representations? The simplest scheme would involve neurons with receptIve fields defined in object-centered
coordinates, i.e., the cells respond to a particular side of an object regardless of
the position and orientation of the object. A recent experiment by Olson and Gettner (1995) supports this possibility. They recorded the activity of neurons in the
supplementary eye field (SEF) while the monkey was performing object-directed
saccades. The task consisted of making a saccade to the right or left side of a
bar, independently of the position of the bar on the screen and according to the
instruction provided by a visual cue. For instance, the cue corresponding to the
instruction 'Go to the right side of the bar' was provided by highlighting the right
side of a small bar briefly flashed at the beginning of the trial (step 2 in figure I-B).
By changing the position of the object on the screen, it is possible to compare
the activity of a neuron for movements involving different saccade directions but
directed to the same side of the object, and vice-versa, for movements involving
the same saccade direction but directed to opposite sides of the object. Olson and
Gettner found that many neurons responded more prior to saccades directed to
a particular side of the bar, independently of the direction of the saccades. For
example, some neurons responded more for an upward right saccade directed to the
left side of the bar but not at all for the same upward right saccade directed to the
right side of the bar (column 1 and 3, figure I-C). This would suggest that these
neurons have bar-centered receptive! fields, i.e., their receptive fields are centered
lwe use the term receptive field in a general sense, meaning either receptive or motor
s. Deneve andA. Pouget
26
...
"Do E
-;:"
i~!:
?c
o~
~.
~
mop
Figure 2: Schematic structure of the network with activity patterns in response
to the horizontal bar shown in the VI map and the command 'Go to the right'.
Only one SEF map is active in this case, the one selective to the right edge of the
bar (where right is defined in retinal coordinates), object orientation of 0? and the
command 'Go to the right'. The letter a, b, c and d indicate which map would be
active for the same command but for various orientations of the object, respectively,
0?, 90?, 180?, 270?. The dotted lines on the maps indicate the outline of the bar.
Only a few representative connections are shown.
on the bar and not on the retina. This would correspond to what we will call an
explicit object-centered representation.
We argue in this paper that these data are compatible with a different type of
representation which is more suitable for the task performed by the monkey. We
describe a neural network which can perform a saccade to the right, or left, boundary
of an object, regardless of its orientation, position or size- a generalization of the
task used by Olson and Gettner. This network uses units with receptive fields
defined in oculocentric coordinates, i.e., they are selective for the direction and
amplitude of saccades with respect to the fixation point, just like collicular neurons.
These tuning curves, however, are also modulated by two types of signals, the
orientation of the object, and the command indicating the side of the object to
which the saccade should be directed. We show that these response properties are
compatible with the Olson and Gettner data and provide predictions for future
experiments. We also show that a simulated lesion leads to object-centered neglect
as observed by Driver et al. (1994).
1
Network Architecture
The network performs a mapping from the image of the bar and the command
(indicating the side of the object to which the saccade must be directed) to the
appropriate motor command in oculocentric coordinates (the kind of command
observed in the frontal eye field, FEF). We use a bar whose left and right sides are
defined with respect to a a triangle appearing on the top of the bar (see figure 2).
The network is composed of four parts. The first two parts of the network models the
field.
Neural Basis of Object-Centered Representations
27
lower areas in visual cortex, where visual features are segmented within retinotopic
maps. In the first layer, the image is projected on a very simple VI-like map (10
by 10 neurons with activity equal to one if a visual feature appears within their
receptive field, and zero otherwise). The second part on the network contains 4
different V2 retinotopic maps, responding respectively to the right, left, top and
bottom boundary of the bar.
This model of V2 is intended to reproduce the response properties of a subset of
cells recently discovered by Zhou et al. (1996). These cells respond to oriented
edges, like VI cells, but when the edge belongs to a closed figure, they also show
a selectivity for the side on which the figure appears with respect to the edge. For
example, a cell might respond to a vertical edge only if this edge is on the right
side of the figure but not on the left (where right and left are defined with respect
to the viewer, not the object itself). This was observed for any orientation of the
edge, but we limit ourselves in this model to horizontal and vertical ones.
The third part of the network models the SEF and is divided into 4 groups of 4
maps, each group receiving connections from the corresponding map in V2 (figure 2).
Within each group of maps, visual activity is modulated by signals related to the
orientation of the object (assumed to be computed in temporal cortex) such that
each of the 4 maps respond best for one particular orientation (respectively 0?, 90?,
180? and 270?). For example, a neuron in the second map of the top group responds
maximally if: 1- there is an edge in its receptive field and the figure is below, and
2- the object has an orientation of 90? counterclockwise. Note that this situation
arises only if the left side of the object appears in the cell's receptive field; it will
never occur for the right side. However, the cell is only partially selective to the
left side of the object, e.g., it does not respond when the left side is in the retinal
receptive field and the orientation of the object is 270? counterclockwise.
These collection of responses can be used to generate an object-centered saccade by
selecting the maps which are partially selective for the side of the object specified
by the command. This is implemented in our network by modulating the SEF maps
by signals related to the command. For example, the unit encoding 'go to the right'
send a positive weight to any map compatible with the right side of the object while
inhibiting the other maps (figure 2).
Therefore, the activity, Bfj' of a neuron at position ij on the map k in the SEF is
the product of three functIOns:
where Vij is the visual receptive field from the V2 map, h((J) is a gaussian function of orientation centered on the cell preferred orientation, (Jk, and gk(C) is the
modulation by the command unit.
Olson and Gettner also used a condition with spatial cueing, viz., the command was
provided by a spatial cue indicating where the saccade should go, as opposed to an
object-centered instruction (see figure I-B). We modeled this condition by simply
multiplying the activity of neurons coding for this location in all the SEF maps by
a fixed constant (10 in the simulations presented here). We also assume that there
is no modulation by orientation of the object since this information is irrelevant in
this experimental condition.
Finally, the fourth part of the network consists of an oculocentric map similar to
the one found in the frontal eye field (FEF) or superior colliculus (SC) in which
the command for the saccade is generated in oculocentric coordinates. The activity
in the output map, {Oij}, is obtained by simply summing the activities of all the
S. Deneve and A. Pouget
28
... ...,-G: ___...
,
I
06L __ _
..t..
I
I
..
; ..t..
.6J
[!1]
B.
Figure 3: A- Polar plots showing the selectivity for saccade direction of a representative SEF units. The first three plots are for various combinations of command (L:
left, R: right, P: spatial cueing) and object orientation. The left plot corresponds
to saccades to a single dot. B- Data for the same unit but for a subset of the
conditions. The first four columns can be directly compared to the experimental
data plotted in figure 1-C from Olson and Gettner. The 5th and 6th columns show
responses when the object is inverted. The seventh columns corresponds to spatial
cueing.
SEF maps. The result is typically a broad two dimensional bell-shaped pattern of
activity from which one can read out the horizontal and vertical components, Xs
and Ys, of the intended saccade by applying a center-of-mass operator.
(1)
2
Results
This network is able to generate a saccade to the right or to left of a bar, whatever
its position, size and orientation. This architecture is basically like a radial basis
function network, i.e., a look-up table with broad tuning curves allowing for interpolation. Consequently, one or two of the SEF maps light up at the appropriate
location for any combination of the command and, position and orientation of the
bar.
Neurons in the SEF maps have the following property: 1- they have an invariant
tuning curve for the direction (and amplitude) of saccadic eye movements in oculacentric coordinates, just like neurons in the FEF or the SC, and 2- the gain of this
tuning curve is modulated by the orientation of the object as well as the command.
Figure 3-A shows how the tuning curve for saccade direction of one particular unit
varies as a function of these two variables. In this particular case, the cell responds
best to a right-upward saccade directed to the left side of the object when the object
is horizontal. Therefore, the SEF units in our model do not have an invariant receptive field in object-centered coordinates but, nevertheless, the gain modulation is
sufficient to perform the object-centered saccade. We predict that similar response
properties should be found in the SEF, and perhaps parietal cortex, of a monkey
trained on a task analogous to the one described here.
Since Olson and Gettner (1995) tested only three positions of the bar and held its
Neural Basis of Object-Centered Representations
29
orientation constant, we cannot determine from their data whether SEF neurons
are gain modulated in the way we just described. However, they found that all
the SEF neurons had oculocentric receptive fields when tested on saccades to a
single dot (personal communication), an observation which is consistent with our
hypothesis (see fourth plot in figure 3-A) while being difficult to reconcile with
an explicit object-centered representation. Second, if we sample our data for the
conditions used by Olson and Gettner, we find that our units behave like real SEF
cells. The first four columns of figure 3-B shows a unit with the same response
properties as the cell represented in figure I-C. Figure 3-B also shows the response
of the same unit when the object is upside down and when we use a spatial cue.
Note that this unit responds to the left side of the bar when the object is upright
(1st column), but not when it is rotated by 1800 (5th column), unless we used a
spatial cue (7th column). The absence of response for the inverted object is due to
the selectivity of the cell to orientation. The cell nevertheless responds in the spatial
cueing conditions because we have assumed that orientation does not modulate the
activity of the units in this case, since it is irrelevant to the task.
Therefore, the gain modulation observed in our units is consistent with available
experimental data and makes predictions for future experiments.
3
Simulation of Neglect
Our representational scheme can account for neglect if the parietal cortex contains
gain modulated cells like the ones we have described and if each cortical hemisphere
contains more units selective for the contralateral side of space. This is known to
be the case for the retinal input; hence most cells in the left hemisphere have their
receptive field on the right hemiretina. We propose that the left hemisphere also
over-represents the right side of objects and vice-versa (where right is defined in
object-centered coordinates).
Recall that the SEF maps in our model are partially selective for the side of objects.
A hemispheric preference for the contralateral side of objects could therefore be
achieved by having all the maps responding to the left side of objects in the right
hemisphere. Clearly, in this case, a right lesion would lead to left object-centered
neglect; our network would no longer be able to perform a saccade to the left side
of an object.
IT we add the retinal gradient and make the previous gradient not quite as binary,
then we predict that a left lesion leads to a syndrome in which the network has
difficulty with sac cades to the left side of an object but more so if the object is
shown in the left hemiretina. Preliminary data from Olson and Gettner (personal
communication) are compatible with this prediction.
The same model can also account for Driver et al. (1994) experiment depicted in
figure I-A. If the hemispheric gradients are as we propose, a right parietal lesion
would lead to a situation in which the overall activity associated with the gap,
i.e., the summed activity of all the neurons responding to this retinal location, is
greater when the object is rotated counterclockwise- the condition in which the
gap is perceived as belonging to the right side of the object- than in the clockwise
condition. This activity difference, which can be thought as being a difference in the
saliency of the upper edge of the triangle, may be sufficient to account for patients'
performance.
Note that object-centered neglect should be observed only if the orientation of the
object is taken into consideration by the SEF units. If the experimental conditions
S. Deneve and A. Pouget
30
are such that the orientation of the object can be ignored by the subject -a situation similar to the spatial cueing condition modeled here- we do not expect to
observe neglect. This may explain why several groups (such as Farah et al., 1990)
have failed to find object-centered neglect even though they used a paradigm similar
to Driver et al. (1994).
4
Discussion
We have demonstrated how object-centered saccades can be performed using neurons with oculocentric receptive fields, gain modulated by the orientation of the
object and the command. The same representational scheme can also account for
object-centered neglect without invoking an explicit object-centered representation,
i.e., representation in which neurons' receptive fields are defined in object-centered
coordinates. The gain modulation by the command is consistent with the single cell
data available [4], but the modulation by the orientation of the object is a prediction
for future experiments.
Whether explicit object-centered representations exist, remains an empirical issue.
In some cases, such representations would be computationally inefficient. In the Olson and Gettner experiment, for instance, having a stage in which motor commands
are specified in object-centered coordinates does not simplify the task. Encoding
the motor command in object-centered coordinates in the intermediate stage of
processing requires (i) recoding the sensory input into object-centered coordinates,
(ii) decoding the object-centered command into an oculocentric command, which
is ultimately what the oculomotor system needs to ,generate the appropriate saccade. Each of these steps are computationally as complex as performing the overall
transformation directly as we have done in this paper.
Therefore, gain modulation provides a simple algorithm for performing objectcentered saccades. Interestingly, the same basic mechanism underlies spatial representations in other frames of reference, such as head-centered and body-centered.
We have shown previously that these responses can be formalized as being basis
functions of their sensory and postures inputs, a set of function which is particularly useful for sensory-motor transformations [5]. The same result applies to the
SEF neurons considered in this paper, suggesting that basis functions may provide
a unified theory of spatial representations in any spatial frame of reference.
References
[1] J. Driver, G. Baylis, S. Goodrich, and R. Rafal. Axis-based neglect of visual
shapes. Neuropsychologia, 32{11}:1353-1365, 1994.
[2] M. Farah, J. Brunn, A. Wong, M. Wallace, and P. Carpenter. Frames of reference for allocating attention to space: evidence from the neglect syndrome.
Neuropsychologia, 28(4):335-47, 1990.
[3] G. Hinton. Mapping part-whole hierarchies into connectionist networks. Artificial Intelligence, 46(1}:47-76, 1990.
[4] C. Olson and S. Gettner. Object-centered direction selectivity in the macaque
supplementary eye. Science, 269:985-988, 1995.
[5] A. Pouget and T. Sejnowski. Spatial transformations in the parietal cortex using
basis functions. Journal of Cognitive Neuroscience, 9(2):222-237, 1997.
[6] H. Zhou, H. Friedman, and R. von der Heydt. Edge selective cells code for
figure-ground in area V2 of monkey visual cortex. In Society For Neuroscience
Abstracts, volume 22, page 160.1, 1996.
| 1439 |@word trial:2 cu:1 briefly:1 middle:1 instruction:3 simulation:2 invoking:1 contains:3 selecting:1 interestingly:1 must:1 tilted:1 shape:1 motor:5 plot:4 cue:6 intelligence:1 beginning:1 provides:1 location:6 preference:1 driver:6 consists:1 fixation:2 wallace:1 provided:3 retinotopic:2 mass:1 what:3 kind:1 monkey:4 unified:1 transformation:3 temporal:1 demonstrates:1 whatever:1 unit:16 positive:1 limit:1 encoding:2 modulation:8 interpolation:1 might:1 directed:10 area:2 empirical:1 bell:1 thought:1 radial:1 suggest:1 cannot:1 operator:1 applying:1 wong:1 map:27 demonstrated:1 center:1 send:1 go:6 regardless:4 attention:1 independently:2 formalized:1 pouget:5 sac:1 coordinate:16 analogous:1 hierarchy:1 us:2 hypothesis:1 recognition:1 giccs:1 located:1 jk:1 particularly:1 bottom:2 observed:5 movement:5 sef:19 asked:1 personal:2 ultimately:1 trained:1 objectcentered:1 basis:9 triangle:4 represented:2 various:2 brunn:1 describe:1 sejnowski:1 goodrich:1 artificial:1 sc:2 whose:3 encoded:1 supplementary:3 quite:1 otherwise:1 itself:1 sequence:1 propose:2 product:1 representational:2 olson:14 hemiretina:2 rotated:2 object:90 cued:1 ij:1 implemented:1 come:1 indicate:2 direction:8 centered:43 argued:1 generalization:2 preliminary:1 viewer:1 considered:1 ground:1 baylis:1 mapping:2 predict:2 mop:1 inhibiting:1 perceived:2 polar:1 modulating:1 vice:2 clearly:1 gaussian:1 zhou:2 command:21 viz:2 detect:2 sense:1 typically:1 selective:7 reproduce:1 upward:3 overall:4 issue:1 orientation:27 spatial:13 summed:1 field:22 equal:1 never:1 shaped:1 washington:1 having:2 represents:1 broad:2 look:1 future:3 connectionist:1 stimulus:1 simplify:1 few:1 retina:1 oriented:1 composed:1 intended:2 ourselves:1 fef:3 friedman:1 investigate:1 possibility:1 light:1 held:1 allocating:1 edge:11 unless:1 plotted:1 instance:2 column:8 lwe:1 compelling:1 subset:2 contralateral:2 delay:1 seventh:1 varies:1 st:1 l5:1 receiving:1 collicular:2 decoding:1 von:1 recorded:1 opposed:1 rafal:1 worse:1 cognitive:2 inefficient:1 account:5 fixating:1 suggesting:1 retinal:5 coding:1 vi:3 performed:2 closed:1 farah:2 responded:2 correspond:1 saliency:1 basically:1 multiplying:1 explain:1 associated:2 gain:10 cueing:8 recall:1 amplitude:2 appears:3 alexandre:1 response:11 maximally:1 done:1 though:1 just:4 stage:2 horizontal:4 ei:2 perhaps:1 consisted:1 hence:2 read:1 flashed:1 oculocentric:8 hemispheric:2 outline:1 performs:1 meaning:1 image:2 consideration:1 recently:3 superior:1 volume:1 versa:2 tuning:5 had:1 dot:2 cortex:8 longer:1 add:1 showed:1 recent:1 belongs:2 irrelevant:2 hemisphere:4 manipulation:1 selectivity:4 binary:1 der:1 inverted:2 greater:1 syndrome:3 determine:1 paradigm:1 period:1 clockwise:3 ii:1 signal:3 upside:1 segmented:1 cross:1 divided:1 e1:1 y:1 schematic:2 prediction:4 involving:3 underlies:1 basic:1 patient:5 achieved:1 cell:18 recording:1 subject:2 counterclockwise:4 neuropsychologia:2 call:1 intermediate:2 architecture:2 opposite:1 whether:2 unilateral:1 ignored:1 useful:1 involve:1 simplest:1 generate:3 exist:1 dotted:2 neuroscience:2 subpart:1 group:5 four:3 demonstrating:1 nevertheless:2 changing:1 deneve:4 egocentric:1 colliculus:1 letter:1 fourth:2 respond:5 layer:1 activity:14 adapted:1 occur:1 bfj:1 alex:1 performing:3 according:1 combination:2 belonging:1 making:1 invariant:2 taken:1 computationally:2 remains:1 previously:1 mechanism:1 available:2 observe:1 v2:5 appropriate:3 appearing:1 existence:1 top:4 responding:3 neglect:13 society:1 posture:1 receptive:18 saccadic:1 responds:4 gradient:3 deficit:1 simulated:1 argue:1 code:1 modeled:2 difficult:1 hemineglect:2 gk:1 perform:7 allowing:1 upper:3 vertical:3 neuron:22 observation:1 behave:1 parietal:6 situation:4 hinton:1 communication:2 head:1 dc:1 frame:4 tnt:1 discovered:1 heydt:1 specified:2 connection:2 macaque:1 able:2 bar:21 below:1 pattern:2 oculomotor:1 suitable:1 difficulty:1 oij:1 scheme:3 eye:7 axis:1 prior:1 georgetown:3 expect:1 sufficient:2 consistent:4 vij:1 compatible:4 side:36 institute:1 recoding:1 boundary:2 curve:5 cortical:1 sensory:3 author:1 collection:1 projected:1 ignore:1 preferred:1 active:2 anda:1 summing:1 assumed:2 why:1 table:1 complex:1 whole:1 reconcile:1 lesion:7 body:1 carpenter:1 representative:2 screen:4 position:9 explicit:4 third:1 down:1 showing:1 x:1 evidence:2 gap:6 easier:1 depicted:1 simply:2 visual:8 failed:1 highlighting:1 neurological:1 partially:3 saccade:29 applies:1 corresponds:2 modulate:1 consequently:1 absence:1 upright:1 sophie:2 experimental:4 indicating:4 support:1 modulated:7 arises:1 frontal:2 tested:2 |
482 | 144 | 662
A PASSIVE SHARED ELEMENT ANALOG
ELECTRICAL COCHLEA
Joe Eisenberg
Bioeng. Group
U.C. Berkeley
David Feld
Dept. Elect. Eng.
207-30 Cory Hall
U.C. Berkeley
Berkeley, CA. 94720
Edwin Lewis
Dept Elect. Eng.
U.C. Berkeley
ABSTRACT
We present a simplified model of the micromechanics of the human
cochlea, realized with electrical elements. Simulation of the model
shows that it retains four signal processing features whose importance
we argue on the basis of engineering logic and evolutionary evidence.
Furthermore, just as the cochlea does, the model achieves massively
parallel signal processing in a structurally economic way, by means of
shared elements. By extracting what we believe are the five essential
features of the cochlea, we hope to design a useful front-end filter to
process acoustic images and to obtain a better understanding of the
auditory system.
INTRODUCTION
Results of psychoacoustical and physiological experiments in humans indicate that
the auditory system creates acoustic images via massively parallel neural computations.
These computations enable the brain to perform voice detection, sound localization, and
many other complex taskS. For example, by recording a random signal with a wide range
of frequency components, and playing this signal simultaneously through both channels
of a stereo headset, one causes the brain to create an acoustical image of a "shsh" sound in
the center of the head. Delaying the presentation of just one frequency component in the
random signal going to one ear and simultaneously playing the original signal to the
other ear, one would still have the image of a "shsh" in the center of the head; however if
one mentally searches the acoustical image space carefully, a clear tone can be found far
off to one side of the head. The frequency of this tone will be that of the component with
the time delay to one ear. Both ears still are receiving wide-band random signals. The
isolated tone will not be perceptible from the signal to either ear alone; but with both
signals together, the brain has enough data to isolate the delayed tone in an acoustical
image. The brain achieves this by massively parallel neural computation.
Because the acoustic front-end fIlter for the brain is the cochlea, people have proposed
that analogs of the cochlea might serve well as front-end fIlters for man-made processors
of acoustical images (Lyon, Mead, 1988). If we were to base a cochlear analog on
current biophysical models of this structure, it would be extraordinarily complicated and
extremely difficult to realize with hardware. Because of this, we want to start with a
cochlear model that incorporates a minimum set of essential ingredients. The ears of
lower vertebrates, such as alligators and frogs, provide some clues to help identify those
ingredients. These animals presumably have to compute acoustic images similar to ours,
A Passive Shared Element Analog Electrical Cochlea
but they do not have cochleas. The acoustic front-end filters in the ears of these animals
evolved independently and in parallel to the evolution of the cochlea. Nevertheless, those
front-end filters share four functional properties with the part of the cochlea which
responds to the lower 7 out of 10 octaves of hearing (20 Hz. to 2560 Hz.):
1.
2.
3.
4.
They are multichannel ftlters with each channel
covering a different part of the frequency spectrum.
Each channel is a relatively broad-band frequency
filter.
Each filter has an extremely steep high-frequency
band edge (typically 60 to 200 db/oct).
Each filter has nearly linear phase shift as a function
of frequency, within its passband.
The front-end acoustical filters of lower vertebrates also have at least one structural feature
in common with the cochlea: namely, the various filter channels share their dynamic
components. This is the fifth property we choose to include. Properties 1 and 3 provide
good resolution in frequency; properties 2 and 4 are what filter designers would add to
provide good resolution in time.
In order to compute acoustical images with the neural networks in our brain, we need
both kinds of resolution: time and frequency. Shared elements, a structural feature, has
obvious advantages for economy of construction. The fact that evolution has come to
these same front-end filter properties repeatedly suggests that these properties have
compelling advantages with respect to an animal's survival. We submit that we can
realize all of these properties very well with the simplest of the modern cochlear models,
namely that of Joseph Zwislocki (1965). This is a transmission line model made
entirely of passive elements.
Figure 1 - Drawing of the ear with the cochlea represented by an
electrical analog.
663
664
Feld, Eisenberg and Lewis
A COCHLEAR MODEL
In order to illustrate Zwislocki's model, a quick review of the mechanics of the
cochlea is useful. Figure 1 depicts the ear with the cochlea represented by an electrical
analog. A sound pressure wave enters the outer ear and strikes the ear drum which, in
turn, causes the three bones of the middle ear to vibrate. The last bone, known as the
stapes, is connected to the oval window (represented in figure 1 by the voltage source at
the beginning of the electrical analog), where the acoustic energy enters the cochlea.
As the acoustic energy is transferred to the oval window, a fluid-mechanical wave is
formed along a structure known as the basilar membrane ([his membrane and the
surrounding fluid is represented by the series and shunt circuit elements of figure 1). As
the basilar membrane vibrates, the acoustical signal is transduced to neural impulses
which travel along the auditory nerve, carrying the data used by the nervous system to
compute auditory images. Figure 2 is taken from a paper by Zweig et al. (1976), and
depicts an uncoiled cochlea. As the fluid-mechanical wave travels through the cochlea: 1)
The wave gradually slows down, and 2) The higher-frequency components of the wave are
absorbed, leaving an increasingly narrower band of low-frequency components proceeding
on toward the far end of the cochlea. If we were to uncoil and enlarge the basilar
membrane it would look like a swim fin (figure 3). If we now were to push on the
basilar membrane, it would push back like a spring. It is most compliant at the wide,
thin end of the fin. Thus as one moves along the basilar membrane from its basal to
apical end, its compliance increases. Zwislocki's transmission-line model was tapered in
this same way.
Scala
vestibuli
Helicotrema
Figure 2 - Uncoiled cochlea (Zweig. 1976).
A Passive Shared Element Analog Electrical Cochlea
Basal End
Figure 3 - Simplified uncoiled and enlarged drawing of the basilar
membrane.
Zwislocki's model of the cochlea is a distributed parameter transmission line.
Figure 4 shows a lumped electrical analog of the model. The series elements (L 1,,, .Lo)
represent the local inertia of the cochlear fluid. The shunt capacitive elements (CI,...Co)
represent the local compliance of the basilar membrane. The shunt resistive elements
(R 1,...Rn) represent the local viscous resistance of the basilar membrane and associated
fluid. The model has one input and a huge 'number of outputs. The input, sound
pressure at the oval window, is represented here as a voltage source. The outputs are
either the displacements or the velocities of the various regions of the basilar membrane.
Figure 4 - Transmission line model of the cochlea represented as an
electrical circuit.
In the electrical analog, shown in figure 4, we have selected velocities as the outputs (in
order to compare our data to real neural tuning curves) and we have represented those
velocities as the currents (11 , .. .In ). The original analysis of Zwislocki's tapered
transmission line model did not produce the steep high frequency band edges that are
observed in real cochleas. This deficiency was a major driving force behind the early
development of more complex cochlear models. Recently, it was found that the original
analysis placed the Zwislocki model in an inappropriate mode of operation (Lewis, 1984).
In this mode,determined by the relative parameter values, the high frequency band edges
had very gentle slopes. Solving the partial differential equations for the model with the
help of a commonly used simplification (the WKB approximation), one finds a second
mode of operation. In this mode, the model produces all five of the properties that we
desire, including extraordinarily steep high-frequency band edges.
665
666
Feld, Eisenberg and Lewis
RESULTS
We were curious to know whether or not the newly-found mode of operation, with
its very steep high-frequency band edges, could be found in a finite-element version of the
model. If so, we should be able to realize a lumped, analog version of the Zwislocki
model for use as a practical front-end filter for acoustical image formation and processing.
We decided to implement the finite element model in SPICE. SPICE is a software
package that is used for electrical circuit simulation. Our SPICE model showed the
following: As long as the model was made up of enough segments, and as long as the
elements had appropriate parameter values, the second mode of operation indeed was
available. Furthermore, it was the predominant mode of operation when the parameter
values of the model were matched to biophysical data for the cochlea.
CD
:3
L\.l
:z:
<
Q:
Q:I
"~
::E
~
u
Z
~
~
<
1-0
~
,.
CI)
100000
L OG
(FREQUENC Y)
]/\
I
1.1
" , " ,,,I
U
I
I
",
",I
"
, ...
FREQUENCY (kHz)
Figure 5 - Frequency response of
Figure 6 - Inverted neural tuning curves
the basilar membrane velocity.
from three afferent fibers of a cat cochlea
(Kiang and Moxon, 1974).
Figure 5 shows the magnitude of the electrical analog's response plotted against
frequency on log-log coordinates. The five curves correspond to five different locations
along our model. The cutoff frequencies span approximately seven octaves. Further
adjustments of the parameters will be needed in order to shift these curves to span the
lower seven octaves of human audition. For comparison, figure 6 shows threshold
response curves of a cat cochlea from a paper by Kiang and Moxon (1974). These curves
are inverted intentionally because Kiang and Moxon plotted stimulus threshold vs.
frequency rather than response amplitude vs. frequency. We use these neural tuning
curves for comparison because direct observations of cochlear mechanics have been
limited to the basal end. Furthermore, in the realm of single frequencies and small
signals, Evans has produced compelling evidence that this is a valid comparison (Evans,
in press). These three curves are typical of the lower seven octaves of hearing. One
obvious discrepancy between Kiang and Moxon's data and our results is that our model
does not exhibit the sharp corners occurring at the band edges. The term sharp corner
denotes the fact that the transition between the shallow rising edge and steep falling edge
of a given curve is abrupt i.e. the corner is not rounded.
A Passive Shared Element Analog Electrical Cochlea
Figure 7 shows what happens to the response curve at a single location along our
model as the number of stages is increased. The curve on the right, in figure 7, was
derived with 500 stages and does not change much as we increase the number of stages
indefinitely. Thus the curve represents a convergence of the solution of the lumped
parameter Zwislocki model to the distributed parameter model. The middle curve was
derived with 100 stages and the left-hand curve was derived with 50. In any lumpedelement transmission line, there occurs an artifactual cutoff which occurs roughly at the
point where the given input wavelength exceeds the dimensions of the lumped elements.
If we do not lump the stages in our model finely enough, we observe this artifactual
cutoff as opposed to the true cutoff of Zwislocki's distributed parameter model. This
phenomena is clearly observed in the curve derived from 50 stages and may account for
the sharper comers in response curves from real cochleas. However, in order to make our
finite element model operate in a manner analogous to that of the distributed parameter
Zwislocki model we need approximately 500 stages.
~
c
~
~
~
~
"'
"'
."
~
0
0
~z
~
.
~
~
10
10 0 00 0
L OG !FREQUENCY )
IdB}
10000 0
L OG (FREQUENCY)
(HZ)
Figure 7 - Convergence of cut-
Figure 8 - Frequency response of the
toff points as the number of
branches increase.
basilar membrane velocity without
the Heliocotrema.
A critical element in the Zwislocki model is a terminating resistor, representing the
heliocotrema (see Rh in figure 3). The heliocotrema is a small hole at the end of the
basilar membrane. Figure 8, shows the effects of removing that resistor. The irregular
frequency characteristics are quite different from the experimental data and represent
possibly wild excursions of the basilar membrane.
Figure 9. shows phase data for the Zwislocki model, which is linear as a function of
frequency. Anderson et aI (1971), show similar results in the squirrel monkey.
With lumped-element analysis we are able to obtain temporal as well as spectral
responses. For a temporal waveform such as an acoustic pulse, the linear relationship
between phase and frequency guarantees that those Fourier components which pass
through the spectral filter will be reassembled with proper phase relationships at the
output of the filter. As it travels down the basilar membrane. the temporal waveform
will simply be smoothed, due to loss of its higher-frequency components and delayed, due
to the linear phase shift. Figure 10 shows the response of our electrical analog to a 1
msec wide square pulse at the input. The curves represent the time courses of basilar
membrane displacement at four equally spaced locations along the cochlea. The curve on
the right represents the response at the apical end of the cochlea (the end farthest from the
input). The curve on the left represents the response at a point 25 percent of the distance
667
668
Feld, Eisenberg and Lewis
input end. The impulse responses of mammalian cochleas and of the auditory filters of
lower vertebrates all show a slight ringing, again indicating a deficiency in our model.
w
~
d
>
Apical End
2000
LINEA.R FREOUENCY {HZ'
04
TIfoilE
(SECONDS)
Figure 9 - Phase response of the
Figure 10 - Traveling square wave
basilar membrane velocity.
pulse along the membrane from the
basal to apical end.
CONCLUSION
Research activity studying the function of higher level brain processing is in its
infancy and little is known about how the various features of the cochlea, such as linear
phase, sharp band edges, as well as nonlinear features, such f1S two-tone suppression and
cubic difference tone excitation, are used by the brain. Therefore, our approach, in
developing a cochlear model, is to incorporate only the most essential ingredients. We
have incorporated the five properties mentioned in the introduction which provide
simplicity of analysis, economy of hardware construction, and the preservation of both
temporal and spectral resolution. The inclusion of these properties is also consistent with
the fact that they are found in numerous species.
We have found that in the correct mode of operation a tapered transmission line
model can exhibit these five important cochlear properties. A lumped-element
approximation can be used to simulate this model as long as at least 500 stages are used.
As observed in figure 7, by decreasing the number of stages below 500, the solution to
the lumped-element model no longer adheres to the Zwislocki model. In fact, the output
of the coursely lumped model more closely resembles the neural tuning data of the
cochlea in that it produces very sharp corners. There is some evidence that indicates the
cochlea is constructed of discrete components. Indeed, the hair cells themselves are
discretized. If this idea is valid, a model constructed of as little as 50 branches may more
accurately represent the cochlear mechanics then the Zwislocki model.
Our simple model has some drawbacks in that it does not replicate various properties
of the cochlea. For example, it does not span the full ten octaves of human audition, nor
does it explain any of the experimentally observed nonlinear aspects seen in the cochlea.
However, we take this approach because it provides us with a powerful analysis tool that
will enable us to study the behavior of lumped-element cochlear models. This tool will
allow us to proceed to the next step; the building of a hardware analog of the cochlea.
A Passive Shared Element Analog Electrical Cochlea
allow US to proceed to the next step; the building of a hardware analog of the cochlea.
RESEARCH DIRECTIONS
In and of itself, the tapered shared element travelling wave structure we have chosen is
interesting to analyze. In order to get even further insight into how this filter works and
to aid in the building of a hardware version of such a filter, we plan to study the
placement of the poles and zeroes of the transfer function at each tap along the structure.
In a travelling wave transmission line we expect that the transfer function at each tap will
have the same denominator. Therefore, it must be the numerators of the transfer
functions which will change, i.e. the zeroes will change from tap to tap. It will be of
interest to see what role the zeroes play in such a ladder structure. Furthennore, it will be
of great interest to us to study what happens to the poles and zeroes of the transfer
function at each tap as the number of stag~s is increased (approaching the distributed
parameter ftlter), or decreased (apprpaching the lumped-element cutoff version of the filter
with sharper corners). We should emphasize that our circuit is bi-directional, i.e. there is
loading from the stages before and after each tap, as in the real cochlea. For this reason,
we must consider carefully the options for hardware realization of our circuit. We might
choose to make a mechanical structure on silicon or some oilier medium, or we could
convert our structure into a uni-directional circuit and build it as a digital or analog
circuit
Using this design we plan to build an acoustic imaging device that will enable us to
explore various signal processing tasks. One such task would be to extract acoustic
signals from noise.
All species need to cope with two types of noise, internal sensor and amplifier noise,
and external noise such as that generated by wind. Spectral decomposition is on effective
way to deal with internal noise. For example, the amplitudes of the spectral components
in the passband of a filter are largely undiminished, whereas the broadband noise, passed
by the filter, is proportional to the square root of the bandwidth. External noise reduction
can be accomplished by spatial decomposition. When temporal resolution is preserved in
signals, spatial decomposition can be achieved by cross correlation of the signals from
two ears. Therefore, from these two properties, spectral and temporal resolution, one can
construct an acoustic imaging system in which signals buried in a sea of noise can be
extracted.
Acknowledgments
We would like to thank Thuan Nguyen for figure 1, Eva Poinar who helped with the
figures, Michael Sneary for valuable discussion, and Bruce Parnas for help with
programming.
669
670
Feld, Eisenberg and Lewis
References
Anderson, DJ., Rose, I.E., Hind, I.E .? Brugge I.F.. Temporal Position of Discharges in
Single Auditory Nerve Fibers Within the Cycle of a Sine-Wave Stimulus: Frequency and
Intensity Effects, J. Acoust. Soc. A~z.., 49. 1131-1139, 1971.
Evans, E.F.? Cochlear Filtering: A View Seen Through the Temporal Discharge Patterns
of Single Cochlear Nerve Fibers. A talk given at the 1988 NATO advanced workshop. to
be published as (J.P. Wilson, D.T. Kemp, eds.) Mechanics of Hearing. Plenum Press.
N.Y.
Kiang, N.Y.S .? Moxon, E.C .? Tails of Tuning Curves of Auditory-Nerve Fibers, J.
Acoust. Soc. Am., 55, 620-630. 1974.
Lewis. E.R., High Frequency Rolloff in a Cochlear Model Without critical-layer
resonance. J. Acoust. Soc. Am.? 76 (3) September, 1984.
Lyon, R.F., Mead, C.A.? An Analog Electronic Cochlea, IEEE Trans.-ASSP, 36. 11191134. 1988.
Zweig. G., Lipes. R.. Pierce. I.R.. The Cochlear Compromise, J. Acoust. Soc. Am.,
59, 975-982. 1976.
Zwislocki.I., Analysis of Some Auditory Characteristics, in Handbook of Mathematical
Psychology, Vol. 3. (Wiley. New York), pp. 1-97, 1965.
| 144 |@word version:4 middle:2 rising:1 loading:1 replicate:1 simulation:2 pulse:3 eng:2 decomposition:3 pressure:2 reduction:1 series:2 ours:1 current:2 must:2 realize:3 evans:3 v:2 alone:1 selected:1 device:1 nervous:1 tone:6 beginning:1 indefinitely:1 provides:1 location:3 five:6 mathematical:1 along:8 constructed:2 direct:1 differential:1 resistive:1 wild:1 uncoiled:3 manner:1 indeed:2 behavior:1 themselves:1 nor:1 mechanic:4 roughly:1 brain:8 discretized:1 decreasing:1 lyon:2 little:2 window:3 inappropriate:1 vertebrate:3 matched:1 circuit:7 transduced:1 kiang:5 medium:1 what:5 evolved:1 kind:1 viscous:1 monkey:1 ringing:1 acoust:4 guarantee:1 temporal:8 berkeley:4 farthest:1 before:1 engineering:1 local:3 headset:1 mead:2 approximately:2 might:2 frog:1 shunt:3 resembles:1 suggests:1 co:1 limited:1 range:1 bi:1 decided:1 practical:1 acknowledgment:1 implement:1 displacement:2 get:1 quick:1 center:2 independently:1 resolution:6 simplicity:1 abrupt:1 insight:1 his:1 coordinate:1 analogous:1 discharge:2 plenum:1 construction:2 play:1 programming:1 element:25 velocity:6 mammalian:1 cut:1 observed:4 role:1 electrical:15 enters:2 toff:1 region:1 eva:1 connected:1 cycle:1 f1s:1 valuable:1 mentioned:1 rose:1 dynamic:1 terminating:1 carrying:1 solving:1 segment:1 compromise:1 serve:1 creates:1 localization:1 basis:1 edwin:1 comer:1 various:5 represented:7 fiber:4 cat:2 talk:1 surrounding:1 effective:1 formation:1 extraordinarily:2 whose:1 quite:1 drawing:2 furthennore:1 itself:1 advantage:2 biophysical:2 realization:1 gentle:1 convergence:2 transmission:8 sea:1 produce:3 help:3 illustrate:1 basilar:16 soc:4 indicate:1 come:1 direction:1 waveform:2 closely:1 correct:1 drawback:1 filter:21 human:4 enable:3 squirrel:1 hall:1 presumably:1 great:1 driving:1 major:1 achieves:2 early:1 travel:3 create:1 tool:2 hope:1 clearly:1 sensor:1 idb:1 rather:1 og:3 voltage:2 wilson:1 derived:4 indicates:1 suppression:1 am:3 economy:2 typically:1 going:1 buried:1 development:1 animal:3 plan:2 spatial:2 bioeng:1 resonance:1 construct:1 enlarge:1 represents:3 broad:1 look:1 nearly:1 thin:1 discrepancy:1 stimulus:2 modern:1 simultaneously:2 delayed:2 phase:7 amplifier:1 detection:1 huge:1 interest:2 predominant:1 behind:1 cory:1 edge:9 partial:1 plotted:2 isolated:1 increased:2 compelling:2 moxon:5 retains:1 hearing:3 apical:4 pole:2 delay:1 front:8 off:1 receiving:1 compliant:1 rounded:1 michael:1 together:1 again:1 ear:13 opposed:1 choose:2 possibly:1 corner:5 external:2 audition:2 wkb:1 account:1 afferent:1 sine:1 bone:2 wind:1 root:1 helped:1 analyze:1 view:1 start:1 wave:9 option:1 parallel:4 complicated:1 slope:1 bruce:1 formed:1 square:3 characteristic:2 largely:1 who:1 correspond:1 identify:1 spaced:1 directional:2 accurately:1 produced:1 published:1 processor:1 explain:1 ed:1 against:1 energy:2 frequency:31 intentionally:1 pp:1 obvious:2 associated:1 auditory:8 newly:1 realm:1 amplitude:2 brugge:1 carefully:2 back:1 nerve:4 higher:3 response:13 scala:1 anderson:2 furthermore:3 just:2 stage:10 correlation:1 traveling:1 hand:1 nonlinear:2 mode:8 impulse:2 believe:1 building:3 effect:2 true:1 evolution:2 deal:1 lumped:10 numerator:1 elect:2 covering:1 excitation:1 octave:5 passive:6 percent:1 image:11 recently:1 common:1 mentally:1 functional:1 khz:1 analog:19 slight:1 tail:1 silicon:1 ai:1 tuning:5 inclusion:1 had:2 dj:1 longer:1 base:1 add:1 showed:1 massively:3 psychoacoustical:1 accomplished:1 inverted:2 seen:2 minimum:1 strike:1 signal:16 preservation:1 branch:2 full:1 sound:4 exceeds:1 cross:1 zweig:3 long:3 equally:1 hair:1 denominator:1 cochlea:41 represent:6 achieved:1 cell:1 irregular:1 preserved:1 whereas:1 want:1 decreased:1 source:2 leaving:1 operate:1 finely:1 recording:1 isolate:1 hz:4 db:1 compliance:2 incorporates:1 lump:1 extracting:1 structural:2 curious:1 enough:3 psychology:1 approaching:1 bandwidth:1 economic:1 drum:1 idea:1 shift:3 whether:1 passed:1 swim:1 stereo:1 resistance:1 micromechanics:1 cause:2 proceed:2 repeatedly:1 york:1 useful:2 clear:1 band:10 ten:1 hardware:6 parnas:1 multichannel:1 simplest:1 spice:3 designer:1 discrete:1 vol:1 group:1 basal:4 four:3 nevertheless:1 threshold:2 falling:1 tapered:4 cutoff:5 imaging:2 convert:1 package:1 powerful:1 electronic:1 excursion:1 entirely:1 layer:1 simplification:1 activity:1 placement:1 deficiency:2 software:1 fourier:1 simulate:1 aspect:1 extremely:2 spring:1 span:3 relatively:1 linea:1 hind:1 transferred:1 developing:1 membrane:18 increasingly:1 stapes:1 perceptible:1 joseph:1 shallow:1 happens:2 gradually:1 taken:1 equation:1 turn:1 needed:1 know:1 end:19 travelling:2 studying:1 available:1 operation:6 observe:1 appropriate:1 spectral:6 voice:1 original:3 capacitive:1 denotes:1 include:1 alligator:1 build:2 passband:2 move:1 realized:1 occurs:2 reassembled:1 responds:1 evolutionary:1 exhibit:2 september:1 distance:1 thank:1 stag:1 outer:1 acoustical:8 cochlear:15 argue:1 seven:3 kemp:1 toward:1 reason:1 relationship:2 difficult:1 steep:5 sharper:2 slows:1 fluid:5 design:2 proper:1 perform:1 observation:1 fin:2 finite:3 incorporated:1 head:3 delaying:1 assp:1 rn:1 smoothed:1 sharp:4 intensity:1 david:1 namely:2 mechanical:3 tap:6 acoustic:11 trans:1 able:2 below:1 pattern:1 vibrate:1 including:1 critical:2 feld:5 force:1 advanced:1 representing:1 ladder:1 numerous:1 extract:1 review:1 understanding:1 eisenberg:5 relative:1 loss:1 expect:1 interesting:1 proportional:1 filtering:1 ingredient:3 digital:1 frequenc:1 consistent:1 playing:2 share:2 cd:1 lo:1 course:1 placed:1 last:1 side:1 allow:2 wide:4 fifth:1 distributed:5 curve:20 dimension:1 valid:2 transition:1 inertia:1 made:3 clue:1 commonly:1 simplified:2 nguyen:1 far:2 cope:1 emphasize:1 uni:1 nato:1 logic:1 handbook:1 spectrum:1 search:1 channel:4 transfer:4 ca:1 adheres:1 complex:2 submit:1 did:1 rh:1 noise:8 enlarged:1 broadband:1 depicts:2 cubic:1 aid:1 wiley:1 structurally:1 position:1 msec:1 resistor:2 infancy:1 down:2 removing:1 physiological:1 evidence:3 survival:1 essential:3 workshop:1 joe:1 importance:1 ci:2 magnitude:1 pierce:1 push:2 occurring:1 hole:1 artifactual:2 wavelength:1 simply:1 explore:1 absorbed:1 desire:1 adjustment:1 lewis:7 extracted:1 oct:1 presentation:1 narrower:1 shared:8 man:1 change:3 experimentally:1 determined:1 typical:1 oval:3 pas:1 specie:2 experimental:1 indicating:1 internal:2 people:1 incorporate:1 dept:2 phenomenon:1 |
483 | 1,440 | Two Approaches to Optimal Annealing
Todd K. Leen
Dept of Compo Sci. & Engineering
Oregon Graduate Institute of
Science and Technology
P.O.Box 91000, Portland,
Oregon 97291-1000
tleen@cse.ogi.edu
Bernhard Schottky and David Saad
Neural Computing Research Group
Dept of Compo Sci. & Appl. Math.
Aston University
Birmingham, B4 7ET, UK
schottba{ saadd}@aston.ac.uk
Abstract
We employ both master equation and order parameter approaches
to analyze the asymptotic dynamics of on-line learning with different learning rate annealing schedules. We examine the relations
between the results obtained by the two approaches and obtain new
results on the optimal decay coefficients and their dependence on
the number of hidden nodes in a two layer architecture.
1
Introduction
The asymptotic dynamics of stochastic on-line learning and it's dependence on the
annealing schedule adopted for the learning coefficients have been studied for some
time in the stochastic approximation literature [1, 2] and more recently in the neural
network literature [3, 4, 5]. The latter studies are based on examining the KramersMoyal expansion of the master equation for the weight space probability densities.
A different approach, based on the deterministic dynamics of macroscopic quantities
called order parameters, has been recently presented [6, 7]. This approach enables
one to monitor the evolution of the order parameters and the system performance
at all times.
In this paper we examine the relation between the two approaches and contrast the
results obtained for different learning rate annealing schedules in the asymptotic
regime. We employ the order parameter approach to examine the dependence of
the dynamics on the number of hidden nodes in a multilayer system. In addition,
we report some lesser-known results on non-standard annealing schedules
T. K. Leen, B. Schottky and D. Saad
302
2
Master Equation
Most on-line learning algorithms assume the form Wt+l = Wt + 1]o/tP H(wt,xt)
where Wt is the weight at time t, Xt is the training example, and H(w,x) is the
weight update. The description of the algorithm's dynamics in terms of weight
space probability densities starts from the master equation
P(w',t+1)=
JdW
(8(w'-w-~~H(w,x)))xP(w,t)
(1)
where ( .. '}x indicates averaging with respect to the measure on x, P(w,t) is the
probability density on weights at time t, and 8( ... ) is the Dirac function. One may
use the Kramers-Moyal expansion of Eq.(l) to derive a partial differential equation
for the weight probability density (here in one dimension for simplicity) {3, 4]
at P (w, t)
=
t
r
(~~) i (7~ a~ [ (Hi (w, x) >x P (w, t)]
t=l
.
(2)
Following {3], we make a small noise expansion for (2) by decomposing the weight
trajectory into a deterministic and stochastic pieces
or
~
= ( TJo)-"(
tP
(w-</>(t))
(3)
where </>( t) is the deterministic trajectory, and ~ are the fluctuations. Apart from
the factor (1]0/ tP)'Y that scales the fluctuations, this is identical to the formulation
for constant learning in {3]. The proper value for the unspecified exponent, will
emerge from homogeneity requirements. Next, the dependence of the jump moments
(Hi (w, x) > on 1]0 is explicated by a Taylor series expansion about the deterministic
path </>. T!{e coefficients in this series expansion are denoted
a~i) == ai (Hi(w,x))x /awilw=tI>
Finally one rewrites (2) in terms of </> and ~ and the expansion of the jump moments,
taking care to transform the differential operators in accordance with (3).
These transformations leave equations of motion for </> and the density
the fluctuations
d</>
dt
=
I1(~,
t) on
(4)
=
=
For stochastic descent H (w, x)
-V w E( w, x) and (4) describes the evolution of
</> as descent on the average cost. The fluctuation equation (5) requires further
manipulation whose form depends on the context. For the usual case of descent
in a quadratic minimum (ail) = -G, minus the cost function curvature), we take
''i
1/2 to insure that for any m, terms in the sum are homogeneous in T}o/t P
=
For constant learning rate (p = 0), rescaling time as t ~ 1]ot allows (5) to be written
in a form convenient for perturbative analysis in 1]0 Typically, the limit 1]0 ~ 0 is
invoked and only the lowest order terms in 1]0 retained (e.g. [3]). These comprise
a diffusion operator, which results in a Gaussian approximation for equilibrium
densities. Higher order terms have been successfully used to calculate corrections
to the equilibrium moments in powers of 1]0 [8].
Two Approaches to Optimal Annealing
303
Of primary interest here is the case of annealed learning, as required for convergence
of the parameter estimates. Again assuming a quadratic bowl and 'Y = 1/2, the first
few terms of (5) are
Gt II = As t -+
00
o < p S 1).
:t Gd~II)
- all)
i; Ge(~II) + ~a~O):
Gl II + 0 (:
r/2.
(6)
the right hand side of (6) is dominated by the first three terms (since
Precisely which terms dominate depends on p.
.
We will first review the classical case p = 1. Asymptotically 1> -+ w"', a local
optimum. The first three leading terms on the right hand side of (6) are all of order
lit. For t -+ 00, we discard the remaining terms. From the resulting equation we
recover a Gaussian equilibrium distribution for ~, or equivalently for Vt (w - w"') ==
Vtv where v is called the weight error. The asymptotically normal distribution for
Vtv has variance 0'0v from which the asymptotic expected squared weight error
can be derived
2 (0)
1
?
E[I v 12] - 0 ' 2 - - 7Jo a 2
1
11m
(7)
t->oo
-,;tv t 27]0 G'" - 1 t
== G(w"') is the curvature at the local optimum.
Positive O',;t v requires T]o > 11 (2G"'). If this condition is not met the expected
where G'"
squared weight offset converges as (1It)1-2'1oG', slower than lit [5, for example,
and references therein]. The above confirms the classical results [1] on asymptotic
normality and convergence rate for 1I t annealing.
For the case 0 < p < 1, the second and third terms on the right hand side of (6)
will dominate as t -+ 00. Again, we have a Gaussian equilibrium density for ~.
Consequently ViP v is asymptotically normal with variance O'~v leading to the
expected squared weight error
0'
2
r;-;;
ytPv
-
1 =
tP
T]o
a~O)
2G
.!..
tP
(8)
Notice that the convergence is slower than lit and that there is no critical value of
the learning rate to obtain a sensible equilibrium distribution. (See [9] for earlier
results on 11tP annealing.)
The generalization error follows the same decay rate as the expected weight offset.
In one dimension, the expected squared weight offset is directly related to excess
generalization error (the generalization error minus the least generalization error
achievable) Eg = G E[v 2 ]. In multiple dimensions, the expected squared weight
offset, together with the maximum and minimum eigenvalues of G'" provide upper
and lower bounds on the excess generalization error proportional to E[lvI 2 ], with
the criticality condition on G'" (for p = 1 )replaced with an analogous condition on
its eigenvalues.
3
Order parameters
In the Master equation approach, one focuses attention on the weight space distribution P( w, t) and calculates quantities of interested by averaging over this density.
An alternative approach is to choose a smaller set of macroscopic variables that are
sufficient for describing principal properties of the system such as the generalization error (in contrast to the evolution of the weights w which are microscopic).
T. K. Leen, B. Schottky and D. Saad
304
Formally, one can replace the parameter dynamics presented in Eq.(1) by the corresponding equation for macroscopic observables which can be easily derived from
the corresponding expressions for w. By choosing an appropriate set of macroscopic
variables and invoking the thermodynamic limit (i.e., looking at systems where the
number of parameters is infinite), one obtains point distributions for the order parameters, rendering the dynamics deterministic.
Several researchers [6, 7] have employed this approach for calculating the tr~ing
dynamics of a soft committee machine (SCM) . The SCM maps inputs x E RN to
a scalar, through a model p{w,x)
2:~lg{Wi' x). The activation function of
the hidden units is g{u) == erf{u/V2) and Wi is the set of input-to-hidden adaptive
weights for the i = 1 ... K hidden nodes. The hidden-to-output weights are set
to 1. This architecture preserves most of the properties of the learning dynamics
and the evolution of the generalization error as a general two-layer network, and
the formalism can be easily extended to accommodate adaptive hidden-to-output
weights [10].
=
Input vectors x are independently drawn with zero mean and unit variance, and the
corresponding targets y are generated by deterministic teacher network corrupted
by additive Gaussian output noise of zero mean and variance O'~. The teacher
network is also a SCM, with input-to-hidden weights wi. The order parameters
sufficient to close the dynamics, and to describe the network generalization error
are overlaps between various input-to-hidden vectors Wi . Wk == Qik, Wi' W~ _
Rin, and w~? w~ == Tnm .
Network performance is measured in terms of the generalization error Eg{W) _
(1/2 [ p(w, x) - Y ]2)~. The generalization error can be expressed in closed form in
terms of the order parameters in the thermodynamiclimit (N -+ 00). The dynamics
of the latter are also obtained in closed form [7]. These dynamics are coupled nonlinear ordinary differential equations whose solution can only be obtained through
numerical integration. However, the asymptotic behavior in the case of annealed
learning is amenable to analysis, and this is one of the primary results of the paper.
We assume an isotropic teacher Tnm = 8 nm and use this symmetry to reduce the
system to a vector of four order parameters uT = (r, q, s, c) related to the overlaps
by Rin = 8in (1 + r) + (1- 8in)S and Qik = 8 i k(1 + q) + {1- 8 ik )C.
?
With learning rate annealing and limt-+oo u = we describe the dynamics in this
vicinity by a linearization of the equations of motion in [7]. The linearization is
d
dt U
where O'~ is the noise variance, b T
-4
M
=
2
3V31T
4
3
--V3
2
3V3
3
= rJ 1\d U + rJ 2 0' 2b ,
(9)
/I
= ~ (0,1/,;3,0,1/2), rJ = rJo/tP,
-~(K -lhI(3)
4 3
-(K - 1)V3
3 2
2
2
3
-(K -1)V3
~
--(K - 1)V3
2
o
--(K-2)+3V3(K - 2)
J3
+~
and M is
-3V3(K - 2)
(10)
+~
V3
The asymptotic equations of motion (9) were derived by dropping terms of order
O(rJlluI12) and higher, and terms of order O{rJ2 u). While the latter are linear in the
order parameters, they are dominated by the rJu and rJ20'~b terms in (9) as t -+ 00.
Two Approaches to Optimal Annealing
305
This choice of truncations sheds light on the approach to equilibrium that is not
implicit in the master equation approach. In the latter, the dominant terms for
the asymptotics of (6) were identified by time scale of the coefficients, there was
no identification of system observables that signal when the asymptotic regime
is entered. For the order parameter approach, the conditions for validity of the
asymptotic approximations are cast in terms of system observables 1JU vs rlu VS
1J2 fI~.
The solution to (9) is
u(t)
= -yet, to) Uo + fI~
j3(t, to) b
(11)
where Uo == u(to) and
-y(t, to) = exp
{M lot dr 1J(r)}
and j3(t, to)
=
t dr -yet, r) 1J2( r).
lto
(12)
The asymptotic order parameter dynamics allow us to compute the generalization
error (to first order in u)
E/
K(l
K-1
= -:;
J3(q- 2r) + -2-(C28) ) .
(13)
Using the solution of Eq.(l1), the generalization error consists of two pieces: a
contribution depending on the actual initial conditions Uo and a contribution due
to the second term on the r.h.s. of Eq.(l1), independent of Uo. The former decays
more rapidly than the latter, and we ignore it in what follows. Asymptotically,
the generalization error is of the form E/ = fI;(CI0l(t) + C202(t)), where Cj are K
dependent coefficients, and OJ are eigenmodes that evolve as
OJ
=-
1J5
1 + (}:j1Jo
[!t _
to'it) Ot;;-(O'it)O+I)] .
(14)
= -~ (~ +2(K -1))
(15)
with eigenvalues (Fig. l(a))
(}:1
= -~ (~
-2)
and
(}:2
The critical learning rate 1J~rit, above which the generalization decays as lit is, for
K>
- 2,
crit
7r
(16)
110 = max - (}:1 , - (}:2
41 J3 - 2 .
(1 1) =
For 1Jo
> 1J~rit
E/
both modes OJ, i = 1,2 decay as lit, and so
C2) 1
= -fI 2 11o2 (Cl
1 + (}:11JO + 1 + (}:21Jo t
1I
2
(
== fIll f 1Jo,K)
t1
(17)
Minimizing the prefactor f (1Jo, K) in (17) minimizes the asymptotic error. The
values 1J~Pt (K) are shown in Fig. 1(b), where the special case of K = 1 (see below)
is also included: There is a significant difference between the values for K = 1 and
K = 2 and a rather weak dependence on K for K ~ 2. The sensitivity of the
generalization error decay factor on the choice of 1Jo is shown in Fig. 1 (c).
The influence of the noise strength on the generalization error can be seen directly
from (17): the noise variance fI; is just a prefactor scaling the lit decay. Neither
the value for the critical nor for the optimal1Jo is influenced by it.
T. K. Leen, B. Schottky and D. Saad
306
The calculation above holds for the case K = 1 (where c and s and the mode 01 are
absent). In this case
- 2 crit(R" -- 1) -__ 2.
110opt(K -_ 1) -1]0
<l:2
_ J311" .
-
2
(18)
Finally, for the general annealing schedule of the form 1] = 1]oltP with 0 < p < 1
the equations of motion (11) can be investigated, and one again finds 11tP decay.
4
Discussion and summary
We employed master equation and order parameter approaches to study the convergence of on-line learning under different annealing schedules. For the lit annealing
schedule, the small noise expansion provides a critical value of 1]0 (7) in terms of
the curvature, above which Vt v is asymptotically normal, and the generalization
decays as lit. The approach is general, but requires knowledge of the first two jump
moments in the asymptotic regime for calculating tl~e relevant properties.
By restricting the order parameters approach to a symmetric task characterized
by a set of isotropic teacher vectors, one can explicitly solve the dynamics in the
asymptotic regime for any number of hidden nodes, and provide explicit expressions
for the decaying generalization error and for the critical (16) and optimal learning
rate prefactors for any number of hidden nodes K. Moreover, one can study the
sensitivity of the generalization error decay to the choice of this prefactor. Similar results have been obtained for the critical learning rate prefactors using both
methods, and both methods have been used to study general11tP annealing. However, the order parameters approach enables one to gain a complete description of
the dynamics and additional insight by restricting the task examined. Finally the
order parameters approach expresses the dynamics in terms of ordinary differential
equations, rather than partial differential equations; a clear advantage for numerical
investigations.
The order parameter approach provides a potentially helpful insight on the passage
into the asymptotic regime. Unlike the truncation of the small noise expansion, the
truncation of the order parameter equations to obtain the asymptotic dynamics is
couched in terms of system observables (c.f. the discussion following (10)). That
is, one knows exactly which observables must be dominant for the system to be
in the asymptotic regime. Equivalently, starting from the full equations, the order parameters approach can tell us when the system is close to the equilibrium
distribution.
Although we obtained a full description of the asymptotic dynamics, it is still unclear
how relevant it is in the larger picture which includes all stages of the training
process, as in many cases it takes a prohibitively long time for the system to reach
the asymptotic regime. It would be interesting to find a way of extending this
framework to gain insight into earlier stages of the learning process.
Acknowledgements: DS and BS would like to thank the Leverhulme Trust for their
support (F J250JK). TL thanks the International Human Frontier Science Program (SF
473-96), and the NSF (ECS-9704094) for their support.
Two Approaches to Optimal Annealing
307
I
wr----~~~-~~~~
" ~.
-2
.....
"'''.
-4
::v .... . . ...
-' II;.
"." ....
", ....
-8
=~~' I
8
. . . . ..
4
6
8
10
12
14
16
18
4
I
(b)
20
18
16
-K=I
...... K=2
- - K=3
.. .. K=5
,.
14
,-,
::.::
0
.,f .
12
I\ . ..
10
~\
~
;;::: 8
.... .. .. ... . . . . .
\\.
\
6
\..~'-----------
4
....... .............
2
0
2
4
6
(c)
8
.....
....?.
OL-~~--~--'-~--~~~~
20
K
(a)
? .. ....? ... .
2
L -_ _....i.---'--....L.~"'--'-.....&--..L.---'''-'?? ........J
2
.. ?
6
.........
'.
10
12
14
16
. ?......
.
14
-'." ".
-6
-12
.. .... f(l1O"'. K)
16
"
-10
-T/"'"
0
18
" .
2
3
4
5
6
7
8
9
10
K
Figure 1: (a) The dependence of the
eigenvalues of M on the the number of hidden units K. Note that
the constant eigenvalue al dominates the convergence for K ~ 2_
(b) 1]gpt and the resulting generalization prefactor / (1]gpt ,K) of (17)
as a function of K_ (c) The dependence of the generalization error decay prefactor / (1]0, K) on the choice
of 1]0.
18
t'lo
References
[1] V. Fabian. Ann. Math . Statist., 39, 1327 1968.
[2] L. Goldstein. Technical Report DRB-306, Dept. of Mathematics, University of
Southern California, LA, 1987.
[3] T. M. Heskes and B. Kappen, Phys. Rev. A 44, 2718 (1991)_
[4] T. K. Leen and J. E. Moody. In Giles, Hanson, and Cowan, editors, Advances in
Neural In/ormation Processing Systems, 5, 451 , San Mateo, CA, 1993. Morgan
Kaufmann.
[5] T. K. Leen and G. B. Orr. In J.D. Cowan, G. Tesauro, and J. Alspector, editors,
Advances in Neural Information Processing Systems 6, 477 ,San Francisco, CA.,
1994. Morgan Kaufmann Publishers.
[6] M. Biehl and H. Schwarze, J. Phys. A 28, 643 (1995).
[7] D. Saad and S_A. Solla Phys. Rev. Lett. 74, 4337 (1995) and Phys. Rev. E 52
4225 (1995).
[8] G. B. Orr. Dynamics and Algorithms for Stochastic Search. PhD thesis, Oregon
Graduate Institute, October 1996.
[9] Naama Barkai. Statistical Mechanics of Learning_ PhD thesis, Hebrew University of Jerusalem, August 1995.
[10] P. Riegler and M. Biehl J_ Phys. A 28, L507 (1995).
| 1440 |@word achievable:1 confirms:1 invoking:1 minus:2 tr:1 accommodate:1 kappen:1 moment:4 initial:1 series:2 o2:1 activation:1 yet:2 perturbative:1 written:1 must:1 additive:1 numerical:2 enables:2 update:1 rjo:1 v:2 isotropic:2 compo:2 provides:2 math:2 node:5 cse:1 c2:1 differential:5 ik:1 consists:1 expected:6 alspector:1 behavior:1 examine:3 nor:1 mechanic:1 ol:1 actual:1 insure:1 moreover:1 lowest:1 what:1 unspecified:1 ail:1 minimizes:1 transformation:1 ti:1 shed:1 exactly:1 prohibitively:1 uk:2 unit:3 uo:4 positive:1 t1:1 engineering:1 accordance:1 todd:1 local:2 limit:2 riegler:1 fluctuation:4 path:1 therein:1 studied:1 examined:1 mateo:1 appl:1 graduate:2 asymptotics:1 convenient:1 close:2 operator:2 context:1 influence:1 deterministic:6 map:1 annealed:2 jerusalem:1 attention:1 starting:1 independently:1 simplicity:1 insight:3 dominate:2 fill:1 analogous:1 target:1 pt:1 homogeneous:1 prefactor:5 calculate:1 ormation:1 solla:1 dynamic:20 rewrite:1 crit:2 rin:2 lto:1 observables:5 bowl:1 easily:2 various:1 describe:2 tell:1 choosing:1 whose:2 larger:1 solve:1 biehl:2 erf:1 transform:1 advantage:1 eigenvalue:5 j2:2 relevant:2 entered:1 rapidly:1 description:3 dirac:1 convergence:5 requirement:1 optimum:2 extending:1 leave:1 converges:1 derive:1 oo:2 ac:1 depending:1 measured:1 eq:4 met:1 stochastic:5 human:1 generalization:21 investigation:1 opt:1 frontier:1 correction:1 hold:1 normal:3 exp:1 equilibrium:7 birmingham:1 gpt:2 successfully:1 gaussian:4 rather:2 og:1 derived:3 focus:1 portland:1 indicates:1 contrast:2 helpful:1 dependent:1 typically:1 hidden:12 relation:2 i1:1 interested:1 j5:1 denoted:1 exponent:1 integration:1 special:1 comprise:1 identical:1 lit:8 report:2 employ:2 few:1 preserve:1 rj2:1 homogeneity:1 replaced:1 interest:1 tjo:1 light:1 amenable:1 partial:2 taylor:1 formalism:1 earlier:2 soft:1 giles:1 tp:8 ordinary:2 cost:2 examining:1 teacher:4 corrupted:1 gd:1 ju:1 density:8 thanks:1 sensitivity:2 international:1 together:1 moody:1 jo:7 again:3 squared:5 nm:1 thesis:2 choose:1 dr:2 leading:2 rescaling:1 lhi:1 orr:2 wk:1 includes:1 coefficient:5 oregon:3 explicitly:1 depends:2 piece:2 lot:1 closed:2 analyze:1 start:1 recover:1 decaying:1 scm:3 contribution:2 variance:6 kaufmann:2 weak:1 identification:1 trajectory:2 researcher:1 influenced:1 reach:1 phys:5 gain:2 knowledge:1 ut:1 cj:1 schedule:7 goldstein:1 higher:2 dt:2 leen:6 formulation:1 box:1 just:1 implicit:1 stage:2 d:1 hand:3 trust:1 nonlinear:1 mode:2 eigenmodes:1 schwarze:1 barkai:1 validity:1 l1o:1 evolution:4 vicinity:1 former:1 symmetric:1 eg:2 couched:1 ogi:1 complete:1 motion:4 l1:2 passage:1 invoked:1 recently:2 fi:5 b4:1 moyal:1 significant:1 ai:1 drb:1 mathematics:1 heskes:1 gt:1 prefactors:2 dominant:2 curvature:3 apart:1 discard:1 manipulation:1 tesauro:1 vt:2 seen:1 minimum:2 additional:1 care:1 morgan:2 employed:2 v3:8 signal:1 ii:5 thermodynamic:1 multiple:1 full:2 rj:3 ing:1 technical:1 characterized:1 calculation:1 long:1 calculates:1 j3:5 multilayer:1 limt:1 qik:2 addition:1 annealing:15 macroscopic:4 publisher:1 saad:5 ot:2 unlike:1 cowan:2 rendering:1 architecture:2 identified:1 reduce:1 lesser:1 absent:1 vtv:2 expression:2 j_:1 clear:1 statist:1 nsf:1 notice:1 leverhulme:1 wr:1 dropping:1 express:1 group:1 four:1 monitor:1 drawn:1 neither:1 schottky:4 diffusion:1 asymptotically:5 sum:1 saadd:1 master:7 scaling:1 tleen:1 layer:2 hi:3 bound:1 quadratic:2 strength:1 precisely:1 dominated:2 tv:1 describes:1 smaller:1 wi:5 rev:3 b:1 equation:20 describing:1 committee:1 know:1 ge:1 vip:1 adopted:1 decomposing:1 v2:1 appropriate:1 alternative:1 slower:2 remaining:1 calculating:2 classical:2 quantity:2 primary:2 dependence:7 usual:1 unclear:1 microscopic:1 southern:1 thank:1 sci:2 sensible:1 assuming:1 retained:1 minimizing:1 hebrew:1 equivalently:2 lg:1 october:1 potentially:1 proper:1 upper:1 fabian:1 descent:3 criticality:1 extended:1 looking:1 rn:1 august:1 david:1 cast:1 required:1 hanson:1 california:1 below:1 regime:7 program:1 oj:3 max:1 power:1 critical:6 overlap:2 l507:1 normality:1 technology:1 aston:2 picture:1 coupled:1 review:1 literature:2 acknowledgement:1 evolve:1 asymptotic:18 interesting:1 proportional:1 sufficient:2 xp:1 kramers:1 editor:2 lo:1 summary:1 gl:1 truncation:3 side:3 allow:1 institute:2 taking:1 emerge:1 naama:1 jdw:1 dimension:3 lett:1 jump:3 adaptive:2 san:2 ec:1 excess:2 obtains:1 ignore:1 bernhard:1 francisco:1 search:1 k_:1 ca:2 symmetry:1 expansion:8 investigated:1 cl:1 lvi:1 noise:7 fig:3 tl:2 explicit:1 sf:1 third:1 xt:2 tnm:2 offset:4 decay:11 dominates:1 restricting:2 phd:2 linearization:2 expressed:1 scalar:1 consequently:1 ann:1 replace:1 included:1 infinite:1 wt:4 averaging:2 principal:1 called:2 la:1 formally:1 rit:2 support:2 latter:5 dept:3 |
484 | 1,441 | On the infeasibility of training neural
networks with small squared errors
Van H. Vu
Department of Mathematics, Yale University
vuha@math.yale.edu
Abstract
We demonstrate that the problem of training neural networks with
small (average) squared error is computationally intractable. Consider a data set of M points (Xi, Yi), i = 1,2, ... , M, where Xi are
input vectors from R d , Yi are real outputs (Yi E R). For a network 10 in some class F of neural networks, (11M) L~l (fO(Xi)Yi)2)1/2 - inlfEF(l/ M) "2:f!1 (f(Xi) - YJ2)1/2 is the (avarage) relative error occurs when one tries to fit the data set by 10. We will
prove for several classes F of neural networks that achieving a relative error smaller than some fixed positive threshold (independent
from the size of the data set) is NP-hard.
1
Introduction
Given a data set (Xi, Yi), i = 1,2, ... , M. Xi are input vectors from R d , Yi are real
outputs (Yi E R). We call the points (Xi, Yi) data points. The training problem
for neural networks is to find a network from some class (usually with fixed number
of nodes and layers), which fits the data set with small error. In the following we
describe the problem with more details.
Let F be a class (set) of neural networks, and a be a metric norm in RM. To
each 1 E F, associate an error vector Ef = (1/(Xd - Yil)f;l (EF depends on the
data set, of course, though we prefer this notation to avoid difficulty of having too
many subindices). The norm of Ej in a shows how well the network 1 fits the data
regarding to this particular norm. Furthermore, let eo:,F denote the smallest error
achieved by a network in F, namely:
eo: F
,
= min
liEf 110:
fEF
In this context, the training problem we consider here is to find
1
E F such that
v.n
372
Vu
IIEfila - ea ,F ~ fF, where fF is a positive number given in advance, and does not
depend on the size M of the data set. We will call fF relative error. The norm a
is chosen by the nature of the training process, the most common norms are:
100 norm:
12 norm:
lem).
Ilvll oo = maxlvi/ (interpolation problem)
IIvl12 = (l/M2::;l v[)1/2, where v = (Vi)t;l
(least square error prob-
The quantity liEf 1112 is usually referred to as the emperical error of the training
process. The first goal of this paper is to show that achieving small emperical error
is NP-hard. From now on, we work with 12 norm, if not otherwise specified.
A question of great importance is: given the data set, F and fF in advance, could
one find an efficient algorithm to solve the training problem formulated above. By
efficiency we mean an algorithm terminating in polynomial time (polynomial in the
size of the input). This question is closely related to the problem of learning neural
networks in polynomial time (see [3]). The input in the algorithm is the data set,
by its size we means the number of bits required to write down all (Xi, Yi).
Question 1. Given F and fF and a data set. Could one find an efficient algorithm
which produces a function f E F such that liEf II < eF + fF
Question 1 is very difficult to answer in general. In this paper we will investigate
the following important sub-question:
Question 2. Can one achieve arbitrary small relative error using polynomial algorithms ?
Our purpose is to give a negative answer for Question 2. This question was posed
by 1. Jones in his seminar at Yale (1996). The crucial point here is that we are
dealing with 12 norm, which is very important from statistical point of view. Our
investigation is also inspired by former works done in [2], [6], [7], etc, which show
negative results in the 100 norm case.
Definition. A positive number f is a threshold of a class F of neural networks if
the training problem by networks from F with relative error less than f is NP-hard
(i.e., computationally infeasible) .
In order to provide a negative answer to Question 2, we are going to show the
existence of thresholds (which is independent from the size of the data set) for the
following classes of networks.
= {flf(x) = (l/n)(2:~=l step (ai x - bi)}
? F~ = {flf(x) = (2:7=1 Cistep (ai x - bd}
? On = {glg(x) = 2:~1 ci<!>i(ai x - bi)}
where n is a positive integer, step(x) = 1 if x is positive and zero otherwise, ai and
? Fn
are vectors from R d , bi are real numbers, and Ci are positive numbel's. It is clear
that the class F~ contains Fn; the reason why we distinguish these two cases is that
the proof for Fn is relatively easy to present, while contains the most important
ideas. In the third class, the functions 1>i are sigmoid functions which satisfy certain
Lipchitzian conditions (for more details see [9])
x
Main Theorem
(i) The classes F1, F2,
F~
and
02 have absolute constant (positive) thresholds
On the Infeasibility of Training Neural Networks with Small Squared Errors
373
(ii) For ellery class F n+2, n > 0, there is a threshold of form (n- 3/'2d- 1 /'2.
(iii) For every
F~+'2' 11
> 0,
(iv) For every class 9n+2, n
there is a threshold of form (n-3/2d-3/'2 .
> 0,
there is a threshold of form (n- 5 / 2 d- 1 / 2 .
In the last three statements. ( is an absolute positive constant .
Here is the key argument of the proof. Assume that there is an algorithm A which
solves the training problem in some class (say Fn ) with relative error f. From
some (properly chosen) NP-hard problem. we will construct a data set so that if f
is sufficiently small, then the solution found by A (given the constructed data set
as input) in Fn implies a solution for the original NP-hard problem. This will give
a lower bound on f, if we assume that the algorithm A is polynomial. In all proofs
the leading parameter is d (the dimension of data inputs). So by polynomial we
mean a polynomial with d as variable. All the input (data) sets constructed
have polynomial size in d.
""ill
The paper is organized as follow. In the next Section, we discuss earlier results
concerning the 100 norm. In Section 3, we display the NP-hard results we will use in
the reduction. In Section 4, we prove the main Theorem for class F2 and mention
the method to handle more general cases. We conclude with some remarks and
open questions in Section 5.
To end this Section, let us mention one important corollary. The Main Theorem
implies that learning F n, F~ and 9n (with respect to 12 norm) is hard. For more
about the connection between the complexity of training and learning problems, we
refer to [3], [5].
Notation: Through the paper Ud denotes the unit hypercube in Rd. For any
number x, Xd denotes the vector (x, X,." x) of length d. In particular, Od denotes
the origin of Rd. For any half space H, fI is the complement of H. For any set A, IAI
is the number of elements in A. A function y( d) is said to have order of magnitude
0(F(d)), if there are c < C positive constants such that c < y(d)jF(d) < C for all
d.
2
Previous works in the loo case
The case Q = 100 (interpolation problem) was considered by several authors for
many different classes of (usually) 2-layer networks (see [6],[2], [7], [8]). Most of the
authors investigate the case when there is a perfect fit, i.e., eleo,F = O. In [2], the
authors proved that training 2-layer networks containing 3 step function nodes with
zero relative error is NP-hard. Their proof can be extended for networks with more
inner nodes and various logistic output nodes. This generalized a former result of
Maggido [8] on data set with rational inputs. Combining the techniques used in
[2] with analysis arguments, Lee Jones [6] showed that the training problem with
relative error 1/10 by networks with two monotone Lipschitzian Sigmoid inner nodes
and linear output node, is also NP-hard (NP-complete under certain circumstances).
This implies a threshold (in the sense of our definition) (1/10)M- 1/ 2 for the class
examined. However, this threshold is rather weak, since it is decreasing in M. This
result was also extended for the n inner nodes case [6].
It is also interesting to compare our results with Judd's. In [7] he considered the
following problem "Given a network and a set of training examples (a data set),
does there exist a set of weights so that the network gives correct output for all
training examples ?" He proved that this problem is NP-hard even if the network is
V. H. Vu
374
required to produce the correct output for two-third of the traing examples. In fact,
it was shown that there is a class of networks and a data sets so that any algorithm
will produce poorly on some networks and data sets in the class. However, from
this result one could not tell if there is a network which is "hard to train" for all
algorithms. Moreover, the number of nodes in the networks grows with the size of
the data set. Therefore, in some sense, the result is not independent from the size
of the data set.
In our proofs we will exploit many techniques provided in these former works. The
crucial one is the reduction used by A. Blum and R. Rivest, which involves the
NP-hardness of the Hypergraph 2-Coloring problem.
3
Sonle NP hard problems
Definition Let B be a CNF formula, where each clause has at most k literals.
Let max(B) be the maximum number of clauses which can be satisfied by a truth
assignment. The APP MAX k-SAT problem is to find a truth assignment which
satisfies (1 - f)max(B) clauses.
The following Theorem says that this approximation problem is NP -hard, for some
small f.
Theorem 3.1.1 Fix k 2: 2. There is fl > 0, such that finding a truth assignment.
which satisfies at least (1- fdmax(B) clauses is NP-h ard.
The problem is still hard, when every literal in B appears in only few clauses, and
every clause contains only few literals. Let B3(5) denote the class of CNFs with at
most 3 literals in a clause and every literal appears in at most 5 clauses (see [1]).
Theorem 3.1.2 There is t2 > 0 such that finding a truth assignment, which satisfies
at least (1 - f)max(B) clauses in a formula B E B3(5) is NP-hard.
The optimal thresholds in these theorems can be computed, due to recent results
in Thereotical Computer Science. Because of space limitation, we do not go into
this matter.
Let H = (V, E) be a hypergraph on the set V, and E is the set of edges (collection
of subsets of V). Elements of V are called vertices. The degree of a vertex is the
number of edges containing the vertex. We could assume that each edge contains
at least two vertices. Color the vertices with color Blue or Red. An edge is colorful
if it contains vertices of both colors, otherwise we call it monochromatic. Let c( H)
be the maximum number of colorful edges one can achieve by a coloring. By a
probabilistic argument, it is easy to show that c(H) is at least IEII2 (in a random
coloring, an edge will be colorful with probability at least 1/2). Using 3.1.2, we
could prove the following theorem (for the proof see [9])
Theorem 3.1.3 There is a constant f3 > 0 such that finding a coloring with at
least (1 - t3)c(H) colorful edges is NP-hard. This statement holds even in the case
when every but one degree in H is at most 10
4
Proof for :F2
We follow the reduction used in [2]. Consider a hypergraph H(V, E) described
Theorem 3.2.1. Let V = {I, 2, . . " d + I}, where with the possible exception of
the vertex d + 1, all other vertices have degree at most 10. Every edge will have
at least 2 and at most 4 vertices. So the number of edges is at least (d + 1) /4.
On the Infeasibility of Training Neural Networks with Small Squared Errors
Let Pi be the
ith
375
unit vector in R d +l , Pi = (0,0 , . .. ,0,1,0, .. . ,0). Furthermore,
Xc = LiE C Pi for every edge C E E. Let S be a coloring with maximum number
of colorful edges. In this coloring denote by Al the set of colorful edges and by A2
the set of monochromatic edges. Clearly IAII = e(H).
Our data set will be the following (inputs are from R d +l instead of from R d , but it
makes no difference)
where (Pd+1,1/2)t and (Od+l , l)t means (Pd+1, 1/2) and (Od+l, 1) are repeated t
times in the data set, resp. Similarly to [2], consider two vectors a and b in R d +l
where
a = (al,"" ad+l), ai = -1 if i is Red and ai
b = (b l , . .. , bd+l) , bi = -1 if i is Blue and bi
= d + 1 otherwise
= d + 1 otherwise
=
It is not difficult to verify that the function fa
(1/2)(step (ax + 1/2) + step (bx
1/2)) fits the data perfectly, thus e:F2 = IIEjal1 = O.
Suppose f
= (1/2) (step (ex -
+
I) + step (dx - 6? satisfies
M
MllEjW = 2)f(Xd - Yi)2 < Mc 2
i=l
Since if f(X i )
Po = l{i.J(Xd
# Yi
then U(Xi )
# Ydl < 4Mc 2 =
-
Yi)2 2: 1/4, the previous inequality implies:
p
The ratio po/Mis called misclassification ratio, and we will show that this ratio
cannot be arbitrary small. In order to avoid unnecessary ceiling and floor symbols,
we assume the upper-bound p is an integer. We choose t
P so that we can also
assume that (Od+l, 1) and (Pd+l, 1/2) are well classified. Let Hl (H2) be the half
space consisting of x: ex - 'Y > 0 (dx - 6 > 0). Note that Od E HI n H2 and
Pd+l E fI I U fI 2. Now let P l denote the set of i where Pi t/:. HI, and P 2 the set of i
such that Pi E Hl n H 2 ? Clearly, if j E P2 , then f(pj) # Yj, hence: IP2 1::; p. Let
Q = {C E EIC n P2 # 0}. Note that for each j E P2, the degree of j is at most 10,
thus: IQI ::; 10!?:?1 ::; lOp
=
Let A~ = {Clf(xc) = I}. Since less than p points are misclassified, IA~ .0. A I I < p.
Color V by the following rule: (1) if Pi E PI, then i is Red; (2) if Pi E P2 , color i
arbitrarily, either Red or Blue; (3) if Pi t/:. P l U P2 , then i is Blue.
Now we can finish the proof by the following two claims:
Claim 1: Every edge in
statement.
A~ \Q
is colorful. It is left to readers to verify this simple
Claim 2: IA~ \QI is close to IAII ?
Notice that:
IAI \(A~ \Q)I ::; IAI.0.A~ 1+ IQI ::; p + lOp = IIp
Observe that the size of the data set is M = d + 2t + lEI, so lEI + d 2: M - 2t =
M - 2p. Moreover, lEI 2: (d + 1)/4, so lEI 2: (1/5)(M - 2p). On the other hand,
IAII2: (1/2)IEI, all together we obtain; IAII2: (1/10)(M - p), which yields:
V. H. Vu
376
=
Choose f
f4 such that k(f4) ~ f3 (see Theortm 3.1.3). Then
for the class ;:2. This completes the proof. Q.E.D.
f4
will be a threshold
Due to space limitation, we omit the proofs for other classes and refer to [9]. However, let us at least describe (roughly) the general method to handle these cases.
The method consists of following steps:
? Extend the data set in the previous proof by a set of (special) points.
? Set the multiplicities of the special points sufficiently high so that those points
should be well-classified.
? If we choose the special points properly, the fact that these points are well-classified
will determine (roughly) the behavior of all but 2 nodes. In general we will show
that all but 2 nodes have little influence on the outputs of non-special data points.
? The problem basically reduces to the case of two nodes. By modifying the previous
proof, we could achieve the desired thresholds.
5
Remarks and open problems
? Readers may argue about the existence of (somewhat less natural) data points of
high multiplicities. We can avoid using these data points by a combinatorial trick
described in [9].
? The proof in Section 4 could be carried out using Theorem 3.1.2. However, we
prefer using the hypergraph coloring terminology (Theorem 3.1.3), which is more
convenient and standard. Moreover, Theorem 3.1.3 itself is interesting, and has not
been listed among well known "approximation is hard" theorems.
? It remains an open question to determine the right order of magnitude of thresholds for all the classes we considered. (see Section 1). By technical reasons, in the
Main theorem, the thresholds for more than two nodes involve the dimension (d).
We conjecture that there are dimension-free thresholds.
Acknowledgement We wish to thank A. Blum, A. Barron and 1. Lovasz for many
useful ideas and discussions.
References
[1] S. Arora and C. Lund Hardness of approximation, book chapter, preprint
[2] A. Blum, R. Rivest Training a 3-node neural network is NP-hard Neutral Networks, Vol 5., p 117-127, 1992
[3] A. Blumer, A. Ehrenfeucht, D. Haussler, M. Warmuth, Learnability and the
Vepnik-Chervonenkis Dimension, Journal ofthe Association for computing Machinery, Vol 36, No.4, 929-965, 1989.
[4] M. Garey and D. Johnson, Computers and intractability: A guide to the theory
of NP-completeness, San Francisco, W.H.Freeman, 1979
On the Infeasibility o/Training Neural Networks with Small Squared Errors
377
[5] D. Haussler, Generalizing the PAC model for neural net and other learning
applications (Tech. Rep. UCSC-CRL-89-30). Santa Cruz. CA: University of
California 1989.
[6] L. J ones, The computational intractability of training sigmoidal neural networks
(preprint)
[7] J. Judd Neutral Networks and Complexity of learning, MIT Press 1990.
[8] N. Meggido, On the complexity of polyhedral separability (Tech. Rep. RJ 5252)
IBM Almaden Research Center, San Jose, CA
[9] V. H. Vu, On the infeasibility of training neural networks with small squared
error. manuscript.
| 1441 |@word polynomial:8 norm:12 open:3 mention:2 reduction:3 contains:5 chervonenkis:1 od:5 dx:2 bd:2 cruz:1 fn:5 half:2 warmuth:1 ith:1 emperical:2 completeness:1 math:1 node:13 sigmoidal:1 constructed:2 ucsc:1 prove:3 consists:1 polyhedral:1 hardness:2 behavior:1 roughly:2 inspired:1 freeman:1 decreasing:1 little:1 provided:1 notation:2 moreover:3 avarage:1 rivest:2 finding:3 every:9 xd:4 rm:1 unit:2 colorful:7 omit:1 positive:9 iqi:2 interpolation:2 lief:3 examined:1 lop:2 bi:5 yj:1 vu:5 convenient:1 cannot:1 close:1 context:1 influence:1 center:1 go:1 yj2:1 m2:1 rule:1 haussler:2 his:1 handle:2 resp:1 suppose:1 origin:1 associate:1 element:2 trick:1 preprint:2 pd:4 complexity:3 traing:1 hypergraph:4 terminating:1 depend:1 efficiency:1 f2:4 po:2 various:1 chapter:1 train:1 describe:2 tell:1 posed:1 solve:1 say:2 otherwise:5 itself:1 net:1 eic:1 combining:1 poorly:1 achieve:3 subindices:1 produce:3 perfect:1 oo:1 ard:1 p2:5 solves:1 involves:1 implies:4 closely:1 correct:2 f4:3 modifying:1 f1:1 fix:1 investigation:1 hold:1 sufficiently:2 considered:3 great:1 claim:3 smallest:1 a2:1 purpose:1 combinatorial:1 iaii:2 lovasz:1 mit:1 clearly:2 rather:1 avoid:3 ej:1 corollary:1 ax:1 properly:2 tech:2 sense:2 going:1 misclassified:1 among:1 ill:1 almaden:1 special:4 construct:1 f3:2 having:1 jones:2 np:18 t2:1 few:2 consisting:1 fef:1 investigate:2 edge:14 machinery:1 iv:1 desired:1 earlier:1 assignment:4 vertex:9 subset:1 neutral:2 johnson:1 too:1 loo:1 learnability:1 answer:3 lee:1 probabilistic:1 together:1 squared:6 satisfied:1 containing:2 choose:3 iip:1 literal:5 book:1 leading:1 bx:1 iei:1 matter:1 satisfy:1 depends:1 vi:1 ad:1 try:1 view:1 red:4 square:1 cnfs:1 t3:1 yield:1 ofthe:1 weak:1 basically:1 mc:2 app:1 classified:3 fo:1 definition:3 garey:1 proof:13 mi:1 rational:1 proved:2 color:5 organized:1 ea:1 coloring:7 appears:2 manuscript:1 follow:2 iai:3 done:1 though:1 furthermore:2 hand:1 logistic:1 lei:4 grows:1 b3:2 verify:2 former:3 hence:1 ehrenfeucht:1 generalized:1 complete:1 demonstrate:1 ef:3 fi:3 common:1 sigmoid:2 clause:9 yil:1 ydl:1 he:2 extend:1 association:1 refer:2 ai:6 rd:2 mathematics:1 similarly:1 etc:1 showed:1 recent:1 certain:2 inequality:1 rep:2 arbitrarily:1 yi:12 somewhat:1 floor:1 eo:2 determine:2 ud:1 ii:2 rj:1 reduces:1 technical:1 concerning:1 qi:1 circumstance:1 metric:1 achieved:1 completes:1 crucial:2 monochromatic:2 call:3 integer:2 iii:1 easy:2 fit:5 finish:1 perfectly:1 inner:3 regarding:1 idea:2 cnf:1 remark:2 useful:1 clear:1 listed:1 involve:1 santa:1 exist:1 notice:1 blue:4 write:1 vol:2 key:1 terminology:1 threshold:15 blum:3 achieving:2 pj:1 monotone:1 prob:1 jose:1 reader:2 prefer:2 bit:1 bound:2 hi:2 layer:3 fl:1 distinguish:1 display:1 yale:3 argument:3 min:1 relatively:1 conjecture:1 department:1 ip2:1 smaller:1 separability:1 lem:1 hl:2 multiplicity:2 ceiling:1 computationally:2 numbel:1 remains:1 discus:1 end:1 observe:1 barron:1 existence:2 original:1 denotes:3 xc:2 lipschitzian:1 exploit:1 hypercube:1 question:11 quantity:1 occurs:1 fa:1 said:1 thank:1 argue:1 clf:1 reason:2 length:1 ratio:3 difficult:2 statement:3 negative:3 upper:1 extended:2 arbitrary:2 complement:1 namely:1 required:2 specified:1 connection:1 california:1 usually:3 lund:1 max:4 ia:2 misclassification:1 difficulty:1 natural:1 arora:1 carried:1 acknowledgement:1 relative:8 interesting:2 limitation:2 h2:2 degree:4 intractability:2 pi:9 ibm:1 course:1 last:1 free:1 infeasible:1 guide:1 absolute:2 van:1 dimension:4 judd:2 author:3 collection:1 san:2 dealing:1 sat:1 conclude:1 unnecessary:1 francisco:1 xi:9 infeasibility:5 why:1 nature:1 ca:2 flf:2 main:4 repeated:1 referred:1 ff:6 sub:1 seminar:1 wish:1 lie:1 third:2 down:1 theorem:15 formula:2 pac:1 symbol:1 intractable:1 importance:1 ci:2 magnitude:2 generalizing:1 truth:4 satisfies:4 goal:1 formulated:1 blumer:1 jf:1 crl:1 hard:18 called:2 exception:1 ex:2 |
485 | 1,442 | Using Helmholtz Machines to analyze
multi-channel neuronal recordings
Virginia R. de Sa
desa@phy.ucsf.edu
R. Christopher deC harms
decharms@phy.ucsf.edu
Michael M. Merzenich
merz@phy.ucsf.edu
Sloan Center for Theoretical Neurobiology and
W. M. Keck Center for Integrative Neuroscience
University of California, San Francisco CA 94143
Abstract
One of the current challenges to understanding neural information
processing in biological systems is to decipher the "code" carried
by large populations of neurons acting in parallel. We present an
algorithm for automated discovery of stochastic firing patterns in
large ensembles of neurons. The algorithm, from the "Helmholtz
Machine" family, attempts to predict the observed spike patterns in
the data. The model consists of an observable layer which is directly
activated by the input spike patterns, and hidden units that are activated through ascending connections from the input layer. The
hidden unit activity can be propagated down to the observable layer
to create a prediction of the data pattern that produced it. Hidden
units are added incrementally and their weights are adjusted to improve the fit between the predictions and data, that is, to increase a
bound on the probability of the data given the model. This greedy
strategy is not globally optimal but is computationally tractable for
large populations of neurons. We show benchmark data on artificially constructed spike trains and promising early results on neurophysiological data collected from our chronic multi-electrode cortical
implant.
1
Introduction
Understanding neural processing will ultimately require observing the response
patterns and interactions of large populations of neurons. While many studies
have demonstrated that neurons can show significant pairwise interactions, and
that these pairwise interactions can code stimulus information [Gray et aI., 1989,
Meister et aI., 1995, deCharms and Merzenich, 1996, Vaadia et al., 1995], there is
currently little understanding of how large ensembles of neurons might function together to represent stimuli. This situation has arisen partly out of the historical
V R. d. Sa, R. C. deCharms and M. M. Merzenich
132
difficulty of recording from large numbers of neurons simultaneously. Now that this
is becoming technically feasible, the remaining analytical challenge is to understand
how to decipher the information carried in distributed neuronal responses.
Extracting information from the firing patterns in large neuronal populations is difficult largely due to the combinatorial complexity of the problem, and the uncertainty about how information may be encoded. There have been several attempts
to look for higher order correlations [Martignon et al., 1997] or decipher the activity
from multiple neurons, but existing methods are limited in the type of patterns they
can extract assuming absolute reliability of spikes within temporal patterns of small
numbers of neurons [Abeles, 1982, Abeles and Gerstein, 1988, Abeles et al., 1993,
Schnitzer and Meister,] or considering only rate codes [Gat and Tishby, 1993,
Abeles et al., 1995]. Given the large numbers of neurons involved in coding sensory
events and the high variability of cortical action potentials, we suspect that meaningful ensemble coding events may be statistically similar from instance to instance
while not being identical. Searching for these type of stochastic patterns is a more
challenging task.
One way to extract the structure in a pattern dataset is to construct a generative
model that produces representative data from hidden stochastic variables. Helmholtz
machines [Hinton et al., 1995, Dayan et al., 1995] efficiently [Frey et al., 1996] produce generative models of datasets by maximizing a lower bound on the log likelihood
of the data. Cascaded Redundancy Reduction [de Sa and Hinton, 1998] is a particularly simple form of Helmholtz machine in which hidden units are incrementally
added. As each unit is added, it greedily attempts to best model the data using all
the previous units. In this paper we describe how to apply the Cascaded Redundancy Reduction algorithm to the problem of finding patterns in neuronal ensemble
data, test the performance of this method on artificial data, and apply the method
to example neuronal spike trains.
1.1
Cascaded Redundancy Reduction
The simplest form of generative model is to model each observed (or input) unit as
a stochastic binary random variable with generative bias bi. This generative input is
passed through a transfer function to give a probability of firing.
p. = a(b?) =
~
~
1
1 + e- bi
(1)
While this can model the individual firing rates of binary units, it cannot account
for correlations in firing between units. Correlations can be modeled by introducing
hidden units with generative weights to the correlated observed units. By cascading
hidden units as in Figure 1, we can represent higher order correlations. Lower units
sum up their total generative input from higher units and their generative bias.
Xi
= bi
+
L
Sj9j,i
(2)
j>i
Finding the optimal generative weights (9j,i, bi) for a given dataset involves an intractable search through an exponential number of possible states of the hidden units.
Helmholtz machines approximate this problem by using forward recognition connections to compute an approximate distribution over hidden states for each data pattern. Cascaded Redundancy Reduction takes this approximation one step further by
approximating the distribution by a single state. This makes the search for recognition and generative weights much simpler. Given a data vector, d, considering the
state produced by the recognition connections as Sd gives a lower bound on the log
Using Helmholtz Machines to Analyze Multi-channel Neuronal Recordings
133
-:.> generative connections
~ recognition connections
TO,k
T
i,k
Figure 1: The Cascaded Redundancy Reduction Network. Hidden units are added
incrementally to help better model the data.
likelihood of the data. Units are added incrementally with the goal of maximizing
this lower bound, C,
C
= 'L[(s% log a(b k ) + (l-s%) log(l-a(bk))+ L
d
sf log a(xf)+ (I-sf) log(l-a(xf))]
i
(3)
Before a unit is added it is considered as a temporary addition. Once its weights
have been learned, it is added to the permanent network only if adding it reduces
the cost on an independent validation set from the same data distribution. This is
to prevent overtraining. While a unit is considered for addition, all weights other
than those to and from the new unit and the generative bias weights are fixed. The
learning of the weights to and from the new unit is then a fairly simple optimization
problem involving treating the unit as stochastic, and performing gradient descent
on the resulting modified lower bound.
2
Method
This generic pattern finding algorithm can be applied to multi-unit spike trains by
treating time as another spatial dimension as is often done for time series data. The
spikes are binned on the order of a few to tens of milliseconds and the algorithm looks
for patterns in finite time length windows by sliding a window centered on each spike
from a chosen trigger channel. An example extracted window using channel 4 as the
trigger channel is shown in Figure 2.
Because the number of spikes can be larger than one, the observed units (bins) are
modeled as discrete Poisson random variables rather than binary random variables
(the hidden units are still kept as binary units). To reflect the constraint that the
expected number of spikes cannot be negative but may be larger than one, the transfer
function for these observed bins was chosen to be exponential. Thus if Xi is the
total summed generative input, Ai, the expected mean number of spikes in bin i, is
calculated as eX; and the probability of finding s spikes in that bin is given by
s!
(4)
V.R. d. Sa, R. C. deChanns and M. M. Merzenich
134
Figure 2: The input patterns for the algorithm are windows from the full spatiatemporal firing patterns. The full dataset is windows centered about every spike in
the trigger channel.
The terms in the lower bound objective function due to the observed bins are modified
accordingly.
3
Experimental Results
Before applying the algorithm to real neural spike trains we have characterized its
properties under controlled conditions. We constructed sample data containing two
random patterns across 10 units spanning 100 msec. The patterns were stochastic
such that each neuron had a probability of firing in each time bin of the pattern.
Sample patterns were drawn from the stochastic pattern templates and embedded in
other "noise" spikes. The sample pattern templates are shown in the first column
of Figure 3. 300 seconds of independent training, validation and test data were
generated. All results are reported on the test data .
After training the network, performance was assessed by stepping through the test
data and observing the pattern of activation across the hidden units obtained from
propagating activity through the forward (recognition) connections and their corresponding generative pattern {Ad obtained from the generative connections from the
binary hidden unit pattern. Typically, many of the theoretically possible 2n hidden
unit patterns do not occur. Of the ones that do, several may code for the noise background. A crucial issue for interpreting patterns in real neural data is to discover
which of the hidden unit activity patterns correspond to actual meaningful patterns.
We use a measure that calculates the quality of the match of the observed pattern
and the generative pattern it invokes. As the algorithm was not trained on the test
data, close matches between the generative pattern and the observed pattern imply
real structure that is common to the training and test dataset. With real neural
data, this question can also be addressed by correlating the occurrence of patterns to
stimuli or behavioural states of the animal.
One match measure we have used to pick out temporally modulated structure is
the cost of coding the observed units using the hidden unit pattern compared to
the cost of using the optimal rate code for that pattern (derived by calculating the
firing rate for each channel in the window excluding the trigger bin). Match values
were calculated for each hidden unit pattern by averaging the results across all its
contributing observed patterns. Typical generative patterns of the added template
patterns (in noise) are shown in the second column of Figure 3. The third column
in the figure shows example matches from the test set, (Le. patterns that activated
the hidden unit pattern corresponding to the generative pattern in column 2). Note
that the instances of the patterns are missing some spikes present in the template,
and are surrounded by many additional spikes.
Using Helmholtz Machines to Analyze Multi-channel Neuronal Recordings
Generative
Pattern
Template
135
Test set
Example
Pattern 1
Pattern 2
Figure 3: Pattern templates, resulting generative patterns after training (showing the
expected number of spikes the algorithm predicts for each bin), and example test set
occurrences. The size and shade of the squares represents the probability of activation
of that bin (or 0/1 for the actual occurrences), the colorbars go from 0 to 1.
We varied both the frequency of the pattern occurrences and that of the added
background spikes. Performance as a function of the frequency of the background
spikes is shown on the left in Figure 4 for a pattern frequency of .4 Hz. Performance
as a function of the pattern frequency for a noise spike frequency of 15Hz is shown
on the right of the Figure. False alarm rates were extremely low ranging from 0-4%
across all the tested conditions. Also, importantly, when we ran three trials with no
added patterns"no patterns were detected by the algorithm.
~!
i
e-
...., .118, . ~1
0(.88, .95, .97)
x(.73, .94, .94)
i
e-
0.8
"i
~
"iN
s::
"2
'"
.~
0.6
co
0 .4
..
0(0, .73, .82)
0(0, .45, .9)
E
CD
Q.
'0
I,
.~
"
~
x(.73, .94, .94)
0.8
pattem 1
pattem 2
shifted Pattem 1
0.2
~I
'"
0.6
.
0.4
.~
E
CD
co
Q.
'0
.(0, .83, .90)
.~
x(O, 0, .51)
Q.
"
~
0(0, 0, .87)
0.2
Q.
x(O, 0, .32)
O L - - L_ _~_ _~-L_ _~_ _~~_ _~
0
14
17
18
19
16
20
average firing 'ale 01 back ground spikas (Hz)
15
21
0.1
0.15
0.2
0.25
0.3
0.35
0.4
frequency 01 pattem occurrence (Hz)
0.45
0 .5
Figure 4: Graphs showing the effect of adding more background spikes (left) and
decreasing the number of pattern occurrences in the dataset (right) on the percentage
of patterns correctly detected. The detection of shifted pattern is due to the presence
of a second spike in channel 4 in the pattern (hits for this case are only calculated
for the times when this spike was present - the others would all be missed). In fact
in some cases the presence of the only slightly probable 3rd bin in channel 4 was
enough to detect another shifted pattern 1. Means over 3 trials are plotted with the
individual trial values given in braces
The algorithm was then applied to recordings made from a chronic array of extracellular microelectrodes placed in the primary auditory cortex of one adult marmoset
monkey and one adult owl monkey [deC harms and Merzenich, 1998]. On some elec-
V.R. d. Sa, R. C. deCharms and M. M. Merzenich
136
Figure 5: Data examples (all but top left) from neural recordings in an awake marmoset monkey that invoke the same generative pattern (top left). The instances are
patterns from the test data that activated the same hidden unit activity pattern resulting in the generative pattern in the top left. The data windows were centered
around all the spikes in channel 4. The brightest bins in the generative pattern represent an expected number of spikes of 1. 7. In the actual patterns, The darkest and
smallest bins represent a bin with 1 spike; each discrete grayscale/size jump represents an additional spike. Each subfigure is indiv!dually normalized to the bin with
the most spikes.
trodes spikes were isolated from individual neurons; others were derived from small
clusters of nearby neurons. Figure 5 shows an example generative pattern (accounting
for 2.8% of the test data) that had a high match value together with example occurrences in the test data. The data were responses recorded to vocalizations played to
the marmoset monkey, channel 4 was used as the trigger channel and 7 hidden units
were added.
4
Discussion
We have introduced a procedure for searching for structure in multineuron spike
trains, and particularly for searching for statistically reproducible stochastic temporal
events among ensembles of neurons. We believe this method has great promise for
exploring the important question of ensemble coding in many neuronal systems, a
crucial part of the problem of understanding neural information coding. The strengths
of this method include the ability to deal with stochastic patterns, the search for any
type of reproducible structure including the extraction of patterns of unsuspected
nature, and its efficient, greedy, search mechanism that allows it to be applied to
large numbers of neurons.
Acknowledgements
We would like to acknowledge Geoff Hinton for useful suggestions in the early stages
of this work, David MacKay for helpful comments on an earlier version of the
manuscript, and the Sloan Foundation for financial support.
Using Helmholtz Machines to Analyze Multi-channel Neuronal Recordings
137
References
[Abeles, 1982] Abeles, M. (1982). Local Cortical Circuits An Electrophysiological
Study, volume 6 of Studies of Brain Function. Springer-Verlag.
[Abeles et al., 1995] Abeles, M., Bergman, H., Gat, I., Meilijson, I., Seidemann, E.,
Tishby, N., and Vaadia, E. (1995). Cortical activity flips among quasi-stationary
states. Proceedings of the National Academy of Science, 92:8616- 8620.
[Abeles et al., 1993] Abeles, M., Bergman, H., Margalit, E., and Vaadia, E. (1993).
Spatiotemporal firing patterns in the frontal cortex of behaving monkeys. Journal
of Neurophysiology, 70(4):1629-1638.
[Abeles and Gerstein, 1988] Abeles, M. and Gerstein, G. L. (1988). Detecting spatiotemporal firing patterns among simultaneously recorded single neurons. Journal
of Neurophysiology, 60(3).
[Dayan et al., 1995] Dayan, P., Hinton, G. E., Neal, R. M., and Zemel, R. S. (1995).
The helmholtz machine. Neural Computation, 7:889-904.
[de Sa and Hinton, 1998] de Sa, V. R. and Hinton, G. E. (1998). Cascaded redundancy reduction. to appear in Network{February).
[deC harms and Merzenich, 1996] deCharms, R. C. and Merzenich, M. M. (1996). Primary cortical representation of sounds by the coordination of action-potential timing. Nature, 381:610-613.
[deCharms and Merzenich, 1998] deCharms, R. C. and Merzenich, M. M. (1998).
Characterizing neurons in the primary auditory cortex of the awake primate using
reverse correlation. this volume.
[Frey et al., 1996] Frey, B. J., Hinton, G. E., and Dayan, P. (1996). Does the wakesleep algorithm produce good density estimators? In Touretzky, D., Mozer, M.,
and Hasselmo, M., editors, Advances in Neural Information Processing Systems 8,
pages 661-667. MIT Press.
[Gat and Tishby, 1993] Gat, I. and Tishby, N. (1993). Statistical modeling of cellassemblies activities in associative cortex of behaving monkeys. In Hanson, S.,
Cowan, J., and Giles, C., editors, Advances in Neural Information Processing Systems 5, pages 945-952. Morgan Kaufmann.
[Grayet al., 1989] Gray, C. M., Konig, P., Engel, A. K., and Singer, W. (1989).
Oscillatory responses in cat visual cortex exhibit inter-columnar synchronization
which reflects global stimulus properties. Nature, 338:334-337.
[Hinton et al., 1995] Hinton, G. E., Dayan, P., Frey, B. J., and Neal, R. M. (1995).
The wake-sleep algorithm for unsupervised neural networks. Science, 268:1158116l.
[Martignon et al., 1997] Martignon, L., Laskey, K., Deco, G., and Vaadia, E. (1997).
Learning exact patterns of quasi-synchronization among spiking neurons from data
on multi-unit recordings. In Mozer, M., Jordan, M., and Petsche, T., editors,
Advances in Neural Information Processing Systems 9, pages 76-82. MIT Press.
[Meister et al., 1995] Meister, M., Lagnado, L., and Baylor, D. (1995). Concerted
signaling by retinal ganglion cells. Science, 270:95-106.
[Schnitzer and Meister,] Schnitzer, M. J. and Meister, M. Information theoretic
identification of neural firing patterns from multi-electrode recordings. in preparation.
[Vaadia et al., 1995] Vaadia, E., Haalman, I., Abeles, M., Bergman, H., Prut, Y.,
Slovin, H., and Aertsen, A. (1995). Dynamics of neuronal interactions in monkey
cortex in relation to behavioural events. Nature, 373:515-518.
| 1442 |@word neurophysiology:2 trial:3 version:1 integrative:1 accounting:1 pick:1 schnitzer:3 reduction:6 phy:3 series:1 existing:1 current:1 activation:2 multineuron:1 treating:2 reproducible:2 stationary:1 greedy:2 generative:25 accordingly:1 detecting:1 simpler:1 constructed:2 consists:1 concerted:1 theoretically:1 pairwise:2 inter:1 expected:4 multi:8 brain:1 globally:1 decreasing:1 little:1 actual:3 window:7 considering:2 discover:1 circuit:1 monkey:7 finding:4 temporal:2 every:1 hit:1 unit:38 appear:1 before:2 frey:4 local:1 sd:1 timing:1 firing:12 becoming:1 might:1 challenging:1 co:2 limited:1 bi:4 statistically:2 signaling:1 procedure:1 indiv:1 cannot:2 close:1 applying:1 demonstrated:1 center:2 chronic:2 maximizing:2 missing:1 go:1 estimator:1 cascading:1 importantly:1 array:1 financial:1 population:4 searching:3 trigger:5 exact:1 bergman:3 helmholtz:9 recognition:5 particularly:2 predicts:1 observed:10 unsuspected:1 ran:1 mozer:2 complexity:1 dynamic:1 ultimately:1 trained:1 technically:1 geoff:1 cat:1 train:5 elec:1 describe:1 artificial:1 detected:2 zemel:1 encoded:1 larger:2 ability:1 vocalization:1 associative:1 vaadia:6 analytical:1 interaction:4 academy:1 konig:1 electrode:2 keck:1 cluster:1 produce:3 help:1 propagating:1 sa:7 involves:1 stochastic:9 centered:3 owl:1 bin:14 require:1 biological:1 probable:1 adjusted:1 exploring:1 around:1 considered:2 ground:1 brightest:1 great:1 predict:1 early:2 smallest:1 combinatorial:1 currently:1 coordination:1 hasselmo:1 create:1 engel:1 reflects:1 mit:2 modified:2 rather:1 derived:2 likelihood:2 greedily:1 detect:1 helpful:1 dayan:5 typically:1 margalit:1 hidden:20 relation:1 quasi:2 issue:1 among:4 animal:1 spatial:1 summed:1 fairly:1 mackay:1 construct:1 once:1 extraction:1 identical:1 represents:2 look:2 unsupervised:1 others:2 stimulus:4 few:1 microelectrodes:1 simultaneously:2 national:1 individual:3 attempt:3 detection:1 activated:4 plotted:1 isolated:1 theoretical:1 subfigure:1 instance:4 column:4 earlier:1 modeling:1 giles:1 cost:3 introducing:1 virginia:1 tishby:4 reported:1 spatiotemporal:2 abele:13 density:1 invoke:1 michael:1 together:2 reflect:1 recorded:2 deco:1 containing:1 desa:1 account:1 potential:2 de:4 retinal:1 coding:5 permanent:1 sloan:2 ad:1 analyze:4 observing:2 meilijson:1 parallel:1 square:1 kaufmann:1 largely:1 efficiently:1 ensemble:6 correspond:1 prut:1 decipher:3 identification:1 produced:2 overtraining:1 oscillatory:1 touretzky:1 martignon:3 frequency:6 involved:1 propagated:1 auditory:2 dataset:5 electrophysiological:1 back:1 manuscript:1 higher:3 response:4 done:1 stage:1 correlation:5 christopher:1 incrementally:4 quality:1 laskey:1 gray:2 believe:1 effect:1 normalized:1 merzenich:10 neal:2 haalman:1 deal:1 theoretic:1 interpreting:1 ranging:1 common:1 spiking:1 stepping:1 volume:2 significant:1 ai:3 rd:1 had:2 reliability:1 cortex:6 behaving:2 seidemann:1 reverse:1 verlag:1 binary:5 morgan:1 additional:2 ale:1 sliding:1 multiple:1 full:2 sound:1 reduces:1 match:6 xf:2 characterized:1 controlled:1 calculates:1 prediction:2 involving:1 poisson:1 represent:4 arisen:1 cell:1 dec:3 addition:2 background:4 addressed:1 wake:1 crucial:2 brace:1 comment:1 recording:9 suspect:1 hz:4 cowan:1 jordan:1 extracting:1 presence:2 enough:1 automated:1 fit:1 passed:1 action:2 useful:1 ten:1 simplest:1 percentage:1 millisecond:1 shifted:3 neuroscience:1 correctly:1 discrete:2 promise:1 redundancy:6 drawn:1 prevent:1 kept:1 graph:1 sum:1 uncertainty:1 pattem:4 family:1 missed:1 gerstein:3 layer:3 bound:6 played:1 sleep:1 activity:7 binned:1 occur:1 strength:1 constraint:1 awake:2 nearby:1 extremely:1 performing:1 extracellular:1 across:4 slightly:1 primate:1 marmoset:3 computationally:1 behavioural:2 mechanism:1 singer:1 flip:1 ascending:1 tractable:1 meister:6 apply:2 generic:1 petsche:1 occurrence:7 darkest:1 top:3 remaining:1 include:1 calculating:1 invokes:1 approximating:1 february:1 objective:1 added:11 question:2 spike:31 strategy:1 primary:3 aertsen:1 exhibit:1 gradient:1 collected:1 spanning:1 assuming:1 code:5 length:1 modeled:2 baylor:1 difficult:1 decharms:7 negative:1 neuron:18 datasets:1 benchmark:1 finite:1 acknowledge:1 descent:1 situation:1 neurobiology:1 variability:1 hinton:9 excluding:1 varied:1 dually:1 bk:1 introduced:1 david:1 connection:7 hanson:1 california:1 learned:1 temporary:1 adult:2 pattern:72 challenge:2 including:1 event:4 difficulty:1 cascaded:6 improve:1 imply:1 temporally:1 carried:2 extract:2 lagnado:1 understanding:4 discovery:1 acknowledgement:1 contributing:1 embedded:1 wakesleep:1 synchronization:2 suggestion:1 validation:2 foundation:1 slovin:1 editor:3 surrounded:1 cd:2 placed:1 l_:2 bias:3 understand:1 template:6 characterizing:1 absolute:1 distributed:1 dimension:1 cortical:5 calculated:3 sensory:1 forward:2 made:1 jump:1 san:1 historical:1 approximate:2 observable:2 global:1 correlating:1 harm:3 francisco:1 xi:2 grayscale:1 search:4 promising:1 nature:4 channel:14 transfer:2 ca:1 correlated:1 artificially:1 noise:4 alarm:1 neuronal:10 representative:1 msec:1 exponential:2 sf:2 third:1 down:1 shade:1 showing:2 intractable:1 false:1 adding:2 gat:4 implant:1 columnar:1 ganglion:1 neurophysiological:1 visual:1 springer:1 extracted:1 goal:1 feasible:1 typical:1 acting:1 averaging:1 total:2 partly:1 experimental:1 merz:1 meaningful:2 support:1 modulated:1 assessed:1 ucsf:3 preparation:1 frontal:1 tested:1 ex:1 |
486 | 1,443 | An Annealed Self-Organizing Map for Source
Channel Coding
Matthias Burger, Thore Graepel, and Klaus Obermayer
Department of Computer Science
Technical University of Berlin
FR 2-1, Franklinstr. 28/29, 10587 Berlin, Germany
{burger, graepel2, oby} @cs.tu-berlin.de
Abstract
We derive and analyse robust optimization schemes for noisy vector
quantization on the basis of deterministic annealing. Starting from a
cost function for central clustering that incorporates distortions from
channel noise we develop a soft topographic vector quantization algorithm (STVQ) which is based on the maximum entropy principle
and which performs a maximum-likelihood estimate in an expectationmaximization (EM) fashion. Annealing in the temperature parameter f3
leads to phase transitions in the existing code vector representation during the cooling process for which we calculate critical temperatures and
modes as a function of eigenvectors and eigenvalues of the covariance
matrix of the data and the transition matrix of the channel noise. A whole
family of vector quantization algorithms is derived from STVQ, among
them a deterministic annealing scheme for Kohonen's self-organizing
map (SOM). This algorithm, which we call SSOM, is then applied to
vector quantization of image data to be sent via a noisy binary symmetric
channel. The algorithm's performance is compared to those of LBG and
STVQ. While it is naturally superior to LBG, which does not take into
account channel noise, its results compare very well to those of STVQ,
which is computationally much more demanding.
1
INTRODUCTION
Noisy vector quantization is an important lossy coding scheme for data to be transmitted
over noisy communication lines. It is especially suited for speech and image data which
in many applieations have to be transmitted under low bandwidth/high noise level conditions. Following the idea of (Farvardin, 1990) and (Luttrell, 1989) of jointly optimizing
the codebook and the data representation w.r.t. to a given channel noise we apply a deterministic annealing scheme (Rose, 1990; Buhmann, 1997) to the problem and develop a
431
An Annealed Self-Organizing Map for Source Channel Coding
soft topographic vector quantization algorithm (STVQ) (cf. Heskes, 1995; Miller, 1994).
From STVQ we can derive a class of vector quantization algorithms, among which we
find SSOM, a deterministic annealing variant of Kohonen's self-organizing map (Kohonen,
1995), as an approximation. While the SSOM like the SOM does not minimize any known
energy function (Luttre11, 1989) it is computationally less demanding than STVQ. The deterministic annealing scheme enables us to use the neighborhood function of the SOM
solely to encode the desired transition probabiliti~;;s of the channel noise and thus opens
up new possibilities for the usage of SOMs with arbitrary neighborhood functions. We
analyse phase transitions during the annealing and demonstrate the performance of SSOM
by applying it to lossy image data compression for transmission via noisy channels.
2
DERIVATION OF A CLASS OF VECTOR QUANTIZERS
Vector quantization is a method of encoding data by grouping the data vectors and providing a representative in data space for each group. Given a set X of data vectors Xi E
~, i
1, ... , D, the objective of vector quantization is to find a set W of code vectors
Wr 1 r = 0, ... , N- 1, and a set M of binary assignment variables IDir. Lr IDir = 1, Vi,
such that a given cost function
=
(1)
r
is minimized. Er (Xi, W) denotes the cost of assigning data point Xi to code vector Wr.
Following an idea by (Luttrell, 1994) we consider the case that the code labels r form a
compressed encoding of the data for the purpose of transmission via a noisy channel (see
Figure 1). The distortion caused by the channel noise is modeled by a matrix H of tran1 , Vr, for the noise induced change of assignment of
sition probabilities hrs. La hrs
a data vector Xi from code vector Wr to code vector W 8 ? After transmission the received
index s is decoded using its code vector w 8 ? Averaging the squared Euclidean distance
IIxi - w sll2 over a11 possible transitions yields the assignment costs
=
(2)
where the factor 1/2 is introduced for computational convenience.
Starting from the cost function E given in Eqs. (1), (2) the Gibbs-distribution
P (M, WI X)
exp ( -,8 E (M, WI X)) can be obtained via the principle of maximum entropy under the constraint of a given average cost (E). The Lagrangian multiplier
,B is associated with {E) and is interpreted as an inverse temperature that determines the
fuzziness of assignments. In order to generalize from the given training set X we calculate the most likely set of code vectors from the probability distribution P (M, WI X)
marginalized over all legal sets of assignments M. For a given value of ,B we obtain
=!
Wr
where P(xi E s)
=
LiXi L 8 hrsP(xi E s)
LiLa hraP(xi E s) '
Vr,
(3)
= (mis).
P (Xi E s)
=
e:x;p
(-~ Lthat Jlxi- Wtll 2 )
Lu exp
(
-~ Lt hut llxi-
Wtll 2
),
(4)
is the assignment probability of data vector Xi to code vector Wa. Solving Eqs. (3), (4) by
fixed-point iteration comprises an expectation-maximization algorithm, where the E-step,
432
M Burger, T. Graepel and K Obermayer
Figure 1: Cartoon of a generic data communication problem. The encoder assigns
input vectors Xi to labeled code vectors
Wr.
Their indices r are the~ transmitted via a noisy channel which is characterized by a set of transition probabilities
hrs? The decoder expands the received index s to its code vector W 8 which represents the data vectors assigned to it during
encoding. The total error is measured via
the squared Euclidean distance between
the original data vector Xi and its representative w 5 averaged over all transitions
r -t s.
Q
I Em'OdPr
r------------.,
: Distortion
"
I
Chlllmel Noiie hn : r - s
I
x, -
w,
X;
II
x; - w? 11 2 :
1.------------..1
I
1?
w,
Decod~r
s -
w,
Eq. (4), determines the assignment probabilities P(xi E s) for all data points Xi and the
old code vectors w 8 and theM-step, Eq. (3), determines the new code vectors Wr from the
new assignment probabilities P(xi E s). In order to find the global minimum ofE, (3 0
is increased according to an annealing schedule which tracks the solution from the easily
solvable convex problem at low f3 to the exact solution of Eqs. (1 ), (2) at infinite j3. In the
following we call the solution of Eqs. (3), (4) soft topographic vector quantizer (STVQ).
=
Eqs. (3), (4) are the starting point for a whole class of vector quantization algorithms (Figure 2). The approximation hrs -t drs applied to Eq. (4) leads to a soft version of Kohonen's self-organzing map (SSOM), if additionally applied to Eq. (3) soft-clustering (SC)
(Rose, 1990) is recovered. f3 -t oo leads to the corresponding "hard" versions topographic
vector quantisation (TVQ) (Luttrell, 1989), self-organizing map (SOM) (Kohonen, 1995),
and LBG. In the following, we will focus on the soft self-organizing map (SSOM). SSOM
is computationally less demanding than STVQ, but offers - in contrast to the traditional
SOM - a robust deterministic annealing optimization scheme. Hence it is possible to extend the SOM approach to arbitrary non-trivial neighborhood functions hrs as required,
e.g. for source channel coding problems for noisy channels.
3
PHASE TRANSITIONS IN THE ANNEALING
From (Rose, 1990) it is known that annealing in f3 changes the representation of the data.
Code vectors split with increasing f3 and the size of the codebook for a fixed f3 is given by
the number of code vectors that have split up to that point. With non-diagonal H, however,
permutation symmetry is broken and the "splitting" behavior of the code vectors changes.
At infinite temperature every data vector Xi is assigned to every code vector w r with equal
probability P 0 (xi E r) = 1/N, where N is the size of the codebook. Hence all code
f:s 2:'::i Xi , Vr, of the data. Expanding the
vectors are located in the center of mass, w~
r.h.s. of Eq. (3) to first order around the fixed point { w~} and assuming hrs
hsr, 'r/ r, s,
we obtain the critical value
=
=
(5)
An Annealed Self-Organizing Map for Source Channel Coding
433
r
hrs-- t5rs
STVQ
hrs-- t5rs
(3---
M- Step
(3-
00
L
TVQ
(3--
.J
00
- - - - ..,
hrs- t5rs
SOM
E-Step
---
!
00
r
hrs- t5rs
.,
sc
SSOM
E-Step
---
LBG
M- Step
L
----
.J
Figure 2: Class of vector quantizers derived from STVQ, together with approximations and
limits (see text). The "S" in front stands for "soft" to indicate the probabilistic approach.
for the inverse temperature, at which the center of mass solution becomes unstable . .\~ax
is the largest eigenvalue of the covariance matrix C = ~ Li XiXf of the data and corresponds to their variance .\~ax = IT~ax along the principal axis which is given by the associated eigenvector v:;;ax and along which code vectors split. .\~ax is the largest eigenvalue
of a matrix G whose elements are given by grt = Ls hrs (hst- h). The rth component
of the corresponding eigenvector v~ax determines for each code vector Wr in which direction along the principal axis it departs from w~ and how it moves relative to the other code
vectors. For SSOM a similar result is obtained with Gin Eq. (5) simply being replaced by
GssoM, g;~oM hrt - ~. See (Graepel, 1997) for details.
=
4 NUMERICAL RESULTS
In the following we consider a binary symmetric channel (BSC) with a bit error rate (BER)
?. Assuming that the length of the code indices is n bits, the matrix elements of the transition matrix Hare
hrs = (1 _ c)n-dH(r,s) cdH(r,s) l
( 6)
where dH (r, s) is the Hamming-distance between the binary representations ofr and s.
4.1
TOY PROBLEM
The numerical analysis of the phase transitions described in the previous section was performed on a toy data set consisting of 2000 data vectors drawn from a two-dimensional
elongated Gaussian distribution P(x) = (211')- 1 ICI-~ exp(-~xTc- 1 x) with diagonal
covariance matrix C = diag(1, 0.04). The size of the codebook was N = 4 corresponding
ton = 2 bits. Figure 3 (left) shows the x-coordinates of the positions of the code vectors in
data space as functions of the inverse temperature {3. At a critical inverse temperature {3*
the code vectors split along the x-axis which is the principal axis of the distribution of data
points. In accordance with the eigenvector v~ax = (1, 0, 0, -1) T for the largest eigenvalue .\~ax of the matrix G two code vectors with Hamming distance dH
2 move to
opposite positions along the principal axis, and two remain at the center. Note the degeneracy of eigenvalues for matrix (6). Figure 3 (right) shows the critical inverse temperature /3*
as a function of the BER for both STVQ (crosses) and SSOM (dots). Results are in very
good agreement with the theoretical predictions of Eq. (5) (solid line). The inset displays
the average cost (E) = ~ Li Lr P(xi E r) Ls hrs \\xi- Ws\\ 2 as a function of f3 for
=
M Burger, T. Graepel and K. Obennayer
434
=
f
0.08 for STVQ and SSOM. The drop of the average cost occurs at the critical inverse
temperature {3"'.
4
3.5
..
X
.,
.
Dl
...
0.5
11
3
I!
I;
'rn2.5
..
?,
-0.5
...
-1
0
2
8
3
4
5
0.05
0.1
BER
0.15
0.2
0.25
Figure 3: Phase transitions in the 2 bit "toy" problem. (left) X-coordinate of code vectors
for the SSOM case plotted vs. inverse temperature {3, f = 0.08. The splitting of the four
code vectors occurs at {3 1.25 which is in very good accordance with the theory. (right)
Critical values of {3 for SSOM (dots) and STVQ (crosses), determined via the kink in the
0.08, top line STVQ), which indicates the phase transition. Solid
average cost (inset: f
lines denote theoretical predictions. Convergence parameter for the fixed-point iteration,
giving the upper limit for the difference in successive code vector positions per dimension,
was d = 5.0E- 10.
=
=
4.2
SOURCE CHANNEL CODING FOR IMAGE DATA
In order to demonstrate the applicability of STVQ and in particular of SSOM to source
channel coding we employed both algorithms to the compression of image data, which
were then sent via a noisy channel and decoded after transmission. As a training set we
used three 512 x 512 pixel 256 gray-value images from different scenes with blocksize
d = 2 x 2. The size of the codebook was chosen to beN = 16 in order to achieve a com2 f3t
pression to 1 bpp. We applied an exponential annealing schedule given by f3t+l
and determined the start value f3o to be just below the critical {3"' for the first split as given
in Eq. (5). Note that with the transition matrix as given in Eq. (6) this optimization corresponds to the embedding of an n = 4 dimensional hypercube in the d = 4 dimensional
data space. We tested the resulting codebooks by encoding our test image Lena1 (Figure 5),
which had not been used for determining the codebook, simulating the transmission of the
indices via a noisy binary symmetric channel with given bit error rate and reconstructing
the image using the codebook.
=
The results are summarized in Figure 4 which shows a plot of the signal-to-noise-ratio
(SNR) as a function of the bit-error rate for STVQ (dots), SSOM (vertical crosses), and
LBG (oblique crosses). STVQ shows the best performance especially for high BERs,
where it is naturally far superior to the LBG-algorithm which does not take into account
channel noise. SSOM, however, performs only slightly worse (approx. 1 dB) than STVQ.
Considering the fact that SSOM is computationally much less demanding than STVQ
1The
Lenna Story can be found at http://www.isr.com/ chuck/lennapgllenna.shtml
An Annealed Self-Organizing Map for Source Channel Coding
435
(O(N) for encoding)- due to the omission of the convolution with hrs in Eq. (4)- theresult demonstrates the efficiency of SSOM for source channel coding. Figure 4 also shows
the generalization behavior of a SSOM codebook optimized for a BER of 0.05 (rectan0.05 it performs worse than approprigles). Since this codebook was optimized fort:
ately trained SSOM codebooks for other values of BER, but still performs better than LBG
except for low values of BERs. At low values, SSOMs trained for the noisy case are outperformed by LBG because robustness w.r.t. channel noise is achieved at the expense of
an optimal data representation in the noise free case. Figure 5, finally, provides a vis..Ial
impression of the performance of the different vector quantizers at a BER of 0.033. While
the reconstruction for STVQ is only slightly better than the one for SSOM, both are clearly
superior to the reconstruction for LBG.
=
Figure 4: Comparison between different vector quantizers for image compression, noisy channel (BSC) transmission and reconstruction. The plot shows
the signal-to-noise-ratio (SNR), defined
as 10 loglo(O'signat/ O'noise). as a function of bit-error rate (BER) for STVQ z
and SSOM, each optimized for the given ~
channel noise, for SSOM, optimized for f
a BER of 0.05, and for LBG. The training set consisted of three 512 x 512 pixel
256 gray-value images with blocksize
2 x 2. The codebook size was N
d
16 corresponding to 1 bpp. The annealing schedule was given by f3t+I = 2 f3t
and Lena was used as a test image. Convergence parameter was 1. 0E - 5.
=
SSQoi5%8ER ?II-?
LOO-..
=
o
5
SlVQ++-
SSQ,I
14
-2
BER
CONCLUSION
We presented an algorithm for noisy vector quantization which is based on deterministic
annealing (STVQ). Phase transitions in the annealing process were analysed and a whole
class of vector quantizers could be derived, includings standard algorithms such as LBG
and "soft" versions as special cases of STVQ. In particular, a fuzzy version of Kohonen's
SOM was introduced, which is computationally more efficient than STVQ and still yields
very good results as demonstrated for noisy vector quantization of image data. The deterministic annealing scheme opens up many new possibilities for the usage of SOMs, in
particular, when its neighborhood function represents non-trivial neighborhood relations.
Acknowledgements This work was supported by TU Berlin (FIP 13/41). We thank
H. Bartsch for help and advice with regard to the image processing example.
References
J. M. Buhmann and T. Hofmann. Robust Vector Quantization by Competitive Learning.
Proceedings ofiCASSP'97, Munich, (1997).
N. Farvardin. A Study of Vector Quantization/or Noisy Channels. IEEE Transactions on
Infonnation Theory, vol. 36, p. 799-809 (1990).
436
M. Burger, T. Graepel and K. Obermayer
Original
STVQ
LBG
SNR9.00dB
SSOM
SNR 4.64 dB
SNR 7.80 dB
Figure 5: Lena transmitted over a binary symmetric channel with BER of 0.033 encoded
and reconstructed using different vector quantization algorithms.
? T. Graepel, M. Burger, and K. Obermayer. Phase Transitions in Stochastic Self-Organizing
Maps. Physical Review E, vol. 56, no. 4, p. 3876-3890 (1997).
T. Heskes and B. Kappen. Self-Organizing and Nonparametric Regression.
Neural Networks- ICANN'95, vol.l,p. 81-86 (1995).
Artificial
T. Kohonen. Self-Organizing Maps. Springer-Verlag, 1995.
S. P. Luttrell. Self-Organisation: A Derivationjromfirst Principles of a Class of Learning
Algorithms. Proceedings of IJCNN'89, Washington DC, vol. 2, p. 495-498 (1989).
S. P. Luttrell. A Baysian Analysis of Self-Organizing Maps. Neural Computation, vol. 6,
p. 767-794 (1994).
D. Miller and K. Rose. Combined Source-Channel Vector Quantization Using Deterministic Annealing. IEEE Transactions on Communications, vol. 42, p. 347-356 (1994).
K. Rose, E. Gurewitz, and G. C. Fox. Statistical Mechanics and Phase Transitions in
Clustering. Physical Review Letters, vol. 65, No.8, p. 945-948 (1990).
| 1443 |@word version:4 compression:3 open:2 covariance:3 solid:2 kappen:1 existing:1 recovered:1 com:1 analysed:1 assigning:1 numerical:2 hofmann:1 enables:1 drop:1 plot:2 v:1 ial:1 oblique:1 lr:2 quantizer:1 provides:1 codebook:10 successive:1 along:5 behavior:2 mechanic:1 lena:2 considering:1 increasing:1 becomes:1 burger:6 mass:2 lixi:1 interpreted:1 eigenvector:3 fuzzy:1 every:2 expands:1 demonstrates:1 lila:1 accordance:2 limit:2 encoding:5 solely:1 averaged:1 jlxi:1 ofe:1 convenience:1 applying:1 www:1 elongated:1 map:12 deterministic:9 lagrangian:1 center:3 annealed:4 demonstrated:1 starting:3 l:2 convex:1 splitting:2 assigns:1 embedding:1 coordinate:2 exact:1 agreement:1 element:2 located:1 cooling:1 labeled:1 signat:1 com2:1 calculate:2 rose:5 broken:1 trained:2 solving:1 efficiency:1 basis:1 easily:1 derivation:1 artificial:1 oby:1 sc:2 klaus:1 neighborhood:5 whose:1 encoded:1 distortion:3 compressed:1 encoder:1 topographic:4 analyse:2 noisy:15 jointly:1 eigenvalue:5 matthias:1 reconstruction:3 luttrell:5 fr:1 tu:2 kohonen:7 organizing:12 achieve:1 decod:1 kink:1 convergence:2 transmission:6 a11:1 ben:1 help:1 derive:2 develop:2 oo:1 bers:2 measured:1 expectationmaximization:1 received:2 eq:15 hrt:1 c:1 hst:1 indicate:1 direction:1 stochastic:1 generalization:1 hut:1 around:1 exp:3 purpose:1 outperformed:1 label:1 infonnation:1 largest:3 bsc:2 clearly:1 gaussian:1 loglo:1 shtml:1 encode:1 derived:3 focus:1 ax:8 likelihood:1 indicates:1 contrast:1 ately:1 w:1 relation:1 germany:1 pixel:2 among:2 special:1 equal:1 f3:7 washington:1 cartoon:1 represents:2 minimized:1 grt:1 phase:9 replaced:1 consisting:1 pression:1 possibility:2 fox:1 euclidean:2 old:1 desired:1 plotted:1 theoretical:2 increased:1 soft:8 assignment:8 maximization:1 cost:9 applicability:1 f3t:4 snr:4 front:1 loo:1 combined:1 probabilistic:1 together:1 squared:2 central:1 hn:1 worse:2 li:2 toy:3 account:2 de:1 rn2:1 coding:9 summarized:1 caused:1 vi:2 performed:1 start:1 competitive:1 minimize:1 om:1 variance:1 miller:2 yield:2 generalize:1 lu:1 energy:1 hare:1 naturally:2 associated:2 mi:1 hamming:2 degeneracy:1 bpp:2 graepel:6 schedule:3 just:1 mode:1 gray:2 lossy:2 thore:1 usage:2 consisted:1 multiplier:1 hence:2 assigned:2 symmetric:4 during:3 self:14 xtc:1 impression:1 demonstrate:2 performs:4 temperature:10 image:13 superior:3 physical:2 extend:1 rth:1 gibbs:1 approx:1 heskes:2 had:1 dot:3 quantisation:1 optimizing:1 verlag:1 binary:6 chuck:1 ofr:1 transmitted:4 minimum:1 employed:1 signal:2 ii:2 technical:1 characterized:1 offer:1 cross:4 f3o:1 j3:1 variant:1 prediction:2 regression:1 sition:1 expectation:1 iteration:2 achieved:1 annealing:17 source:9 induced:1 sent:2 db:4 cdh:1 incorporates:1 call:2 split:5 bandwidth:1 opposite:1 codebooks:2 idea:2 speech:1 eigenvectors:1 nonparametric:1 iixi:1 http:1 wr:7 track:1 per:1 vol:7 group:1 four:1 drawn:1 inverse:7 letter:1 franklinstr:1 family:1 bit:7 display:1 ijcnn:1 constraint:1 scene:1 department:1 munich:1 according:1 remain:1 slightly:2 em:2 reconstructing:1 wi:3 computationally:5 legal:1 bartsch:1 drs:1 apply:1 generic:1 simulating:1 robustness:1 original:2 top:1 denotes:1 clustering:3 cf:1 marginalized:1 giving:1 especially:2 hypercube:1 objective:1 move:2 occurs:2 traditional:1 diagonal:2 obermayer:4 gin:1 distance:4 thank:1 berlin:4 decoder:1 unstable:1 trivial:2 assuming:2 code:28 length:1 modeled:1 index:5 providing:1 ratio:2 quantizers:5 expense:1 obennayer:1 upper:1 vertical:1 convolution:1 wtll:2 communication:3 dc:1 arbitrary:2 omission:1 introduced:2 fort:1 required:1 optimized:4 baysian:1 below:1 critical:7 demanding:4 buhmann:2 hr:14 solvable:1 scheme:7 axis:5 gurewitz:1 text:1 review:2 acknowledgement:1 determining:1 relative:1 permutation:1 principle:3 story:1 supported:1 free:1 blocksize:2 ber:10 isr:1 regard:1 dimension:1 transition:16 stand:1 ici:1 far:1 transaction:2 reconstructed:1 fip:1 global:1 xi:19 additionally:1 channel:29 robust:3 expanding:1 symmetry:1 som:10 diag:1 icann:1 whole:3 noise:15 advice:1 representative:2 fashion:1 vr:3 position:3 decoded:2 lbg:12 comprises:1 exponential:1 hsr:1 departs:1 inset:2 er:2 ton:1 organisation:1 grouping:1 dl:1 quantization:16 suited:1 entropy:2 lt:1 simply:1 likely:1 springer:1 corresponds:2 determines:4 dh:3 llxi:1 fuzziness:1 change:3 hard:1 infinite:2 determined:2 except:1 averaging:1 principal:4 lenna:1 total:1 la:1 tested:1 |
487 | 1,444 | Regression with Input-dependent Noise:
A Gaussian Process Treatment
Paul W. Goldberg
Department of Computer Science
University of Warwick
Coventry, CV 4 7AL, UK
pvgGdcs.varvick.ac.uk
Christopher K.I. Williams
Neural Computing Research Group
Aston University
Birmingham B4 7ET, UK
c.k.i.villiamsGaston.ac.uk
Christopher M. Bishop
Microsoft Research
St. George House
1 Guildhall Street
Cambridge, CB2 3NH, UK
cmbishopOmicrosoft.com
Abstract
Gaussian processes provide natural non-parametric prior distributions over regression functions. In this paper we consider regression
problems where there is noise on the output, and the variance of
the noise depends on the inputs. If we assume that the noise is
a smooth function of the inputs, then it is natural to model the
noise variance using a second Gaussian process, in addition to the
Gaussian process governing the noise-free output value. We show
that prior uncertainty about the parameters controlling both processes can be handled and that the posterior distribution of the
noise rate can be sampled from using Markov chain Monte Carlo
methods. Our results on a synthetic data set give a posterior noise
variance that well-approximates the true variance.
1
Background and Motivation
A very natural approach to regression problems is to place a prior on the kinds of
function that we expect, and then after observing the data to obtain a posterior.
The prior can be obtained by placing prior distributions on the weights in a neural
494
P. W Goldberg, C. K. L Williams and C. M. Bishop
network, although we would argue that it is perhaps more natural to place priors directly over functions. One tractable way of doing this is to create a Gaussian process
prior. This has the advantage that predictions can be made from the posterior using
only matrix multiplication for fixed hyperparameters and a global noise level. In
contrast, for neural networks (with fixed hyperparameters and a global noise level)
it is necessary to use approximations or Markov chain Monte Carlo (MCMC) methods. Rasmussen (1996) has demonstrated that predictions obtained with Gaussian
processes are as good as or better than other state-of-the art predictors.
In much of the work on regression problems in the statistical and neural networks
literatures, it is assumed that there is a global noise level, independent of the input
vector x. The book by Bishop (1995) and the papers by Bishop (1994), MacKay
(1995) and Bishop and Qazaz (1997) have examined the case of input-dependent
noise for parametric models such as neural networks. (Such models are said to
heteroscedastic in the statistics literature.) In this paper we develop the treatment
of an input-dependent noise model for Gaussian process regression, where the noise
is assumed to be Gaussian but its variance depends on x. As the noise level is nonnegative we place a Gaussian process prior on the log noise level. Thus there are
two Gaussian processes involved in making predictions: the usual Gaussian process
for predicting the function values (the y-process), and another one (the z-process)
for predicting the log noise level. Below we present a Markov chain Monte Carlo
method for carrying out inference with this model and demonstrate its performance
on a test problem.
1.1
Gaussian processes
A stochastic process is a collection of random variables {Y(x)lx E X} indexed by
a set X. Often X will be a space such as 'R,d for some dimension d, although it
could be more general. The stochastic process is specified by giving the probability
distribution for every finite subset of variables Y(Xl), ... , Y(Xk) in a consistent
manner. A Gaussian process is a stochastic process which can be fully specified
by its mean function J.L(x) = E[Y(x)] and its covariance function Cp(x,x') =
E[(Y(x)-J.L(x?)(Y(x')-J.L(x'?]; any finite set of points will have a joint multivariate
Gaussian distribution. Below we consider Gaussian processes which have J.L(x) == O.
This assumes that any known offset or trend in the data has been. removed. A
non-zero I' (x ) is easily incorporated into the framework at the expense of extra
notational complexity.
A covariance junction is used to define a Gaussian process; it is a parametrised
function from pairs of x-values to their covariance. The form of the covariance
function that we shall use for the prior over functions is given by
Cy(x(i),x U? =vyexp
(-~ tWYl(x~i) _x~j?2) + J y 8(i,j)
(1)
1=1
where vy specifies the overall y-scale and W;:/2 is the length-scale associated with
the lth coordinate. Jy is a "jitter" term (as discussed by Neal, 1997), which is
added to prevent ill-conditioning of the covariance matrix of the outputs. J y is a
typically given a small value, e.g. 10- 6 .
For the prediction problem we are given n data points
1)
= ((Xl,t1),(X2,t2),
Input-dependent Noise: A Gaussian Process Treatment
495
... , (x n , t n ?), where ti is the observed output value at Xi. The t's are assumed
to have been generated from the true y-values by adding independent Gaussian
noise whose variance is x-dependent. Let the noise variance at the n data points
be r = (r(xl),r(x2), ... ,r(x n )). Given the assumption of a Gaussian process prior
over functions, it is a standard result (e.g. Whittle, 1963) that the predictive distribution P(t*lx*) corresponding to a new input x* is t* "'" N(t(X*),0'2(X*)), where
i(x*)
0'2(X*)
-
k~(x*)(Ky
+ KN)-lt
Cy(x*, x*)
+ r(x*)
- k~(x*)(Ky
(2)
+ KN )-lky(x*)
(3)
where the noise-free covariance matrix K y satisfies [KY] ij = C y (x i, X j ), and
ky(x*) = (Cy(x*,xd, ... ,Cy(x*,xn?T, KN = diag(r) and t = (tb ... ,tn)T,
and V0'2(X*) gives the "error bars" or confidence interval of the prediction.
In this paper we do not specify a functional form for the noise level r(x) but we do
place a prior over it. An independent Gaussian process (the z-process) is defined to
be the log of the noise level. Its values at the training data points are denoted by
z = (zl, . .. ,zn),sothatr = (exp(zl), ... ,exp(zn?. The priorforz has a covariance
function CZ(X(i), xU? similar to that given in equation 1, although the parameters
vz and the WZI'S can be chosen to be different to those for the y-process. We also
add the jitter term J z t5(i,j) to the covariance function for Z, where Jz is given the
value 10- 2 ? This value is larger than usual, for technical reasons discussed later.
We use a zero-mean process for z which carries a prior assumption that the average
noise rate is approximately 1 (being e to the power of components of z). This is
suitable for the experiment described in section 3. In general it is easy to add an
offset to the z-process to shift the prior noise rate.
2
An input-dependent noise process
We discuss, in turn, sampling the noise rates and making predictions with fixed values of the parameters that control both processes, and sampling from the posterior
on these parameters.
2.1
Sampling the Noise Rates
The predictive distribution for t*, the output at a point x*, is P(t*lt) =
f P(t*lt,r(z?P(zlt)dz. Given a z vector, the prediction P(t*lt,r(z? is Gaussian with mean and variance given by equations 2 and 3, but P(zlt) is difficult to
handle analytically, so we use a Monte Carlo approximation to the integral. Given
a representative sample {Zb ... ' Zk} of log noise rate vectors we can approximate
the integral by the sum E j P(t*lt,r(zj?.
i
We wish to sample from the distribution P(zlt). As this is quite difficult, we sample
instead from P(y, zit); a sample for P(zlt) can then be obtained by ignoring the
y values. This is a similar approach to that taken by Neal (1997) in the case of
Gaussian processes used for classification or robust regression with t-distributed
noise. We find that
(4)
P(y, zit) oc P(tly, r(z?P(y)P(z).
We use Gibbs sampling to sample from P(y, zit) by alternately sampling from
P(zly, t) and P(ylz, t). Intuitively were are alternating the "fitting" of the curve (or
P. W. Goldberg, C. K. 1. Williams and C. M Bishop
496
y-process) with "fitting" the noise level (z-process) . These two steps are discussed
in turn .
? Sampling from P(ylt, z)
For y we have that
P(ylt, z) ex P(tly, r(z?P(y)
(5)
where
P(tly, r(z?
=
TI
n
1
(
(21l'Ti)l/2 exp -
(ti - Yi)2 )
2Ti
.
(6)
Equation (6) can also be written as P(tly,r(z? '" N(t,KN) ' Thus P(ylt,z) is
a multivariate Gaussian with mean (Kyl + Ki/)-l K;/t and covariance matrix
(Kyl + KN1)-1 which can be sampled by standard methods .
? Sampling from P(zlt,y)
For fixed y and t we obtain
P(zly, t) ex P(tly, z)P(z).
(7)
The form of equation 6 means that it is not easy to sample z as a vector. Instead
we can sample its components separately, which is a standard Gibbs sampling algorithm. Let Zi denote the ith component of z and let Z-i denote the remaining
components. Then
(8)
P(Zilz-i) is the distribution of Zi conditioned on the values of Z-i'
The computation of P(zilz-i) is very similar to that described by equations (2) and (3),
except that Cy ( " .) is replaced by C z ( " .) and there is no noise so that T (.) will be
identically zero.
We sample from P(zilz-i' y, t) using rejection sampling. We first sample from
P(zilz-i), and then reject according to the term exp{ -Zi/2 - Hti - Yi)2 exp( -Zi)}
(the likelihood of local noise rate Zi), which can be rescaled to have a maximum
value of lover Zi. Note that it is not necessary to perform a separate matrix
inversion for each i when computing the P(zilz-i) terms; the required matrices
can be computed efficiently from the inverse of K z. We find that the average
rejection rate is approximately two-thirds, which makes the method we currently use
reasonably efficient. Note that it is also possible to incorporate the term exp( -Zi/2)
from the likelihood into the mean of the Gaussian P(zilz-i) to reduce the rejection
rate.
As an alternative approach, it is possible to carry out Gibbs sampling for P(zilz-i' t)
without explicitly representing y, using the fact that 10gP(tlz) = -~logIKI?
T K-1t + canst, where K = K y + K N . We have implemented this and found
similar results to those obtained using sampling of the y's. However, explicitly
representing the y-process is useful when adapting the parameters, as described in
section 2.3.
!t
Input-dependent Noise: A Gaussian Process Treatment
2.2
497
Making predictions
So far we have explained how to obtain a sample from P(zlt). To make predictions
we use
P(t*lt)
~ ~ l: P(t*lt, r(zj)).
(9)
j
However, P(t*lt,r(zj)) is not immediately available, as z*, the noise level at x* is
unknown. In fact
(10)
P(z*IZj, t) is simply a Gaussian distribution for z* conditioned on Zj, and is obtained in a similar way to P(zilz-i). As P(t*lz*, t, r(zj)) is a Gaussian distribution
as given by equations (2) and (3), P(t*\t, r(z j)) is an infinite mixture of Gaussians
with weights P(z*IZj) . Note, however, that each ofthese components has the same
mean i(x*) as given by equation (2), but a different variance.
We approximate P(t*lt, r(zj)) by taking s = 10 samples of P(z*lzj) and thus obtain
a mixture of s Gaussians as the approximating distribution. The approximation for
P(t*lt) is then obtained by averaging these s-component mixtures over the k samples
Z1> ??? , Zk to obtain an sk-component mixture of Gaussians.
2.3
Adapting the parameters
Above we have described how to obtain a sample from the posterior distribution
P(z\t) and to use this to make predictions, based on the assumption that the
parameters Oy (Le. Vy,Jy,WYl, . .. ,WYd) and Oz (Le. vz,JZ,WZl, ... ,WZd) have
been set to the correct values. In practice we are unlikely to know what these
settings should be, and so introduce a hierarchical model, as shown in Figure l.
This graphical model shows that the joint probability distribution decomposes as
P(Oy,OZ, y, z, t) = P(Oy)P(Oz)P(yIOy )P(z\Oz)P(t\y, z).
Our goal now becomes to obtain a sample from the posterior P(Oy,Oz,y,zlt),
which can be used for making predictions as before. (Again, the y samples are
not needed for making predictions, but they will turn out to be useful for sampling
Oy .) Sampling from the joint posterior can be achieved by interleaving updates of
Oy and Oz with y and Z updates. Gibbs sampling for Oy and Oz is not feasible
as these parameters are buried deeply in the K y and K N matrices, so we use the
Metropolis algorithm for their updates. As usual, we consider moving from our
current state 0 = (Oy,Oz) to a new state 0 using a proposal distribution J(O,O). In
practice we take J to be an isotropic Gaussian centered on 0?. Denote the ratio of
P(Oy)P(Oz)P(yIOy)P(z\Oz) in states 9 and 0 by r. Then the proposed state 0 is
accepted with probability min{r, 1}.
It would also be possible to use more sophisticated MCMC algorithms such as the
Hybrid Monte Carlo algorithm which uses derivative information, as discussed in
Neal (1997).
3
Results
We have tested the method on a one-dimensional synthetic problem. 60 data points
P. W. Goldberg, C. K. l Williams and C. M Bishop
498
Figure 1: The hierarchical model including parameters.
were generated from the function y = 2 sin(271"x) on [0, 1] by adding independent
Gaussian noise. This noise has a standard deviation that increases linearly from 0.5
at x 0 to 1.5 at x
1. The function and the training data set are illustrated in
Figure 2(a).
=
=
As the parameters are non-negative quantities, we actually compute with their log
values. logvy, logvz, logwy and log Wz were given N(O, 1) prior distributions. The
jitter values were fixed at J y = 10- 6 and J z = 10- 2 ? The relatively large value
for J z assists the convergence of the Gibbs sampling, since it is responsible for
most of the variance of the conditional distribution P(Zi/Z-i}. The broadening of
this distribution leads to samples whose likelihoods are more variable, allowing the
likelihood term (used for rejection) to be more influential.
In our simulations, on each iteration we made three Metropolis updates for the
parameters, along with sampling from all of the y and Z variables. The Metropolis
proposal distribution was an isotropic Gaussian with variance 0.01. We ran for
3000 iterations, and discarded the first one-third of iterations as "burn-in", after
which plots of each of the parameters seemed to have settled down. The parameters
and Z values were stored every 100 iterations. In Figure 2(b) the average standard
deviation of the inferred noise has been plotted, along with with two standard
deviation error-bars. Notice how the standard deviation increases from left to right,
in close agreement with the data generator.
Studying the posterior distributions of the parameters, we find that the ylength scale Ay d;j (wy) -1/2 is well localized around 0.22 ? 0.1, in good agreement with the wavelength of the sinusoidal generator. (For the covariance function
in equation 1, the expected number of zero crossings per unit length is 1/7I"Ay.)
(WZ)-1/2 is less tightly constrained, which makes sense as it corresponds to a longer
wavelength process, and with only a short segment of data available there is still
considerable posterior uncertainty.
4
Conclusions
We have introduced a natural non-parametric prior on variable noise rates, and
given an effective method of sampling the posterior distribution, using a MCMC
499
Input-dependent Noise: A Gaussian Process Treatment
.'
..
"
.'
..
-2
o.
?O~~
O I~~
0 2--0~'~O~'~
O ~'~"--O~7~O~'~
"~
(a)
(b)
Figure 2: (a) shows the training set (crosses); the solid line depicts the x-dependent
mean of the output. (b) The solid curve shows the average standard deviation of
the noise process, with two standard deviation error bars plotted as dashed lines.
The dotted line indicates the true standard deviation of the data generator.
method. When applied to a data set with varying noise, the posterior noise rates
obtained are well-matched to the known structure. We are currently experimenting
with the method on some more challenging real-world problems.
Acknowledgements
This work was carried out at Aston University under EPSRC Grant Ref. GR/K 51792
Validation and Verification of Neural Network Systems.
References
[1] C.M. Bishop (1994). Mixture Density Networks. Technical report NCRG/94/001, Neural Computing Research Group, Aston University, Birmingham, UK.
[2] C.M. Bishop (1995). Neural Networks for Pattern Recognition. Oxford University Press.
[3] C.M. Bishop and C. Qazaz (1997). Regression with Input-dependent Noise: A Bayesian
Treatment. In M. C. Mozer, M. I. Jordan and T. Petsche (Eds) Advances in Neural
Information Processing Systems 9 Cambridge MA MIT Press.
[4] D. J. C. MacKay (1995). Probabilistic networks: new models and new methods. In
F. Fogelman-Soulie and P. Gallinari (Eds), Proceedings ICANN'95 International Conference on Neural Networks, pp. 331-337. Paris, EC2 & Cie.
[5] R. Neal (1997). Monte Carlo Implementation of Gaussian Process Models for Bayesian
Regression and Classification. Technical Report 9702, Department of Statistics, University of Toronto. Available from http://vvv.cs.toronto.edurradford/.
[6] C.E. Rasmussen (1996). Evaluation of Gaussian Processes and Other Methods for Nonlinear Regression. PhD thesis, Department of Computer Science, University of Toronto.
Available from http://vwv . cs .utoronto. carcarl/ .
[7] C.K.I. Williams and C.E. Rasmussen (1996). Gaussian Processes for Regression. In
D. S. Touretzky, M. C. Mozer and M. E . Hasselmo Advances in Neural Information
Processing Systems 8 pp. 514-520, Cambridge MA MIT Press.
[8] P. Whittle (1963). Prediction and regulation by linear least-square methods. English
Universities Press.
| 1444 |@word inversion:1 simulation:1 covariance:10 solid:2 carry:2 current:1 com:1 written:1 plot:1 update:4 lky:1 xk:1 isotropic:2 ith:1 short:1 toronto:3 lx:2 along:2 vwv:1 fitting:2 introduce:1 manner:1 expected:1 wzd:1 becomes:1 matched:1 what:1 kind:1 every:2 ti:5 xd:1 uk:6 zl:2 control:1 unit:1 grant:1 gallinari:1 t1:1 before:1 local:1 oxford:1 approximately:2 burn:1 examined:1 challenging:1 heteroscedastic:1 zilz:8 responsible:1 practice:2 cb2:1 reject:1 adapting:2 confidence:1 close:1 demonstrated:1 dz:1 williams:5 immediately:1 handle:1 coordinate:1 controlling:1 goldberg:4 us:1 agreement:2 trend:1 crossing:1 recognition:1 observed:1 epsrc:1 cy:5 removed:1 rescaled:1 deeply:1 ran:1 mozer:2 complexity:1 carrying:1 segment:1 predictive:2 easily:1 joint:3 effective:1 monte:6 qazaz:2 whose:2 quite:1 larger:1 warwick:1 statistic:2 gp:1 advantage:1 wzi:1 oz:10 ky:4 convergence:1 develop:1 ac:2 ij:1 zit:3 implemented:1 c:2 correct:1 stochastic:3 centered:1 tlz:1 around:1 guildhall:1 exp:6 birmingham:2 currently:2 vz:2 hasselmo:1 create:1 mit:2 gaussian:34 varying:1 canst:1 notational:1 likelihood:4 indicates:1 experimenting:1 contrast:1 ylt:3 sense:1 inference:1 dependent:10 typically:1 unlikely:1 buried:1 kn1:1 fogelman:1 overall:1 classification:2 ill:1 denoted:1 art:1 constrained:1 mackay:2 sampling:17 placing:1 t2:1 report:2 tightly:1 replaced:1 microsoft:1 izj:2 evaluation:1 mixture:5 parametrised:1 chain:3 integral:2 necessary:2 indexed:1 plotted:2 zn:2 deviation:7 subset:1 predictor:1 gr:1 stored:1 kn:4 lzj:1 synthetic:2 st:1 density:1 international:1 ec2:1 probabilistic:1 again:1 thesis:1 settled:1 book:1 derivative:1 sinusoidal:1 whittle:2 explicitly:2 depends:2 later:1 observing:1 doing:1 square:1 variance:11 efficiently:1 bayesian:2 carlo:6 touretzky:1 ed:2 pp:2 involved:1 associated:1 sampled:2 treatment:6 sophisticated:1 actually:1 specify:1 governing:1 christopher:2 nonlinear:1 perhaps:1 true:3 analytically:1 alternating:1 neal:4 illustrated:1 sin:1 oc:1 ay:2 demonstrate:1 tn:1 cp:1 functional:1 b4:1 conditioning:1 nh:1 ncrg:1 discussed:4 approximates:1 cambridge:3 gibbs:5 cv:1 moving:1 longer:1 v0:1 add:2 posterior:12 multivariate:2 yi:2 kyl:2 george:1 dashed:1 smooth:1 technical:3 cross:1 jy:2 prediction:13 regression:11 iteration:4 cie:1 cz:1 achieved:1 proposal:2 addition:1 background:1 separately:1 interval:1 extra:1 zlt:7 vvv:1 lover:1 jordan:1 easy:2 identically:1 zi:8 reduce:1 shift:1 handled:1 assist:1 useful:2 wzl:1 http:2 specifies:1 vy:2 zj:6 notice:1 dotted:1 per:1 shall:1 group:2 prevent:1 sum:1 tly:5 ylz:1 inverse:1 uncertainty:2 jitter:3 place:4 ki:1 nonnegative:1 x2:2 min:1 relatively:1 department:3 influential:1 according:1 metropolis:3 ofthese:1 making:5 intuitively:1 explained:1 taken:1 equation:8 discus:1 turn:3 needed:1 know:1 tractable:1 studying:1 junction:1 available:4 gaussians:3 hierarchical:2 petsche:1 alternative:1 assumes:1 remaining:1 graphical:1 giving:1 approximating:1 added:1 quantity:1 parametric:3 usual:3 said:1 separate:1 street:1 argue:1 reason:1 length:2 ratio:1 difficult:2 regulation:1 expense:1 negative:1 implementation:1 unknown:1 perform:1 allowing:1 markov:3 discarded:1 finite:2 incorporated:1 inferred:1 introduced:1 pair:1 required:1 specified:2 paris:1 z1:1 alternately:1 bar:3 below:2 wy:1 pattern:1 tb:1 including:1 wz:2 power:1 suitable:1 natural:5 hybrid:1 predicting:2 representing:2 aston:3 carried:1 prior:15 literature:2 acknowledgement:1 multiplication:1 fully:1 expect:1 oy:9 localized:1 generator:3 validation:1 verification:1 consistent:1 zly:2 free:2 rasmussen:3 english:1 taking:1 distributed:1 curve:2 dimension:1 xn:1 world:1 soulie:1 t5:1 seemed:1 made:2 collection:1 lz:1 far:1 approximate:2 global:3 assumed:3 xi:1 sk:1 decomposes:1 jz:2 zk:2 robust:1 reasonably:1 ignoring:1 broadening:1 diag:1 icann:1 linearly:1 motivation:1 noise:44 paul:1 hyperparameters:2 ref:1 xu:1 representative:1 depicts:1 wish:1 xl:3 house:1 third:2 hti:1 interleaving:1 down:1 bishop:10 utoronto:1 offset:2 adding:2 phd:1 conditioned:2 rejection:4 lt:10 simply:1 wavelength:2 corresponds:1 satisfies:1 ma:2 conditional:1 lth:1 goal:1 feasible:1 considerable:1 infinite:1 except:1 averaging:1 zb:1 accepted:1 incorporate:1 mcmc:3 tested:1 ex:2 |
488 | 1,445 | Structure Driven Image Database
Retrieval
Jeremy S. De Bonet &, Paul Viola
Artificial Intelligence Laboratory
Learning & Vision Group
545 Technology Square Massachusetts Institute of Technology
. Cambridge, MA 02139
EMAIL: jsdCOaLmit. edu & violaCOaLmit. edu
HOMEPAGE: http://www.ai . mit. edu/pro j ects/l v
Abstract
A new algorithm is presented which approximates the perceived
visual similarity between images. The images are initially transformed into a feature space which captures visual structure, texture and color using a tree of filters. Similarity is the inverse of
the distance in this perceptual feature space. Using this algorithm
we have constructed an image database system which can perform
example based retrieval on large image databases. Using carefully
constructed target sets, which limit variation to only a single visual
characteristic, retrieval rates are quantitatively compared to those
of standard methods.
1
Introduction
Without supplementary information, there exists no way to directly measure the
similarity between the content of images. In general, one cannot answer a question of
the form: "is image A more like image B or image C?" without defining the criteria
by which this comparison is to be made. People perform such tasks by inferring
some criterion, based on their visual experience or by complex reasoning about the
situations depicted in the images. Humans are very capable database searchers.
They can perform simple searches like, "find me images of cars", or more complex
or loosely defined searches like, "find me images that depict pride in America".
In either case one must examine all, or a large portion, of the database. As the
prevalence and size of multimedia databases increases, automated techniques will
become critical in the successful retrieval of relevant information. Such techniques
must be able to measure the similarity between the visual content of natural images.
Structure Driven Image Database Retrieval
867
Many algorithms have been proposed for image database retrieval. For the most
part these techniques compute a feature vector from an image which is made up of
a handful of image measurements. Visual or semantic distance is then equated with
feature distance. Examples include color histograms, texture histograms, shape
boundary descriptors, eigenimages, and hybrid schemes (QBIC, ; Niblack et aI.,
1993; Virage, ; Kelly, Cannon and Hush, 1995; Pentland, Picard and ScIaroff, 1995;
Picard and Kabir, 1993; Santini and Jain, 1996). A query to such a system typically
consists of specifying two types of parameters: the target values of each of the
measurements; and a set of weights, which determine the relative importance of
deviations from the target in each measurement dimension. The features used by
these systems each capture some very general property of images. As a result of
their lack of specificity however, many images which are actually very different in
content generate the same feature responses.
In contrast our approach extracts thousands of very specific features. These features
measure both local texture and global structure. The feature extraction algorithm
computes color, edge orientation, and other local properties at many resolutions.
This sort of multi-scale feature analysis is of critical importance in visual recognition and has been used successfully in the context of object recognition (von der
Malsburg, 1988; Rao and Ballard, 1995; Viola, 1996)
Our system differs from others because it detects not only first order relationships,
such as edges or color, but also measures how these first order relationships are
related to one another. Thus by finding patterns between image regions with particular local properties, more complex - and therefore more discriminating - features can be extracted. This type of repeated, non-linear, feature detection bears
a strong resemblance to the response properties of visual cortex cells (Desimone
et al., 1984). While the mechanism for the responses of these cells is not yet clear,
this work supports the conclusion that this type of representation is very useful in
practical visual processing.
2
Computing the Characteristic Signature
The "texture-of-texture" measurements are based on the outputs of a tree of nonlinear filtering operations. Each path through the tree creates a particular filter
network, which responds to certain structural organization in the image. Measuring the appropriately weighted difference between the signatures of images in the
database and the set of query-images, produces a similarity measure which can be
used to rank and sort the images in the database.
The computation of the characteristic signature is straightforward. At the highest
level of resolution the image is convolved with a set of 25 local linear features
including oriented edges and bars. The results of these convolutions are 25 feature
response images. These images are then rectified by squaring, which extracts the
texture energy in the image, and then downsampled by a factor of two. At this
point there are 25 half scale output images which each measure a local textural
property of the input image. For example one image is sensitive to vertical edges,
and responds strongly to both skyscrapers and picket fences.
Convolution, rectification and downsampling is then repeated on each of these 25
half resolution images producing 625 quarter scale "texture-of-texture" images. The
second layer will respond strongly to regions where the texture specified in the first
layer has a particular spatial arrangement. For example if horizontal alignments
of vertical texture are detected, there will be a strong response to a picket fence
and little response to a skyscraper. With additional layers additional specificity is
868
J. S. D. Bonet and P. A. Viola
achieved; repeating this procedure a third time yields 15,625 meta-texture images
at eighth scale.
Each of the resulting meta-textures is then summed to compute a single value and
provides one element in the characteristic signature. When three channels of color
are included there are a total of 46,875 elements in the characteristic signature. Once
computed, the signature elements are normalized to reduce the effects of contrast
changes.
More formally the characteristic signature of an image is given by:
Si,j,k,e(I) =
L
(1)
Ei,j,k(Ie)
pixels
where I is the image, i, j and k index over the different types of linear filters, and
Ie are the different color channels of the image. The definition of E is:
Ei(I)
Ei,j (I)
Ei,j,k(I)
2 .j.. [(Fi 0 1)2]
2.j.. [(Fj 0 Ei(I))2]
(2)
(3)
2.j.. [(Fk 0 E i ,j(I))2] .
(4)
where F j is the ith filter and 2.j.. is the downsampling operation.
3
U sing Characteristic Signatures To Form Image Queries
In the "query by image" paradigm, we describe similarity in terms of the difference
between an image and a group of example query images. This is done by comparing the characteristic signature of the image to the mean signature of the query
images. The relative importance of each element of the characteristic signature in
determining similarity is proportional to the inverse variance of that element across
the example-image group:
L= -
L L L L [S;,j,k,O( Iq) - S;,j,k,o (I.".)
i
j
k
e
Var[Si,j,k,e(lq )]
r
(5)
where Si,j,k,e(Iq ) and Var [Si,j,k,c(Iq )] are the mean and variance of the characteristic signatures computed over the query set. This is a diagonal approximation of the
Mahalanobis distance (Duda and Hart, 1973). It has the effect of normalizing the
vector-space defined by the characteristic signatures, so that characteristic elements
which are salient within the group of query images contribute more to the overall
similarity of an image.
In Figure 1 three 2D projections of these 46,875 dimensional characteristic signature
space are shown. The data points marked with circles are generated by the 10 images
shown at the top of Figure 3. The remaining points are generated by 2900 distract or
images. Comparing (a) and (b) we see that in some projections the images cluster
tightly, while in others they are distributed. Given a sample of images from the
target set we can observe the variation in each possible projection axis. Most of the
time the axes shown in (a) will be strongly discounted by the algorithm because
these features are not consistent across the query set. Similarly the axes from (b)
will receive a large weight because the target images have very consistent values.
The axes along which target groups cluster, however, differ from target group to
target group. As a result it is not possible to conclude that the axes in (b) are
simply better than the axes in (a). In Figure 1 (c) the same projection is shown
again this time with a different target set highlighted (with asterisks).
Structure Driven Image Database Retrieval
869
9
~i~~:;:,~?~~~~'?:.~ .'
.' ~~~" 9.:'.):
t'K.
o
o
(a)
(b)
(c)
Figure 1: In some projections target groups do not cluster (a), and they do in
others (b). However, different target groups will not necessarily cluster in the same
projections (c).
4
Experiments
In the first set of experiments we use a database of 2900 images from 29 Corel Photo
CD (collections 1000-2900.) Figure 2 shows the results of typical user query on this
system. The top windows in each Figure contain the query-images submitted by
the user. The bottom windows show the thirty images found to be most similar;
similarity decreases from upper left (most similar) to lower right.
Though these examples provide an anecdotal indication that the system is generating similarity measures which roughly conform to human perception, it is difficult
to fully characterize the performance of this image retrieval technique. This is a
fundamental problem of the domain. Images vary from each other in an astronomical number of ways, and similarity is perceived by human observers based upon
complex interactions between recognition, cognition, and assumption. It seems unlikely that an absolute criterion for image similarity can ever be determined, if one
truly exists. However using sets of images which we believe are visually similar, we
can establish a basis for comparing algorithms.
To better measure the performance of the system we added a set of 10 images to the
2900 image database and attempted to retrieve these new images. We compare the
performance of the present system to ten other techniques. Though these techniques
are not as sophisticated as those used in systems developed by other researchers,
they are indicative of the types of methods which are prevalent in the literature.
In each experiment we measure the retrieval rates for a set of ten target images
which we believe to be visually similar because they consist of images of a single
scene. Images in the target set differ due to variation of a single visual characteristic.
In some of the target sets the photographic conditions have been changed, either by
moving the camera, the objects or the light. In other target sets post photograph
image manipulation has been performed. Two example target sets are shown in
Figure 3.
In each experiment we perform 45 database queries using every possible pair of
images from the target set. The retrieval methods compared are: ToT The current
textures-of-textures system; RGB-216(or 512)C R,G,B color histograms using
216 (or 512) bins by dividing each color dimension into 6 (or 8) regions. The
target histogram is generated by combining the histograms from the two model images; HSV-216(or 512)C same using H,S,V color space; RGB(or HSV)-216(or
870
J.
s. D.
Figure 2: Sample queries and top 30 responses.
Bonet and P. A. Viola
Structure Driven Image Database Retrieval
871
,
.~t
.
?
., 1::
'" ~
.
"t- '.
:tic
~~ ~~
p ..
~.
-'
'i'~,.'
..
,L1 ~
~<
..
J, rf.
-"
~
~
.
'\\
'
";"~.
..
.,'"
" :". : ... . '
~' ..
. "r;IiB
,~~-o:
~
.,
~,~
Figure 3: Two target sets used in the retrieval experiments. Top: Variation of
object location. BOTTOM: Variation of hard shadows.
c::
o
+='
U
Q)
0.5
CD
~
~
O~--~L-~LL~?
Bnghtness
Camera Position
Contrast
Hard Shadows
Soft shadows
NOise 0.75
Object Location
Object Pose
-
c::
o
U
Q) 0.5
Q)
:3~
o
NOise 0.25
NOise 0 .5
-
c::
o
U
Q) 0 .5
-
I
~
ROe-l16C
RGe- O,, 2C
"lGa-2I SNN
~GB - ~l2N.'"
. ~S\l-2 16C
H'>V-2"i~
I"+S"'~ l 2N"
COr-lU"M
c .~
,,'
Q)
:3~
Occlusion
Rotation (360deg)
Rotation (60deg)
Translation
Zoom
Figure 4: The percentage of target images ranked above all the distract or images,
shown for 15 target sets such as those in Figure 3. The textures-of-textures model
presented here achieves perfect performance in 13 of the 15 experiments.
512)NN histograms in which similarity is measured using a nearest neighbor metric; COR-full(and low)res full resolution (and 4x downsampled) image correlation.
J. S. D. Bonet and P. A. Viola
872
Rankings for all 10 target images in each of the 45 queries are obtained for each
variation. To get a comparative sense of the overall performance of each technique,
we show the number of target images retrieved with a Neyman-Pearson criterion
of zero, i.e. no false positives Figure 4. The textures-of-textures model substantially outperforms all of the other techniques, achieving perfect performance in 13
experiments.
5
Discussion
We have presented a technique for approximating perceived visual similarUsing
ity, by measuring the structural content similarity between images.
the high dimensional "characteristic signature" space representation, we diA world wide
rectly compare database-images to a set of query-images.
web interface to system has been created and is available via the URL:
http://www.ai.mit.edu/~jsd/Research/lmageDatabase/Demo
Experiments indicate that the present system can retrieve images which share visual characteristics with the query-images, from a large non-homogeneous database.
Further, it greatly outperforms many of the standard methods which form the basis
of other systems.
Though the results presented here are encouraging, on real world queries, the retrieved images often contain many false alarms, such as those in Figure 2; however,
we believe that with additional analysis performance can be improved.
References
Desimone, R., Albright, T. D., Gross, C. G., and Bruce, C. (1984) . Stimulus selective
properties of inferior temporal neurons in the macaque. Journal of Neuroscience,
4:2051-2062.
Duda, R . and Hart, P. (1973) . Pattern Classification and Scene Analysis. John Wiley and
Sons.
Kelly, M., Cannon, T. M., and Hush, D. R. (1995). Query by image example: the candid
approach. SPIE Vol. 2420 Storage and Retrieval for Image and Video Databases III,
pages 238-248.
Niblack, V., Barber, R ., Equitz, W ., Flickner, M. , Glasman, E. , Petkovic, D., Yanker,
P., Faloutsos, C., and Taubin, G. (1993). The qbic project: querying images by
content using color, texture, and shape. ISf3TjSPIE 1993 International Symposium
on Electronic Imaging: Science f3 Technology, 1908:173-187.
Pentland, A., Picard, R. W., and Sclaroff, S. (1995). Photobook: Content-based manipulation of image databases. Technical Report 255, MIT Media Lab.
Picard, R . W. and Kabir, T . (1993) . Finding similar patterns in large image databases.
ICASSP, V:161-164.
QBIC. The ibm qbic project. Web: http://wwwqbic.almaden.ibm.comf.
Rao, R. P. N. and Ballard, D. (1995). Object indexing using an iconic sparse distributed
memory. Technical Report TR-559, University of Rochester.
Santini, S. and Jain, R. (1996) . Gabor space and the development of preattentive similarity.
In Proceedings of ICPR 96. International Conference on Pattern Recognition, Vienna.
Viola, P. (1996). Complex feature recognition: A bayesian approach for learning to recognize objects. Technical Report 1591, MIT AI Lab.
Virage. The virage project. Web: http://www.virage.com/ .
von der Malsburg, C. (1988). Pattern recognition by labeled graph matching. Neural
Networks, 1:141-148.
| 1445 |@word duda:2 seems:1 rgb:2 tr:1 outperforms:2 current:1 comparing:3 com:1 si:4 yet:1 must:2 tot:1 john:1 shape:2 depict:1 petkovic:1 intelligence:1 half:2 indicative:1 ith:1 provides:1 contribute:1 hsv:2 location:2 along:1 constructed:2 become:1 ect:1 symposium:1 consists:1 roughly:1 examine:1 multi:1 detects:1 discounted:1 little:1 snn:1 window:2 encouraging:1 taubin:1 project:3 homepage:1 medium:1 tic:1 substantially:1 developed:1 finding:2 temporal:1 every:1 producing:1 positive:1 local:5 textural:1 limit:1 path:1 specifying:1 practical:1 thirty:1 camera:2 differs:1 prevalence:1 procedure:1 gabor:1 projection:6 matching:1 specificity:2 downsampled:2 get:1 cannot:1 storage:1 context:1 www:3 straightforward:1 resolution:4 retrieve:2 ity:1 variation:6 target:23 user:2 homogeneous:1 element:6 recognition:6 iib:1 database:20 labeled:1 bottom:2 capture:2 thousand:1 region:3 decrease:1 highest:1 gross:1 signature:15 creates:1 upon:1 basis:2 icassp:1 eigenimages:1 america:1 jain:2 describe:1 artificial:1 query:18 detected:1 pearson:1 supplementary:1 highlighted:1 indication:1 interaction:1 relevant:1 combining:1 cluster:4 produce:1 generating:1 perfect:2 comparative:1 object:7 iq:3 pose:1 measured:1 nearest:1 strong:2 dividing:1 shadow:3 indicate:1 differ:2 filter:4 human:3 bin:1 visually:2 cognition:1 vary:1 achieves:1 perceived:3 sensitive:1 successfully:1 weighted:1 anecdotal:1 mit:4 cannon:2 ax:5 fence:2 jsd:1 iconic:1 rank:1 prevalent:1 greatly:1 contrast:3 sense:1 squaring:1 nn:1 typically:1 unlikely:1 initially:1 transformed:1 selective:1 pixel:1 overall:2 classification:1 orientation:1 almaden:1 development:1 spatial:1 summed:1 once:1 f3:1 extraction:1 others:3 stimulus:1 quantitatively:1 report:3 oriented:1 tightly:1 zoom:1 recognize:1 occlusion:1 detection:1 organization:1 picard:4 alignment:1 truly:1 light:1 edge:4 capable:1 desimone:2 experience:1 tree:3 loosely:1 circle:1 re:1 soft:1 rao:2 measuring:2 deviation:1 successful:1 characterize:1 answer:1 fundamental:1 international:2 discriminating:1 ie:2 von:2 again:1 jeremy:1 de:1 ranking:1 performed:1 observer:1 lab:2 portion:1 sort:2 rochester:1 bruce:1 square:1 descriptor:1 characteristic:16 variance:2 yield:1 bayesian:1 lu:1 rectified:1 researcher:1 submitted:1 email:1 definition:1 energy:1 spie:1 massachusetts:1 color:10 car:1 astronomical:1 carefully:1 actually:1 sophisticated:1 response:7 improved:1 done:1 though:3 strongly:3 correlation:1 horizontal:1 web:3 ei:5 bonet:4 nonlinear:1 lack:1 resemblance:1 believe:3 effect:2 normalized:1 contain:2 laboratory:1 semantic:1 mahalanobis:1 ll:1 inferior:1 criterion:4 l1:1 interface:1 pro:1 fj:1 reasoning:1 image:85 fi:1 rotation:2 quarter:1 corel:1 approximates:1 measurement:4 cambridge:1 ai:4 fk:1 similarly:1 moving:1 similarity:15 cortex:1 rectly:1 retrieved:2 driven:4 manipulation:2 certain:1 meta:2 santini:2 der:2 additional:3 determine:1 paradigm:1 full:2 photographic:1 technical:3 retrieval:13 hart:2 post:1 vision:1 metric:1 searcher:1 histogram:6 roe:1 achieved:1 cell:2 receive:1 appropriately:1 structural:2 iii:1 automated:1 reduce:1 gb:1 url:1 useful:1 clear:1 repeating:1 ten:2 http:4 generate:1 percentage:1 neuroscience:1 conform:1 vol:1 group:9 salient:1 achieving:1 imaging:1 graph:1 inverse:2 respond:1 electronic:1 layer:3 handful:1 scene:2 icpr:1 across:2 son:1 indexing:1 rectification:1 neyman:1 mechanism:1 cor:2 photo:1 dia:1 available:1 operation:2 observe:1 faloutsos:1 convolved:1 top:4 remaining:1 include:1 malsburg:2 vienna:1 establish:1 approximating:1 question:1 pride:1 arrangement:1 added:1 responds:2 diagonal:1 distance:4 me:2 barber:1 index:1 relationship:2 downsampling:2 difficult:1 perform:4 upper:1 vertical:2 convolution:2 neuron:1 sing:1 pentland:2 viola:6 defining:1 situation:1 ever:1 pair:1 specified:1 hush:2 macaque:1 able:1 bar:1 pattern:5 perception:1 eighth:1 rf:1 including:1 memory:1 video:1 niblack:2 critical:2 natural:1 hybrid:1 ranked:1 scheme:1 technology:3 picket:2 axis:1 created:1 extract:2 kelly:2 literature:1 determining:1 relative:2 fully:1 bear:1 filtering:1 proportional:1 querying:1 skyscraper:2 var:2 asterisk:1 consistent:2 share:1 cd:2 qbic:4 translation:1 ibm:2 changed:1 institute:1 neighbor:1 wide:1 absolute:1 sparse:1 distributed:2 boundary:1 dimension:2 world:2 computes:1 equated:1 made:2 collection:1 deg:2 global:1 conclude:1 demo:1 search:2 ballard:2 channel:2 distract:2 complex:5 necessarily:1 domain:1 noise:3 paul:1 alarm:1 repeated:2 wiley:1 inferring:1 position:1 lq:1 perceptual:1 third:1 specific:1 normalizing:1 exists:2 consist:1 false:2 importance:3 texture:19 sclaroff:1 depicted:1 photograph:1 simply:1 visual:12 extracted:1 ma:1 marked:1 content:6 change:1 hard:2 included:1 typical:1 determined:1 rge:1 multimedia:1 total:1 albright:1 attempted:1 preattentive:1 formally:1 support:1 people:1 photobook:1 |
489 | 1,446 | Factorizing Multivariate Function Classes
Juan K. Lin*
Department of Physics
University of Chicago
Chicago, IL 60637
Abstract
The mathematical framework for factorizing equivalence classes of
multivariate functions is formulated in this paper. Independent
component analysis is shown to be a special case of this decomposition. Using only the local geometric structure of a class representative, we derive an analytic solution for the factorization. We
demonstrate the factorization solution with numerical experiments
and present a preliminary tie to decorrelation.
1
FORMALISM
In independent component analysis (ICA), the goal is to find an unknown linear
coordinate system where the joint distribution function admits a factorization into
the product of one dimensional functions. However, this decomposition is only
rarely possible. To formalize the notion of multivariate function factorization, we
begin by defining an equivalence relation.
Definition. We say that two functions f, 9 : IRn -t IR are equivalent if there exists
A,b and c such that: f(x) = cg(Ax+b), where A is a non-singular matrix and c f. O.
Thus, the equivalence class of a function consists of aU invertible linear transformations of it. To avoid confusion, equivalence classes will be denoted in upper
case, and class representatives in lower case. We now define the product of two
equivalence classes. Consider representatives b : IRn -t IR, and c : IRm -t IR of corresponding equivalence classes Band C. Let Xl E IRn , x"2 E IRm , and x = (xl, x"2).
From the scalar product of the two functions, define the function a : IRn+m -+ IR
by a(x) = b(xI)c(x"2). Let the product of Band C be the equivalence class A with
*
Current address: E25-201, MIT, Cambridge, MA 02139. Email: jklin@ai.mit.edu
1. K Lin
564
representative a(x). This product is independent of the choice of representatives of
Band C, and hence is a well defined operation on equivalence classes. We proceed
to define the notion of an irreducible class.
Definition. Denote the equivalence class of constants by I. We say that A is
irreducible if A = BC implies either B = A, C = I, or B = I, C = A.
From the way products of equivalence classes are defined, we know that all equivalence classes of one dimensional functions are irreducible. Our formulation of the
factorization of multivariate function classes is now complete. Given a multivariate
function, we seek a factorization of the equivalence class of the given representative
into a product of irreducibles. Intuitively, in the context of joint distribution functions, the irreducible classes constitute the underlying sources. This factorization
generalizes independent component analysis to allow for higher dimensional "vector" sources. Consequently, this decomposition is well-defined for all multivariate
function classes. We now present a local geometric approach to accomplishing this
factorization.
2
LOCAL GEOMETRIC INFORMATION
Given that the joint distribution factorizes into a product in the "source" coordinate
system, what information can be extracted locally from the joint distribution in a
"mixed" coordinate frame? We assume that the relevant multivariate function is
twice differentiable in the re!jion of interest, and denote H f, the Hessian of I, to be
the matrix with elements Hij = oiod, where Ok = a~k'
Proposition: H' is block diagonal everywhere, oiojllso = 0 for all points
and all i ~ k, j > k, il and only il 1 is separable into a sum I(SI,"" sn)
g( SI, ... , Sk) + h( Sk+l, ... , sn) for some functions 9 and h.
Proof - Sufficiency:
Given l(sl, . .. , sn) = g(SI, . .. , Sk)
So
=
+ h(sk+1, . .. , Sn),
0 21 _ ~ Oh(Sk+1,"" Sn) _ 0
OSiOSj - OSi
OSj
everywhere for all i
~
k, j
>
k.
Necessity:
From H{n = 0, we can decompose
1 into
l(sl, S2,???, sn) = 9(SI, ... , sn-t}
for some functions
j > k, we find
9 and h.
+ h(S2"'" sn),
Continuing by imposing the constraints H [. = 0 for all
J
l(sl, S2,"" sn) = 9(S1, ... , Sk)
+ h(S2"'" sn).
Combining with Htj = 0 for all j > k yields
I(SI, S2,???, sn) = 9(SI, ... , Sk) + h(S3, ... , sn).
Finally, inducting on i, from the constraints Ht.
arrive at the desired functional form
~
I(SI,S2"",Sn)
= 0 for
all i <
- k and J' > k , we
= g(SI,. ",Sk) + h(Sk+l, ... ,Sn).
Factorizing Multivariate Function Classes
565
More explicitly, a twice-differentiable function satisfies the set of coupled partial
differential equations represented by the block diagonal structure of H if and only
if it admits the corresponding separation of variables decomposition. By letting
log p = f, the additive decomposition of f translates to a product decomposition of
p. The more general decomposition into an arbitrary number of factors is obtained
by iterative application of the above proposition. The special case of independent
component analysis corresponds to a strict diagonalization of H. Thus, in the
context of smooth joint distribution functions, pairwise conditional independence is
necessary and sufficient for statistical independence.
To use this information in a transformed "mixture" frame, we must understand how
the matrix Hlog p transforms. From the relation between the mixture and source
As, we have 8~;
Aji 8~j' where we use Eincoordinate systems given by fl
stein's convention of summation over repeated indices. From the relation between
the joint distributions in the mixture and source frames, Ps(S) = IAIPx (fl), direct
differentiation gives
8 2 10gps(S) _ A A 8 2 10gPx(i)
=
OSi 8s 1
.
H ..
L et tmg
lJ
=
.
8 2 10gp (8)
8s; 8Sj
an d H-lJ..
=
ji
-
=
kl
8x j 8 x k
8 2 10gp (x).
Z
8x; 8Xj
, m
.
.
.
matnx
notatIOn
we have
H = AT if A. In other words, H is a second rank (symmetric) covariant tensor.
The joint distribution admits a product decomposition in the source frame if and
only if H and hence AT if A has the corresponding block diagonal structure. Thus
multivariate function class factorization is solved by joint block diagonalization of
symmetric matrices, with constraints on A of the form AjiifjkAkl = 0.
Because the Hessian is symmetric, its diagonalization involves only (n choose 2)
constraints. Consequently, in the independent component analysis case where the
joint distribution function admits a factorization into one dimensional functions,
if the mixing transformation is orthogonal, the independent component coordinate
system will lie along the eigenvector directions of if. Generally however, n(n 1) independent constraints corresponding to information from the Hessian at two
points are needed to determine the n arbitrary coordinate directions.
3
NUMERICAL EXPERIMENTS
In the simplest attack on the factorization problem, we solve the constraint equations from two points simultaneously. The analytic solution is demonstrated in two
dimensions. Without loss of generality, the mixing matrix A is taken to be of the
form
A=(~ ~) .
The constraints from the two points are: ax + b(xy + 1) + cy = 0, and
a'x + b'(xy + 1) + e'y = 0, where Hu = a, H21 = H12 = b and H22 = e at the
first point, and the primed coefficients denote the values at the second point.
Solving the simultaneous quadratic equations, we find
x
a'e - ae' ? v(ale - ae' )2 - 4(a' b - ab' ) (b'e - be')
2(ab' - a'b)
566
1. K. Lin
y
a'e - ae' ? v(ale - ae')2 - 4(a'b - ab')(b'e - be')
2(bc' - b'e)
The ? double roots is indicative of the (x, y) ~ (l/y, l/x) symmetry in the equations, and together only give two distinct orientation solutions. These independent
component orientation solutions are given by 81 = tan-l(l/x) and 82 = tan-ley).
3.1
Natural Audio Sources
To demonstrate the analytic factorization solution, we present some proof of concept
numerics. Generality is pursued over optimization concerns. First, we perform the
standard separation of two linearly mixed natural audio sources. The input dataset
consists of 32000 un-ordered datapoints, since no use will be made of the temporal
information. The process for obtaining estimates of the Hessian matrix if is as
follows. A histogram of the input distribution was first acquired and smoothed by a
low-pass Gaussian mask in spatial-frequency space. The elements of if were then
obtained via convolution with a discrete approximation of the derivative operator.
The width of the Gaussian mask and the support of the derivative operator were
chosen to reduce sensitivity to low spatial-frequency uncertainty. It should be noted
that the analytic factorization solution makes no assumptions about the mixing
transformation, consequently, a blind determination of the smoothing length scale
is not possible because of the multiplicative degree of freedom in each source.
Because of the need to take the logarithm of p before differentiation, or equivalently
to divide by p afterwards, we set a threshold and only extracted information from
points where the number of counts was greater than threshold. This is justified from
a counting uncertainty perspective, and also from the understanding that regions
with vanishing probability measure contain no information.
With our sample of 32000 datapoints, we considered only the bin-points with a
corresponding bin count greater than 30. From the 394 bin locations that satisfied
this constraint, the solutions (8 1 Jh) for all (394 choose 2) = (394?393/2) pairs of
the corresponding factorization equations are plotted in Fig. 1. A histogram ofthese
solutions are shown in Fig. 2. The two peaks in the solution histogram correspond
to orientations that differ from the two actual independent component orientations
by 0.008 and 0.013 radians. The signal to mixture ratio of the two outputs generated
from the solution are 158 and 49.
3.2
Effect of Noise
Because the solution is analytic, uncertainty in the sampling just propagates through
to the solution, giving rise to a finite width in the solution's distribution. We
investigated the effect of noise and counting uncertainty by performing numerics
starting from analytic forms for the source distributions. The joint distribution in
the source frame was taken to be:
Normalization is irrelevant since a function's decomposition into product form is
preserved in scalar multiplication. This is also reflected in the equivalence between
Hiogp and Hiog cp for e an arbitrary positive constant. The joint distribution in
the mixture frame was obtained from the relation Px(x) = IAI-Ips(S'). To simulate
Factorizing Multivariate Function Classes
. .
567
._ ... -'.o r' ~ . . . .
~.
n/4
82
..????? n:I..:-l
re/4
??-Tt/4
Figure 1: Scatterplot of the independent component orientation solutions. All
unordered solution pairs ((h, (J2) are plotted. The solutions are taken in the range
from -7r /2 to 7r /2.
~~~~r--------'r--------'.--------,--------~--------~--------~
??_????n/2
?_??. ?n!4
e
n/4
rr/2
Figure 2: Histogram of the orientation solutions plotted in the previous figure.
The range is still taken from -7r /2 to 7r /2, with the histogram wrapped around
to ease the circular identification. The mixing matrix used was: all = 0.0514,
a21 = 0.779, a12 = 0.930, a22 = -0.579, giving independent component orientations
at -0.557 and 1.505 radians. Gaussian fit to the centers of the two solution peaks
give -0.570 ? 0.066 and 1.513 ? 0.077 radians for the two orientations.
sampling, Px (x) was multiplied with the number of samples M, onto which was
added Gaussian distributed noise with amplitude given by the (M Px(X))1/2. This
reflects the fact that counting uncertainty scales as the square root of the number
of counts. The result was rounded to the nearest integer, with all negative count
values set to zero. The subsequent processing coincided with that for natural audio
sources. From the source distribution equation above, the minimum number of
expected counts is M, and the maximum is 9M. The results in Figures 3 and 4
show that, as expected, increasing the number of samplings decreases the widths
of the solution peaks. By fitting Gaussians to the two peaks, we find that the
uncertainty (peak widths) in the independent component orientations changes from
0.06 to 0.1 radians as the sampling is decreased from for M = 20 to M = 2. So
even with few samplings, a relatively accurate determination of the independent
component coordinate system can be made.
1. K Lin
568
- IT.!::!
e
- rr/4
rr/ 2
rr/ 4
Figure 3: Histogram of the independent component orientation solutions for four
different samplings. Solutions were generated from 20000 randomly chosen pairs
of positions. The curves, from darkest to lightest, correspond to solutions for the
noiseless, M = 20,11 and 2 simulations. The noiseless solution histogram curve
extends to a height of approximately 15000 counts, and is accurate to the width of
the bin. The slight scatter is due to discretization noise. Spikes at (} = 0 and -7r /2
correspond to pairs of positions which contain no information.
. . . . .
rrJ2
. . . . . .
~4-
-?~---_ - -;r-- _ "?_- "? +- - -" _ - -r-- ?""? Y- " ?~'"
e -
+''---%-- .____4- -
. . . . . .
-+''.---3;-'' .. - .-3;- ---T...
-
-Y-".".+ - -
-
-rr./4 --rr./2 p ..... -T .........;!;- ,,- ?? ;E-..
...-T ........;!;-??.. ... ;E-.. " .?. .;E-. . .. .;1;-- ???? .;!;- .. ?? ;E- .
. . . . .
",.....
-r ..".. .;E- .
.-1'-" .. '"
.... -:E-.
os
""
.
... -:E-.
..:r ??cr?
=
=
=
M
Figure 4: The centers and widths of the solution peaks as a function of the minimum
expected number of counts M . From the source distribution, the maximum expected
number of counts is 9M. Information was only extracted from regions with more
than 2M counts. The actual independent component orientation as determined
from the mixing matrix A are shown by the two dashed lines. The solutions are
very accurate even for small samplings.
4
RELATION TO DECORRELATION
Ideally, if a mixed tensor (transforms as J = A-I j A) with the full degrees of
freedom can be found which is diagonal if and only if the joint distribution appears
in product form, then the independent component coordinate directions will coincide
with that of the tensor's eigenvectors. However, the preceding analysis shows that
a maximum of n(n -1)/2 constraints contain all the information that exists locally.
This, however, provides a nice connection with decorrelation.
Starting with the characteristic function of log p(i) , ?(k) =
the off diagonal terms of Hlogp are given by
J ei k.;E
logp(X) di,
which can loosely be seen as the second order cross-moments in ?(k). Thus di-
Factorizing Multivariate Function Classes
569
agonalization of Hiog P roughly translates into decor relation in ?(k). It should be
noted that ?(k) is not a proper distribution function. In fact, it is a complex valued
function with ?ek) = ?* (-k). Consequently, the summation in the above equation
is not an expectation value, and needs to be interpreted as a superposition of plane
waves with specified wavelengths, amplitudes and phases.
5
DISCUSSION
The introduced functional decomposition defines a generalization of independent
component analysis which is valid for all multivariate functions. A rigorous notion of the decomposition of a multivariate function into a set of lower dimensional
factors is presented. With only the assumption of local twice differentiability, we
derive an analytic solution for this factorization [1]. A new algorithm is presented,
which in contrast to iterative non-local parametric density estimation ICA algorithms [2, 3, 4], performs the decomposition analytically using local geometric information. The analytic nature of this approach allows for a proper treatment of
source separation in the presence of uncertainty, while the local nature allows for a
local determination of the source coordinate system. This leaves open the possibility of describing a position dependent independent component coordinate system
with local linear coordinates patches.
The presented class factorization formalism removes the decomposition assumptions needed for independent component analysis, and reinforces the well known
fact that sources are recoverable only up to linear transformation. By modifying
the equivalence class relation, a rich underlying algebraic structure with both multiplication and addition can be constructed. Also, it is clear that the matrix of
second derivatives reveals an even more general combinatorial undirected graphical
structure of the multivariate function. These topics, as well as uniqueness issues of
the factorization will be addressed elsewhere [5].
The author is grateful to Jack Cowan, David Grier and Robert Wald for many
invaluable discussions.
References
[1] J. K. Lin, Local Independent Component Analysis, Ph. D. thesis, University of
Chicago, 1997.
[2] A. J. Bell and T. J . Sejnowski, Neural Computation 7, 1129 (1995).
[3] S. Amari, A. Cichocki, and H. Yang, in Advances in Neural and Information
Processing Systems, 8, edited by D. S. Touretzky, M. C. Mozer, and M. E.
Hasselmo (MIT Press, Cambridge, MA, 1996), pp. 757-763.
[4] B. A. Pearlmutter and L. Parra, in Advances in Neural and Information Processing Systems, 9, edited by M. C. Mozer, M. I. Jordan, and T. Petsche (MIT
Press, Cambridge, MA, 1997), pp. 613-619.
[5] J. K. Lin, Graphical Structure of Multivariate Functions, in preparation.
| 1446 |@word grier:1 open:1 hu:1 simulation:1 seek:1 decomposition:13 tmg:1 moment:1 necessity:1 bc:2 current:1 discretization:1 si:8 scatter:1 must:1 chicago:3 numerical:2 additive:1 subsequent:1 analytic:8 remove:1 pursued:1 leaf:1 indicative:1 plane:1 vanishing:1 gpx:1 provides:1 location:1 attack:1 height:1 mathematical:1 along:1 constructed:1 direct:1 differential:1 consists:2 fitting:1 acquired:1 pairwise:1 mask:2 expected:4 ica:2 roughly:1 actual:2 increasing:1 begin:1 underlying:2 notation:1 what:1 interpreted:1 eigenvector:1 htj:1 transformation:4 differentiation:2 temporal:1 tie:1 before:1 positive:1 local:10 approximately:1 twice:3 au:1 equivalence:14 ease:1 factorization:17 oiod:1 range:2 matnx:1 block:4 aji:1 bell:1 word:1 onto:1 operator:2 context:2 equivalent:1 demonstrated:1 center:2 starting:2 oh:1 datapoints:2 notion:3 coordinate:10 tan:2 gps:1 element:2 solved:1 cy:1 region:2 decrease:1 edited:2 mozer:2 ideally:1 grateful:1 solving:1 joint:12 represented:1 distinct:1 sejnowski:1 solve:1 valued:1 say:2 amari:1 gp:2 ip:1 jion:1 differentiable:2 rr:6 h22:1 product:12 j2:1 relevant:1 combining:1 osi:2 mixing:5 double:1 p:1 derive:2 nearest:1 a22:1 involves:1 implies:1 convention:1 differ:1 direction:3 ley:1 modifying:1 a12:1 bin:4 generalization:1 preliminary:1 decompose:1 proposition:2 summation:2 parra:1 around:1 considered:1 uniqueness:1 estimation:1 combinatorial:1 superposition:1 hasselmo:1 reflects:1 mit:4 gaussian:4 primed:1 avoid:1 cr:1 factorizes:1 ax:2 rank:1 contrast:1 rigorous:1 cg:1 dependent:1 lj:2 relation:7 irn:4 transformed:1 issue:1 orientation:11 denoted:1 spatial:2 special:2 smoothing:1 sampling:7 few:1 irreducible:4 randomly:1 simultaneously:1 phase:1 ab:3 freedom:2 interest:1 circular:1 possibility:1 mixture:5 accurate:3 partial:1 necessary:1 xy:2 orthogonal:1 continuing:1 irm:2 logarithm:1 re:2 desired:1 divide:1 plotted:3 loosely:1 formalism:2 logp:1 density:1 peak:6 sensitivity:1 physic:1 off:1 invertible:1 rounded:1 together:1 e25:1 thesis:1 satisfied:1 choose:2 juan:1 ek:1 derivative:3 unordered:1 coefficient:1 explicitly:1 blind:1 multiplicative:1 root:2 wave:1 il:3 ir:4 square:1 accomplishing:1 characteristic:1 yield:1 correspond:3 identification:1 simultaneous:1 touretzky:1 email:1 definition:2 frequency:2 pp:2 proof:2 di:2 radian:4 dataset:1 treatment:1 formalize:1 amplitude:2 appears:1 ok:1 higher:1 reflected:1 iai:1 formulation:1 sufficiency:1 generality:2 just:1 ei:1 o:1 defines:1 effect:2 concept:1 contain:3 hence:2 analytically:1 symmetric:3 wrapped:1 width:6 noted:2 complete:1 demonstrate:2 confusion:1 tt:1 performs:1 cp:1 pearlmutter:1 invaluable:1 jack:1 functional:2 ji:1 slight:1 cambridge:3 imposing:1 ai:1 multivariate:15 perspective:1 irrelevant:1 seen:1 minimum:2 greater:2 preceding:1 determine:1 ale:2 signal:1 recoverable:1 dashed:1 afterwards:1 full:1 smooth:1 determination:3 cross:1 lin:6 wald:1 ae:4 noiseless:2 expectation:1 histogram:7 normalization:1 justified:1 preserved:1 addition:1 decreased:1 addressed:1 singular:1 source:17 strict:1 undirected:1 cowan:1 jordan:1 integer:1 counting:3 presence:1 yang:1 independence:2 xj:1 fit:1 reduce:1 translates:2 algebraic:1 proceed:1 hessian:4 constitute:1 generally:1 clear:1 eigenvectors:1 transforms:2 stein:1 band:3 locally:2 ph:1 differentiability:1 simplest:1 sl:3 s3:1 lightest:1 reinforces:1 discrete:1 four:1 threshold:2 ht:1 sum:1 everywhere:2 uncertainty:7 arrive:1 extends:1 separation:3 h12:1 patch:1 fl:2 quadratic:1 constraint:9 simulate:1 performing:1 separable:1 px:3 relatively:1 department:1 ofthese:1 s1:1 intuitively:1 taken:4 equation:7 describing:1 count:9 needed:2 know:1 letting:1 generalizes:1 operation:1 gaussians:1 multiplied:1 petsche:1 darkest:1 graphical:2 giving:2 tensor:3 added:1 spike:1 parametric:1 diagonal:5 topic:1 irreducibles:1 length:1 index:1 ratio:1 equivalently:1 hlog:1 robert:1 hij:1 negative:1 rise:1 numerics:2 proper:2 unknown:1 perform:1 upper:1 convolution:1 finite:1 defining:1 frame:6 smoothed:1 arbitrary:3 introduced:1 david:1 pair:4 kl:1 specified:1 connection:1 address:1 decorrelation:3 natural:3 coupled:1 cichocki:1 sn:14 nice:1 geometric:4 understanding:1 multiplication:2 loss:1 mixed:3 degree:2 sufficient:1 propagates:1 elsewhere:1 allow:1 understand:1 jh:1 distributed:1 curve:2 dimension:1 valid:1 rich:1 author:1 made:2 coincide:1 sj:1 reveals:1 xi:1 factorizing:5 un:1 iterative:2 sk:9 osj:1 nature:2 symmetry:1 obtaining:1 investigated:1 complex:1 linearly:1 s2:6 noise:4 repeated:1 fig:2 representative:6 position:3 a21:1 xl:2 lie:1 coincided:1 admits:4 concern:1 exists:2 scatterplot:1 diagonalization:3 wavelength:1 ordered:1 scalar:2 covariant:1 corresponds:1 satisfies:1 extracted:3 ma:3 conditional:1 goal:1 formulated:1 consequently:4 change:1 determined:1 pas:1 rarely:1 h21:1 support:1 preparation:1 audio:3 |
490 | 1,447 | An Improved Policy Iteratioll Algorithm
for Partially Observable MDPs
Eric A. Hansen
Computer Science Department
University of Massachusetts
Amherst, MA 01003
hansen@cs.umass.edu
Abstract
A new policy iteration algorithm for partially observable Markov
decision processes is presented that is simpler and more efficient than
an earlier policy iteration algorithm of Sondik (1971,1978). The key
simplification is representation of a policy as a finite-state controller.
This representation makes policy evaluation straightforward. The paper's contribution is to show that the dynamic-programming update
used in the policy improvement step can be interpreted as the transformation of a finite-state controller into an improved finite-state controller. The new algorithm consistently outperforms value iteration
as an approach to solving infinite-horizon problems.
1
Introduction
A partially observable Markov decision process (POMDP) is a generalization of the
standard completely observable Markov decision process that allows imperfect information about the state of the system. First studied as a model of decision-making in
operations research, it has recently been used as a framework for decision-theoretic
planning and reinforcement learning with hidden state (Monahan, 1982; Cassandra,
Kaelbling, & Littman, 1994; Jaakkola, Singh, & Jordan, 1995).
Value iteration and policy iteration algorithms for POMDPs were first developed by
Sondik and rely on a piecewise linear and convex representation of the value function
(Sondik, 1971; Smallwood & Sondik,1973; Sondik, 1978). Sondik's policy iteration
algorithm has proved to be impractical, however, because its policy evaluation step is
extremely complicated and difficult to implement. As a result, almost all subsequent
work on dynamic programming for POMDPs has used value iteration. In this paper,
we describe an improved policy iteration algorithm for POMDPs that avoids the difficulties of Sondik's algorithm. We show that these difficulties hinge on the choice of
a policy representation and can be avoided by representing a policy as a finite-state
E. A. Hansen
1016
controller. This representation makes the policy evaluation step easy to implement
and efficient. We show that the policy improvement step can be interpreted in a natural way as the transformation of a finite-state controller into an improved finite-state
controller. Although it is not always possible to represent an optimal policy for an
infinite-horizon POMDP as a finite-state controller, it is always possible to do so when
the optimal value function is piecewise linear and convex. Therefore representation of
a poiicy as a finite-state controller is no more limiting than representation of the value
function as piecewise linear and convex. In fact, it is the close relationship between
representation of a policy as a finite-state controller and representation of a value
function as piecewise linear and convex that the new algorithm successfully exploits.
The paper is organized as follows. Section 2 briefly reviews the POMDP model and
Sondik's policy iteration algorithm. Section 3 describes an improved policy iteration
algorithm. Section 4 illustrates the algorithm with a simple example and reports a
comparison of its performance to value iteration. The paper concludes with a discussion of the significance of this work.
2
Background
Consider a discrete-time POMDP with a finite set of states 5, a finite set of actions
A, and a finite set of observations e. Each time period, the system is in some state
i E 5, an agent chooses an action a E A for which it receives a reward with expected
value ri, the system makes a transition to state j E 5 with probability pij' and the
agent observes () E e with probability tje. We assume the performance objective is
to maximize expected total discounted reward over an infinite horizon.
Although the state of the system cannot be directly observed, the probability that it is
in a given state can be calculated. Let 7r denote a vector of state probabilities, called
an information state, where 7ri denotes the probability that the system is in state i. If
action a is taken in information state 7r and () is observed, the successor information
state is determined by revising each state probability using Bayes' theorem: trj =
LiEs 7riPijQje/ Li,jES 7riPijQje' Geometrically, each information state 7r is a point in
the (151 - I)-dimensional unit simplex, denoted II.
It is well-known that an information state 7r is a sufficient statistic that summarizes
all information about the history of a POMDP necessary for optimal action selection.
Therefore a POMDP can be recast as a completely observable MDP with a continuous
state space II and it can be theoretically solved using dynamic programming. The key
to practical implementation of a dynamic-programming algorithm is a piecewise-linear
and convex representation of the value function. Smallwood and Sondik (1973) show
that the dynamic-programming update for POMDPs preserves the piecewise linearity
and convexity of the value function. They also show that an optimal value function fot
a finite-horizon POMDP is always piecewise linear and convex. For infinite-horizon
POMDPs, Sondik (1978) shows that an optimal value function is sometimes piecewise
linear and convex and can be aproximated arbitrarily closely by a piecewise linear and
convex function otherwise.
A piecewise linear and convex value function V can be represented by a finite set
of lSI-dimensional vectors, r = {aO,a i , ?.. }, such that V(7r) = maxkLi s7riaf. A
dynamic-programming update transforms a value function V representedEfiy a set r
of a-vectors into an improved value function V' represented by a set r' of a-vectors.
Each possible a-vector in r' corresponds to choice of an action, and for each possible
observation, choice of a successor vector in r. Given the combinatorial number of
choices that can be made, the maximum n4mber of vectors in r' is IAllfll91. However
most of these potential vectors are not needed to define the updated value function
and can be pruned. Thus the dynamic-programming update problem is to find a
An Improved Policy Iteration Algorithmfor Partially Observable MDPs
J017
minimal set of vectors r' that represents V', given a set of vectors r that represents
V . Several algorithms for performing this dynamic-programming update have been
developed but describing them is beyond the scope of this paper. Any algorithm for
performing the dynamic-programming update can be used in the policy improvement
step of policy iteration. The algorithm that is presently the fastest is described by
(Cassandra, Littman, & Zhang, 1997).
For value iteration, it is sufficient to have a representation of the value function because
a policy is defined implicitly by the value function, as follows,
8(11") = a(arg mF
L
1I"i o
f),
(1)
iES
where a(k) denotes the action associated with vector ok. But for policy iteration,
a policy must be represented independently of the value function because the policy
evaluation step computes the value function of a given policy. Sondik's choice of
a policy representation is influenced by Blackwell's proof that for a continuous-space
infinite-horizon MDP, there is a stationary, deterministic Markov policy that is optimal
(Blackwell, 1965). Based on this result, Sondik restricts policy space to stationary and
deterministic Markov policies that map the continuum of information space II into
action space A. Because it is important for a policy to have a finite representation,
Sondik defines an admissible policy as a mapping from a finite number of polyhedral
regions of II to A . Each region is represented by a set of linear inequalities, where
each linear inequality corresponds to a boundary of the region.
This is Sondik's canonical representation of a policy, but his policy iteration algorithm
makes use of two other representations. In the policy evaluation step, he converts a
policy from this representation to an equivalent, or approximately equivalent, finitestate controller. Although no method is known for computing the value function of
a policy represented as a mapping from II to A, the value function of a finite-state
controller can be computed in a straightforward way. In the policy improvement
step, Sondik converts a policy represented implicitly by the updated value function
and equation (1) back to his canonical representation. The complexity of translating
between these different policy representations - especially in the policy evaluation step
- makes Sondik's policy iteration algorithm difficult to implement and explains why
it is not used in practice.
3
Algorithm
We now show that policy iteration for POMDPs can be simplified - both conceptually
and computationally - by using a single representation of a policy as a finite-state
controller.
3.1
Policy evaluation
As Sondik recognized, policy evaluation is straightforward when a policy is represented
as a finite-state controller. An o-vector representation of the value function of a finitestate controller is computed by solving the system of linear equations,
k _
0i -
a(k)
ri
+ (3'"'
a(k) a(k) s(k ,8)
L.JPij qj8 OJ
,
(2)
j ,8
where k is an index of a state of the finite-state controller, a(k) is the action associated
with machine state k, and s(k,O) is the index of the successor machine state if 0 is
observed. This value function is convex as well as piecewise linear because the expected
value of an information state is determined by assuming the controller is started in
the machine state that optimizes it.
E. A. Hansen
1018
1. Specify an initial finite-state controller, <5, and select f. for detecting conver-
gence to an f.-optimal policy.
2. Policy evaluation: Calculate a set r of a-vectors that represents the value
function for <5 by solving the system of equations given by equation 2.
3. Policy improvement: Perform a dynamic-programming update and use the
new set of vectors r' to transform <5 into a new finite-state controller, <5', as
follows:
(a) For each vector a in r':
l. If the action and successor links associated with a duplicate those of
a machine state of <5, then keep that machine state unchanged in 8'.
ii. Else if a pointwise dominates a vector associated with a machine state
of <5, change the action and successor links of that machine state to
those used to create a. (If it pointwise dominates the vectors of more
than one machine state, they can be combined into a single machine
state.)
iii. Otherwise add a machine state to <5' that has the same action and
successor links used to create a.
(b) Prune any machine state for which there is no corresponding vector in
r', as long as it is not reachable from a machine state to which a vector
in r' does correspond.
4. Termination test. If the Bellman residual is less than or equal to f.(1 - /3)//3,
exit with f.-optimal policy. Otherwise set <5 to <5' and go to step 2.
Figure 1: Policy iteration algorithm.
3.2
Policy improvement
The policy improvement step uses the dynamic-programming update to transform a
value function V represented by a set r of a-vectors into an improved value function
V' represented by a set r' of a-vectors. We now show that the dynamic-programming
update can also be interpreted as the transformation of a finite-state controller 8 into
an improved finite-state controller <5'. The transformation is made based on a simple
comparison of r' and r.
First note that some of the a-vectors in r' are duplicates of a-vectors in r, that is,
their action and successor links match (and their vector values are pointwise equal).
Any machine state of <5 for which there is a duplicate vector in r' is left unchanged.
The vectors in r' that are not duplicates of vectors in r indicate how to change the
finite-state controller. If a non-duplicate vector in r' pointwise dominates a vector
in r, the machine state that corresponds to the pointwise dominated vector in r is
changed so that its action and successor links match those of the dominating vector
in r'. If a non-duplicate vector in r' does not pointwise dominate a vector in r, a
machine state is added to the finite-state controller with the same action and successor
links used to generate the vector. There may be some machine states for which there
is no corresponding vector in r' and they can be pruned, but only if they are not
reachable from a machine state that corresponds to a vector in r'. This last point is
important because it preserves the integrity of the finite-state controller.
A policy iteration algorithm that uses these simple transformations to change a finitestate controller in the policy improvement step is summarized in Figure 1. An algorithm that performs this transformation is easy to implement and runs very efficiently
because it simply compares the a-vectors in r' to the a-vectors in r and modifies the
finite-state controller accordingly. The policy evaluation step is invoked to compute
the value function of the transformed finite-state controller. (This is only necessary
An Improved Policy Iteration Algorithmfor Partially Observable MDPs
1019
if a machine state has been changed, not if machine states have simply been added.)
It is easy to show that the value function of the transformed finite-state controller /j'
dominates the value function of the original finite-state controller, /j, and we omit the
proof which appears in (Hansen, 1998).
Theorem 1 If a finite-state controller is not optimal, policy improvement transforms
it into a finite-state controller with a value function that is as good or better for every
information state and better for some information state.
3.3
Convergence
If a finite-state controller cannot be improved in the policy improvement step (Le., all
the vectors in r' are duplicates of vectors in r), it must be optimal because the value
function satisfies the optimality equation. However policy iteration does not necessarily converge to an optimal finite-state controller after a finite number of iterations
because there is not necessarily an optimal finite-state controller. Therefore we use
the same stopping condition used by Sondik to detect t-optimality: a finite-state controller is t-optimal when the Bellman residual is less than or equal to t(l- {3) / {3, where
{3 denotes the discount factor. Representation of a policy as a finite-state controller
makes the following proof straightforward (Hansen, 1998).
Theorem 2 Policy iteration converges to an t-optimal finite-state controller after a
finite number of iterations.
4
Example and performance
We illustrate the algorithm using the same example used by Sondik: a simple twostate, two-action, two-observation POMDP that models the problem of finding an
optimal marketing strategy given imperfect information about consumer preferences
(Sondik,1971,1978). The two states of the problem represent consumer preference
or lack of preference for the manufacturers brand; let B denote brand preference
and ....,B denote lack of brand preference. Although consumer preferences cannot be
observed, they can be infered based on observed purchasing behavior; let P denote
purchase of the product and let ....,p denote no purchase. There are two marketing
alternatives or actions; the company can market a luxury version of the product (L)
or a standard version (S). The luxury version is more expensive to market but can
bring greater profit. Marketing the luxury version also increases brand preference.
However consumers are more likely to purchase the less expensive, standard product.
The transition probabilities, observation probabilities, and reward function for this
example are shown in Figure 2. The discount factor is 0.9.
Both Sondik's policy iteration algorithm and the new policy iteration algorithm converge in three iterations from a starting policy that is equivalent to the finite-state
AClions
Markel
luxury
producl (L)
Markel
slandard
producl (S)
Transilion
probabililies
B -B
B/O.8/0.2\
-B 0.5 0.5
B -B
-BB~
0.4 o.
Observalion
probabililies
P
-p
B10.81 0.2\
-B 0.60.4
P
Expecled
reward
-BB?j
?4
-p
-BB~
O. 0.
-BBbj
?3
Figure 2: Parameters for marketing example of Sondik (1971,1978) .
E A. Hansen
1020
.
~; " -~ ''''
..
" a=L \
~ 9,96
:
'- 18.86 <,8=-P
'~~9~~_:~~~;~<~._
= S " \\
14.82! \
: ' .1
:
\
18.20 /
\
', __ __ - '~~ P \.,
''',
,.;,\
/""---.. ,9=-p,y:
" a.= S \ .. ,'"
:
14.86
t ____ . .
\_: 8.1~ / 8=P
(.)
(b)
(e)
(d)
(e)
Figure 3: (a) shows the initial finite-state controller, (b) uses dashed circles to show the
vectors in r' generated in the first policy improvement step and (c) shows the transformed
finite-state controller, (d) uses dashed circles to show the vectors in r' generated in the second
policy improvement step and (e) shows the transformed finite-state controller after policy
evaluation. The optimality of this finite-state controller is detected on the third iteration,
which is not shown. Arcs are labeled with one of two possible observations and machine
states are labeled with one of two possible actions and a 2-dimensional vector that contains
a value for each of the two possible system states.
controller shown in Figure 3a. Figure 3 shows how the initial finite-state controller
is transformed into an optimal finite-state controller by the new algorithm. In the
first iteration, the updated set of vectors r' (indicated by dashed circles in Figure 3b)
includes two duplicate vectors and one non-duplicate that results in an added machine
state. Figure 3c shows the improved finite-state controller after the first iteration. In
the second iteration, each of the three vectors in the updated set of vectors r' (indicated by dashed circles in Figure 3d) pointwise dominates a vector that corresponds
to a current machine state. Thus each of these machine states is changed. Figure 4e
shows the improved finite-state controller after the second iteration. The optimality
of this finite-state controller is detected in the third iteration.
This is the only example for which Sondik reports using policy iteration to find an optimal policy. For POMDPs with more than two states, Sondik's algorithm is especially
difficult to implement. Sondik reports that his algorithm finds a suboptimal policy
for an example described in (Smallwood & Sondik, 1973). No further computational
experience with his algorithm has been reported.
The new policy iteration algorithm described in this paper easily finds an optimal
finite-state controller for the example described in (Smallwood & Sondik, 1973) and
has been used to solve many other POMDPs. In fact, it consistently outperforms value
iteration. We compared its performance to the performance of value iteration on a
suite of ten POMDPs that represent a range of problem sizes for which exact dynamicprogramming updates are currently feasible. (Presently, exact dynamic-prorgramming
updates are not feasible for POMDPs with more than about ten or fifteen states,
actions, or observations.) Starting from the same point, we measured how soon each
algorithm converged to f-optimality for f values of 10.0, 1.0, 0.1 , and 0.01. Policy
iteration was consistently faster than value iteration by a factor that ranged from a
low of about 10 times faster to a high of over 120 times faster. On average, its rate
of convergence was between 40 and 50 times faster than value iteration for this set
of examples. The finite-state controllers it found had as many as several hundred
machine states, although optimal finite-state controllers were sometimes found with
just a few machine states.
An Improved Policy Iteration Algorithm for Partially Observable MDPs
5
1021
Discussion
We have demonstrated that the dynamic-programming update for POMDPs can be
interpreted as the improvement of a finite-state controller. This interpretation can
be applied to both value iteration and policy iteration. It provides no computational
speedup for value iteration, but for policy iteration it results in substantial speedup by
making policy evaluation straightforward and easy to implement. This representation
also has the advantage that it makes a policy easier to understand and execute than
representation as a mapping from regions of information space to actions. In particular, a policy can be executed without maintaining an information state at run-time.
It is well-known that policy iteration converges to f-optimality (or optimality) in
fewer iterations than value iteration. For completely observable MDPs, this is not a
clear advantage because the policy evaluation step is more computationally expensive
than the dynamic-programming update. But for POMDPs, policy evaluation has loworder polynomial complexity compared to the worst-case exponential complexity of
the dynamic-programming update (Littman et al., 1995). Therefore, policy iteration
appears to have a clearer advantage over value iteration for POMDPs. Preliminary
testing bears this out and suggests that policy iteration significantly outperforms value
iteration as an approach to solving infinite-horizon POMDPs.
Acknowledgements
Thanks to Shlomo Zilberstein and especially Michael Littman for helpful discussions.
Support for this work was provided in part by the National Science Foundation under
grants IRI-9409827 and IRI-9624992.
References
Blackwell, D. {1965} Discounted dynamic programming. Ann. Math. Stat. 36:226235.
Cassandra, A.; Kaelbling, L.P.; Littman, M.L. {1994} Acting optimally in partially
observable stochastic domains. In Proc. 13th National Conf. on AI, 1023-1028.
Cassandra, A.; Littman, M.L.; & Zhang, N.L. (1997) Incremental pruning: A simple,
fast, exact algorithm for partially observable Markov decision processes. In Proc. 13th
A nnual Con/. on Uncertainty in AI.
Hansen, E.A. (1998). Finite-Memory Control of Partially Observable Systems. PhD
thesis, Department of Computer Science, University of Massachusetts at Amherst.
Jaakkola, T.; Singh, S.P. ; & Jordan, M.I. (1995) Reinforcement learning algorithm for
partially observable Markov decision problems. In NIPS-7.
Littman, M.L.; Cassandra, A.R.; & Kaebling, L.P. (1995) Efficient dynamicprogramming updates in partially observable Markov decision processes. Computer
Science Technical Report CS-95-19, Brown University.
Monahan, G.E. (1982) A survey of partially observable Markov decision processes:
Theory, models, and algorithms. Management Science 28:1-16.
Smallwood, R.D. & Sondik, E.J. (1973) The optimal control of partially observable
Markov processes over a finite horizon. Operations Research 21:1071-1088.
Sondik, E.J. (1971) The Optimal Control of Partially Observable Markov Processes.
PhD thesis, Department of Electrical Engineering, Stanford University.
Sondik, E.J. (1978) The optimal control of partially observable Markov processes over
the infinite horizon: Discounted costs. Operations Research 26:282-304.
| 1447 |@word version:4 briefly:1 polynomial:1 termination:1 fifteen:1 profit:1 initial:3 contains:1 uma:1 outperforms:3 current:1 must:2 subsequent:1 shlomo:1 update:15 stationary:2 fewer:1 twostate:1 accordingly:1 provides:1 detecting:1 loworder:1 math:1 preference:7 simpler:1 zhang:2 polyhedral:1 theoretically:1 market:2 expected:3 behavior:1 planning:1 tje:1 bellman:2 discounted:3 company:1 provided:1 linearity:1 interpreted:4 developed:2 revising:1 finding:1 transformation:6 impractical:1 suite:1 every:1 control:4 unit:1 grant:1 omit:1 engineering:1 approximately:1 studied:1 suggests:1 fastest:1 range:1 practical:1 fot:1 testing:1 practice:1 implement:6 significantly:1 cannot:3 close:1 selection:1 equivalent:3 deterministic:2 map:1 demonstrated:1 modifies:1 straightforward:5 go:1 starting:2 independently:1 convex:10 survey:1 pomdp:8 iri:2 dominate:1 smallwood:5 his:4 limiting:1 updated:4 exact:3 programming:16 us:4 expensive:3 labeled:2 observed:5 solved:1 electrical:1 worst:1 calculate:1 region:4 observes:1 substantial:1 convexity:1 complexity:3 reward:4 littman:7 dynamic:17 singh:2 solving:4 eric:1 exit:1 completely:3 conver:1 easily:1 represented:9 fast:1 describe:1 detected:2 stanford:1 dominating:1 solve:1 otherwise:3 statistic:1 transform:2 advantage:3 product:3 convergence:2 incremental:1 converges:2 illustrate:1 clearer:1 stat:1 measured:1 c:2 indicate:1 closely:1 stochastic:1 successor:9 translating:1 explains:1 ao:1 generalization:1 preliminary:1 scope:1 mapping:3 continuum:1 proc:2 combinatorial:1 currently:1 hansen:8 create:2 successfully:1 always:3 jaakkola:2 zilberstein:1 improvement:13 consistently:3 detect:1 helpful:1 stopping:1 hidden:1 transformed:5 arg:1 denoted:1 equal:3 represents:3 purchase:3 simplex:1 report:4 piecewise:11 duplicate:9 few:1 preserve:2 national:2 luxury:4 evaluation:14 necessary:2 experience:1 circle:4 minimal:1 earlier:1 kaelbling:2 cost:1 hundred:1 optimally:1 reported:1 chooses:1 combined:1 thanks:1 amherst:2 michael:1 thesis:2 management:1 conf:1 li:1 potential:1 summarized:1 includes:1 sondik:30 bayes:1 complicated:1 contribution:1 efficiently:1 correspond:1 conceptually:1 pomdps:14 history:1 converged:1 influenced:1 finitestate:3 proof:3 associated:4 con:1 proved:1 massachusetts:2 organized:1 jes:1 back:1 appears:2 ok:1 specify:1 improved:14 execute:1 marketing:4 just:1 receives:1 lack:2 defines:1 indicated:2 mdp:2 ranged:1 brown:1 theoretic:1 performs:1 bring:1 invoked:1 recently:1 he:1 interpretation:1 ai:2 had:1 reachable:2 add:1 integrity:1 dynamicprogramming:2 optimizes:1 inequality:2 arbitrarily:1 greater:1 prune:1 recognized:1 converge:2 maximize:1 period:1 dashed:4 ii:6 technical:1 match:2 faster:4 long:1 y:1 controller:50 iteration:53 represent:3 sometimes:2 background:1 else:1 jordan:2 iii:1 easy:4 suboptimal:1 imperfect:2 action:19 algorithmfor:2 clear:1 transforms:2 discount:2 ten:2 generate:1 lsi:1 restricts:1 canonical:2 discrete:1 key:2 geometrically:1 convert:2 run:2 uncertainty:1 almost:1 decision:9 summarizes:1 simplification:1 ri:3 dominated:1 extremely:1 optimality:7 pruned:2 performing:2 speedup:2 department:3 describes:1 making:2 presently:2 trj:1 taken:1 computationally:2 equation:5 nnual:1 describing:1 needed:1 operation:3 manufacturer:1 alternative:1 original:1 denotes:3 hinge:1 maintaining:1 exploit:1 especially:3 unchanged:2 objective:1 added:3 strategy:1 link:6 gence:1 assuming:1 consumer:4 index:2 relationship:1 pointwise:7 difficult:3 executed:1 implementation:1 policy:85 perform:1 observation:6 markov:12 arc:1 finite:57 blackwell:3 nip:1 beyond:1 recast:1 oj:1 memory:1 difficulty:2 rely:1 natural:1 residual:2 representing:1 mdps:5 started:1 concludes:1 b10:1 review:1 acknowledgement:1 bear:1 foundation:1 agent:2 purchasing:1 pij:1 sufficient:2 changed:3 last:1 soon:1 understand:1 boundary:1 calculated:1 transition:2 avoids:1 computes:1 monahan:2 reinforcement:2 made:2 avoided:1 simplified:1 bb:3 pruning:1 observable:18 implicitly:2 keep:1 infered:1 continuous:2 why:1 necessarily:2 domain:1 significance:1 exponential:1 lie:1 third:2 admissible:1 theorem:3 dominates:5 phd:2 illustrates:1 horizon:9 cassandra:5 easier:1 mf:1 simply:2 likely:1 kaebling:1 partially:15 corresponds:5 satisfies:1 ma:1 ann:1 markel:2 feasible:2 change:3 infinite:7 determined:2 acting:1 total:1 called:1 brand:4 select:1 support:1 |
491 | 1,449 | Adaptation in Speech Motor Control
John F. Houde*
UCSF Keck Center
Box 0732
San Francisco, CA 94143
Michael I. Jordan
MIT Dept. of Brain and Cognitive Sci.
EI0-034D
Cambridge, MA 02139
houde~phy.ucsf.edu
jordan~psyche.mit.edu
Abstract
Human subjects are known to adapt their motor behavior to a
shift of the visual field brought about by wearing prism glasses
over their eyes. We have studied the analog of this effect in speech.
Using a device that can feed back transformed speech signals in
real time, we exposed subjects to alterations of their own speech
feedback. We found that speakers learn to adjust their production
of a vowel to compensate for feedback alterations that change the
vowel's perceived phonetic identity; moreover, the effect generalizes
across consonant contexts and to different vowels.
1
INTRODUCTION
For more than a century, it has been know that humans will adapt their reaches
to altered visual feedback [8]. One of the most studied examples of this adaptation
is prism adaptation, which is seen when a subject reaches to targets while wearing
image-shifting prism glasses [2]. Initially, the subject misses the targets, but he
soon learns to compensate and reach accurately. This compensation is retained
beyond the time that the glasses are worn: when the glasses are removed, the
subject's reaches now overshoot targets in the direction that he compensated. This
retained compensation is called adaptation, and its generation from exposure to
altered sensory feedback is called sensorimotor adaptation (SA).
In the study reported here, we investigated whether SA could be observed in a
motor task that is quite different from reaching - speech production. Specifically,
we examined whether the control of phonetically relevant speech features would
respond adaptively to altered auditory feedback. By itself, this is an important
theoretical question because various aspects of speech production have already been
shown to be sensitive to auditory feedback [5, 1, 4]. Moreover, we were particularly
*To whom correspondence should be addressed.
39
Adaptation in Speech Motor Control
interested in whether speech SA would also exhibit generalization. If so, speech SA
could be used to examine the organization of speech motor control. For example,
suppose we observed adaptation of [c) in "get". We could then examine whether
we also see adaptation of [c) in "peg". IT so, then producing [c) in the two different
words must access a common, adapted representation - evidence for a hierarchical
speech production system in which word productions are composed from smaller
units such as phonemes. We could also examine whether adapting [c) in "get"
causes adaptation of [re) in "gat". IT so, then the production representations of [c)
and [re] could not be independent, supporting the idea that vowels are produced by
controlling a common set of features. Such theories about the organization of the
speech production system have been postulated in phonology and phonetics, but
the empirical evidence supporting these theories has generally been observational
and hence not entirely conclusive [7,6].
2
METHODS
To study speech SA, we focused on vowel production because the phonetically relevant features of vowel sounds are formant frequencies, which are feasible to alter
in real time. 1
To alter the formants of a subject's speech feedback, we built the apparatus shown
in Figure 1. The subject wears earphones and a microphone and sits in front of
a PC video monitor that presents words to be spoken aloud. The signal from the
microphone is sent to a Digital Signal Processing board, which collects a 64ms
time interval from which a magnitude spectrum is calculated. From this spectrum,
formant frequencies and amplitudes are estimated. To alter the speech, the first
three formant frequencies are shifted, and the shifted formants drive a formant
synthesizer that creates the output speech sent to the subject's earphones. This
analysis-synthesis process was accomplished with only 16ms of feedback delay. To
minimize how much the subject directly heard of his own voice via bone conduction,
the subject produced only whispered speech, masked with mild noise.
altered
feedback
Altered Fonnants
I
earphones.
;-------...
....
I ?.
/
pep
DSP board
in PC
I
?
PC video monitor
microphone .'
.'"
L....-_~
.....
Fl.F2.F3
Frequency
Alteration.
'.
Fonnants
V\IFonnant
/\ ?Estimation I
L.....---' . .
~.
Magnitude Spectrum
Figure 1: The apparatus used in the study.
For each subject in our experiment, we shifted formants along the path defined
by the (F1,F2,F3) frequencies of a subject's productions of the vowels [i) , [t.]' [c], [re],
1 See
[3] for detailed discussion of the methods used in this study.
J. F. Houde and M. I Jordan
40
and [a].2 Figure 2 shows examples of this shifting process in (Fl,F2) space for the
feedback transformations that were used in the study. TOIshift formants along the
subject's [i]-[a] path, we extend the path at both ends and we number the endpoints
and vowels to make a path position measure that normalizes the distances between
vowels. The formants of each speech sound F produced by the subject were then
re-represented in terms of path projection - the path posiUon of nearest path point
P, and path deviation - the distance D to this point P. Feedback transformations
were constructed to alter path projections while preserving path deviations. Two
different transformations were used. The +2.0 transformation added 2.0 to path
projections: under this transform, if the subject produced speech sound F (a sound
near [cD, he heard instead sound F+ (a sound near [aD. The subject could compensate for this transform and hear sound F only by shifting his production of F to
F- (a sound near [iD. The -2.0 transformation subtracted 2.0 from path projections:
under this transformation, if the subject produced F, he heard F-. Thus, in this
case, the subject could compensate by shifting production to F+.
F-
F+
F+
[ah)O 5
end /)6
end /)6
Fl
Fl
(a) +2.0 Transformation
(b) -2.0 Transformation
Figure 2: Feedback transformations used in the study.
These feedback transformations were used in an experiment in which a subject was
visually prompted to whisper words with a 300ms target duration_ Word promptings
occurred in groups of ten called epochs. Within each epoch, the first six words came
from a set of training words and the last four came from a set of testing words.
The subject heard feedback of his first five word productions in each epoch, while
masking noise blocked his hearing for his remaining five word productions in the
epoch. Thus, the subject only heard feedback of his production of the first five
training words and never heard his productions of the testing words.
2Where possible, we use standard phonetic symbols for vowel sounds: [i] as in "seat",
[L] as in "hit", [c] as in "get", [re] as in "hat", and [a] as in "pop". Where font limitations
prevent us from using these symbols, we use the alternate notation of [i], [ib], [eh] , rae],
and lab], respectively, for the same vowel sounds.
Adaptation in Speech Motor Control
41
The experiment lasted 2 hours and consisted of 422 epochs divided over five phases:
1. A 10 minute warmup phase used to acclimate the subject to the experimen-
tal setup.
2. A 17 minute baseline phase used to measure formants of the subject's normal vowel productions.
3. A 20 minute ramp phase in which the subject's feedback was increasingly
altered up to a maximum value.
4. A 1 hour training phase in which the subject produced words while the
feedback was maximally altered.
5. A 17 minute test phase used to measure formants of the subject's postexposure vowel productions while his feedback was maximally altered.
By the end of the ramp phase, feedback alteration reached its maximum strength,
which was +2.0 for half the subjects and -2.0 for the other subjects. In addition,
all subjects were run in a control experiment in which feedback was never altered.
The two word sets from which prompted words were selected were both sets of
eve words. Training words (in which adaptation was induced) were all bilabials
with [c] as the vowel ("pep", "peb", "bep", and "beb"). Testing words (in which
generalization of the training word adaptation was measured) were divided into two
subsets, each designed to measure a different type of generalization: (1) context
generalization words, which had the same vowel [c] as the training words but varied
the consonant context ("peg", "gep", and "teg"); (2) vowel target generalization
words, which had the same consonant context as the training words but varied the
vowel ("pip,", "peep,", "pap" , and "pop").
Eight male MIT students participated in the study. All were native speakers of
North American English and all were naive to the purpose of the study.
3
RESULTS
To illustrate how we measured compensation and adaptation in the experiments,
we first show the results for an individual subject. Figure 3 shows (F1,F2) plots of
response of subject OB in both the adaptation experiment (in which he was exposed
to the -2.0 feedback transformation) and the control experiment. In each figure, the
dotted line is OB's [i]-[a] path.
Figure 3(a) shows OB's compensation responses, which were measured from his
productions of the training words made when he heard feedback of his whispering.
The solid arrow labeled "-2.0 xform" shows how much his mean vowel formants
changed (testing phase - baseline phase) after being exposed to the -2.0 feedback
transformation. It shows he shifted his production of [c] to something a bit past
[re], which corresponds to a p;;tth projection change of slightly more than one vowel
interval towards [a]. Thus, since the path projection shift of the transform was -2.0
(2.0 vowel intervals towards liD, the figure shows that OB compensates for over
half the action of the transformation. The hollow arrow in Figure 3(a) shows how
OB heard his compensation. It shows he heard his actual production shift from [c]
towards [a] as a shift from [i] back towards [c].
Figure 3(b) shows how much of OB's compensation was retained when he whispered
the training words with feedback blocked by noise .. This retained compensation is
called adaptation, and it was measured from path projection changes by the same
method used to measure compensation. In the figure, we see OB's adaptation
J. F. Houde and M. I Jordan
42
response (the solid "-2.0 xform" arrow) is a path projection shift of slightly less
than one vowel interval, so his adaptation is slightly less than half. Thus, the figure
shows that OB retains an appreciable amount of his compensation in the absence
of feedback.
Finally, in both plots of Figure 3, the almost non-existent "control" arrows show
that OB exhibited almost no formant change in the control experiment - as we
would expect since feedback was never altered in this experiment.
2600
2600
2400
2400
[i) ?.
[i) ~.
:~
2200
N
2000
.\\.
N
1800
\.
control "':
~
?2.0 dorm
C'I
"'
[ab) ?
..
[eb) '-. ~ .
1800
control , \. 2.0 dorm
~
[ae) \
1600
'.
1400
[ib)",
'-'
rae) \
1600
\~
2000
X
[eb;??? ~ ~
X
C'I
'\\
2200
\
[ab)?
1400
1200
1200
300
SOO
700
900
Fl (Hz)
(a) compensation
II 00
300
SOO
700
900
1100
Fl (Hz)
(b) adaptation
Figure 3: Subject OB compensation and adaptation.
The plots in Figure 4 show that there was significant compensation and adaptation
across all subjects. In these plots, the vertical scale indicates how much the changes
in mean vowel formants (testing phase - baseline phase) in each subject's productions of the training words compensated for the action of the feedback transformation he was exposed to. The filled circles linked by the solid line show compensation
(Figure 4(a? and retained compensation, or adaptation (Figure 4(b)) across subjects in the adaptation experiment in which feedback was altered; the open circles
linked by the dotted line show the same measures from the control experiment in
which feedback was not altered. (The solid and dotted lines facilitate comparison of
results across subjects but do not signify any relationship between subjects.) In the
control experiment, for each subject, compensation and adaptation were measured
with respect to the feedback transformation used in the adaptation experiment.
The plots show that there are large variations in compensation and adaptation
across subjects, but overall there was significantly more compensation (p < 0.006)
and adaptation (p < 0.023) in the adaptation experiments that in the control experiments.
Figure 5 shows plots of how much of the adaptation observed in the training words
carried over the the testing words. For each testing word shown, a measure of this
carryover called mean generalization is plotted, which was calculated as a ratio of
adaptations: the adaptation seen in the testing word divided by the adaptation seen
in the training words (adaptation values observed in the control experiment were
Adaptation in Speech Motor Control
43
1.2
1.2
1.0
1.0
0 .80
0.80
0 .60
0 .60
0.40
?
;
0.20
\
/rs \"
/i
0.00
0.40
ro
0.20
\~_
vs
.J: .
.~
ah
ty
-0.20
vs
sr
0.00
sr
-0.20
ro
ab
cw
-0.40
-0 .40
(a) mean compensation
(b) mean adaptatioll
Figure 4: Mean compensation and adaptation across all subjects.
subtracted out to remove any effects not arising from exposure to altered feedback).
Figure 5(a) shows mean generalization for the context generalization words except
for "pep" (since "pep" was also a training word) . The plot shows large variance
in mean generalization for each of the three words, but overall there was significant (p < 0.040) mean generalization. Thus, there was significant carryover of the
adaptation of [c] in the training words to different consonant contexts.
Figure 5(b) shows mean generalization for the vowel target generalization words.
Not all of these words are shown: unfortunately, we weren't able to accurately
estimate the formants of [i] and [a], so "peep" and "pop" were dropped from our
generalization analysis. For the remaining two vowel target generalization words,
the plot shows large variance in mean generalization for each of the words, but
overall there was significant (p < 0.013) mean generalization. Thus, there was
significant generalization of the adaptation of [c] to the vowels [t] and [eel.
4
DISCUSSION
Several conclusions can be drawn from the experiment described above. First, comparison of the adaptation and control experiment results seen in Figure 4 shows
a clear effect of exposure to the altered feedback: this exposure caused compensation responses in most subjects. Furthermore, the adaptation results show that
this compensation was retained in absence of acoustic feedback. Next, the context
generalization results seen in Figure 5(a) show that some adapted representation of
[c] is shared across the training and testing words. These results provide evidence
for a hierarchical speech production system in which words are composed from
smaller phoneme-like units. Finally, the vowel target generalization results seen in
Figure 5(b) show that the production representations of [t], [cl, and [ee] are not
independent, suggesting that these vowels are produced by controlling a common
set of features.
1. F. Houde and M. I. Jordan
44
2.0
2.0
1.5
1.5
peg
.,:
Q)
.,:
1.0
0
c:
<'S
Q)
~
gep
teg
0
c:
<'S
0.50
pap
1.0
Q)
Q)
~
0.50
0.00
0.00
-0.50
?0.50
(a) context gen. words
pip
A
~
(b) vowel target gen. words
Figure 5: Mean generalization for the testing words, averaged across subjects.
Thus, in summary, our study has shown (1) that speech production, like reaching,
can be made to exhibit sensorimotor adaptation, and (2) that this adaptation effect
exhibits generalization that can be used to make inferences about the structure of
the speech production system.
Acknowledgments
We thank J. Perkell, K. Stevens, R. Held and P. Sabes for helpful discussions.
References
[1] V. L. Gracco et al., (1994) J. Acoust. Soc. Am. 95:2821
[2] H. V. Helmholtz, (1867) Treatise on physiological optics, Vol. 3 (Eng. Trans.
by Optical Soc. of America, Rochester, NY, 1925)
[3] J. F. Houde (1997), Sensorimotor Adaptation in Speech Production, Doctoral
Dissertation, M. I. T., Cambridge, MA.
[4) H. Kawahara (1993) J. Acoust. Soc. Am. 94:1883.
[5] B. S. Lee (1950) J. Acoust. Soc. Am. 22:639.
[6] W. J. M. Levelt (1989), Speaking: from intention to articulation, MIT Press,
Cambridge, MA.
[7] A. S. Meyer (1992), Cognition 42:18l.
[8) R. B. Welch (1986), Handbook of Perception and Human Performance, K. R.
Boff, L. Kaufman, J. P. Thomas Eds., John Wiley and Sons, New York.
| 1449 |@word mild:1 open:1 r:1 eng:1 solid:4 phy:1 xform:2 past:1 synthesizer:1 must:1 john:2 motor:7 remove:1 designed:1 plot:8 v:2 half:3 selected:1 device:1 dissertation:1 sits:1 five:4 warmup:1 along:2 constructed:1 behavior:1 examine:3 formants:10 brain:1 actual:1 moreover:2 notation:1 kaufman:1 spoken:1 acoust:3 transformation:15 ro:2 hit:1 control:17 unit:2 producing:1 dropped:1 apparatus:2 id:1 whisper:1 path:16 eb:2 studied:2 examined:1 doctoral:1 collect:1 averaged:1 acknowledgment:1 testing:10 empirical:1 adapting:1 significantly:1 projection:8 word:44 intention:1 get:3 context:8 center:1 compensated:2 worn:1 exposure:4 focused:1 welch:1 bep:1 seat:1 his:16 century:1 variation:1 target:9 controlling:2 suppose:1 perkell:1 helmholtz:1 particularly:1 native:1 labeled:1 observed:4 removed:1 existent:1 overshoot:1 exposed:4 creates:1 f2:4 represented:1 various:1 america:1 dorm:2 kawahara:1 aloud:1 quite:1 ramp:2 compensates:1 formant:5 transform:3 itself:1 adaptation:41 relevant:2 gen:2 boff:1 gep:2 keck:1 houde:6 illustrate:1 measured:5 nearest:1 sa:5 soc:4 direction:1 stevens:1 human:3 observational:1 f1:2 generalization:21 weren:1 normal:1 visually:1 cognition:1 purpose:1 perceived:1 estimation:1 sensitive:1 mit:4 brought:1 reaching:2 beb:1 pip:2 dsp:1 indicates:1 lasted:1 baseline:3 am:3 glass:4 helpful:1 inference:1 initially:1 transformed:1 interested:1 overall:3 field:1 f3:2 never:3 experimen:1 alter:4 carryover:2 composed:2 individual:1 phase:11 vowel:28 ab:3 organization:2 rae:2 adjust:1 male:1 pc:3 held:1 filled:1 re:6 circle:2 plotted:1 theoretical:1 retains:1 hearing:1 deviation:2 subset:1 masked:1 delay:1 front:1 reported:1 conduction:1 adaptively:1 lee:1 eel:1 michael:1 synthesis:1 cognitive:1 american:1 prompting:1 suggesting:1 alteration:4 student:1 north:1 postulated:1 caused:1 ad:1 bone:1 lab:1 linked:2 reached:1 masking:1 rochester:1 minimize:1 phonetically:2 phoneme:2 variance:2 accurately:2 produced:7 drive:1 ah:2 levelt:1 reach:4 ed:1 ty:1 sensorimotor:3 frequency:5 auditory:2 amplitude:1 back:2 feed:1 response:4 maximally:2 box:1 furthermore:1 facilitate:1 effect:5 consisted:1 hence:1 speaker:2 m:3 phonetics:1 image:1 common:3 endpoint:1 analog:1 he:10 extend:1 occurred:1 pep:4 significant:5 blocked:2 cambridge:3 had:2 wear:1 access:1 treatise:1 something:1 own:2 phonetic:2 prism:3 came:2 accomplished:1 seen:6 preserving:1 signal:3 ii:1 sound:10 earphone:3 adapt:2 compensate:4 dept:1 divided:3 ae:1 addition:1 participated:1 signify:1 addressed:1 interval:4 exhibited:1 sr:2 hz:2 subject:42 induced:1 sent:2 jordan:5 eve:1 near:3 ee:1 idea:1 shift:5 whether:5 six:1 speech:26 speaking:1 cause:1 york:1 action:2 generally:1 heard:9 detailed:1 clear:1 amount:1 ten:1 tth:1 peg:3 shifted:4 dotted:3 estimated:1 arising:1 vol:1 group:1 four:1 monitor:2 drawn:1 prevent:1 pap:2 run:1 respond:1 almost:2 ob:10 bit:1 entirely:1 fl:6 correspondence:1 adapted:2 strength:1 optic:1 tal:1 aspect:1 optical:1 alternate:1 across:8 smaller:2 son:1 increasingly:1 psyche:1 slightly:3 lid:1 ei0:1 know:1 end:4 generalizes:1 eight:1 hierarchical:2 subtracted:2 voice:1 hat:1 thomas:1 remaining:2 phonology:1 question:1 already:1 added:1 font:1 exhibit:3 distance:2 cw:1 thank:1 sci:1 whom:1 retained:6 relationship:1 prompted:2 ratio:1 setup:1 unfortunately:1 vertical:1 compensation:21 supporting:2 varied:2 gracco:1 conclusive:1 acoustic:1 pop:3 hour:2 trans:1 beyond:1 able:1 perception:1 articulation:1 hear:1 built:1 soo:2 video:2 shifting:4 eh:1 altered:14 eye:1 sabes:1 carried:1 naive:1 epoch:5 expect:1 generation:1 limitation:1 digital:1 cd:1 production:26 normalizes:1 changed:1 summary:1 last:1 soon:1 english:1 feedback:33 calculated:2 sensory:1 made:2 san:1 handbook:1 francisco:1 consonant:4 spectrum:3 learn:1 ca:1 peep:2 investigated:1 cl:1 arrow:4 noise:3 board:2 ny:1 wiley:1 position:1 meyer:1 ib:2 learns:1 minute:4 symbol:2 physiological:1 evidence:3 gat:1 magnitude:2 visual:2 corresponds:1 ma:3 identity:1 towards:4 appreciable:1 shared:1 absence:2 feasible:1 change:5 specifically:1 except:1 miss:1 microphone:3 called:5 ucsf:2 hollow:1 wearing:2 whispered:2 |
492 | 145 | 459
A BIFURCATION THEORY APPROACH TO THE
PROGRAMMING OF PERIODIC A TTRACTORS IN
NETWORK MODELS OF OLFACTORY CORTEX
Bill Baird
Department of Biophysics
U.C. Berkeley
ABSTRACT
A new learning algorithm for the storage of static
and periodic attractors in biologically inspired
recurrent analog neural networks is introduced.
For a network of n nodes, n static or n/2 periodic
attractors may be stored. The algorithm allows
programming of the network vector field independent of the patterns to be stored. Stability of
patterns, basin geometry, and rates of convergence
may be controlled. For orthonormal patterns, the
l~grning operation reduces to a kind of periodic
outer product rule that allows local, additive,
commutative, incremental learning. Standing or
traveling wave cycles may be stored to mimic the
kind of oscillating spatial patterns that appear
in the neural activity of the olfactory bulb and
prepyriform cortex during inspiration and suffice,
in the bulb, to predict the pattern recognition
behavior of rabbits in classical conditioning experiments. These attractors arise, during simulated inspiration, through a multiple Hopf bifurcation, which can act as a critical "decision pOint"
for their selection by a very small input pattern.
INTRODUCTION
This approach allows the construction of biological models and
the exploration of engineering or cognitive networks that
employ the type of dynamics found in the brain. Patterns of 40
to 80 hz oscillation have been observed in the large scale activity of the olfactory bulb and cortex(Freeman and Baird 86)
and even visual neocortex(Freeman 87,Grey and Singer 88), and
found to predict the olfactory and visual pattern recognition
responses of a trained animal. Here we use analytic methods of
bifurcation theory to design algorithms for determining synaptic weights in recurrent network architectures, like those
460
Baird
found in olfactory cortex, for associative memory storage of
these kinds of dynamic patterns.
The "projection algorithm" introduced here employs higher
order correlations, and is the most analytically transparent
of the algorithms to come from the bifurcation theory approach(Baird 88). Alternative numerical algorithms employing
unused capacity or hidden units instead of higher order correlations are discussed in (Baird 89). All of these methods
provide solutions to the problem of storing exact analog attractors, static or dynamic, in recurrent neural networks, and
allow programming of the ambient vector field independent of
the patterns to be stored. The stability of cycles or equilibria, geometry of basins of attraction, rates of convergence
to attractors, and the location in parameter space of primary
and secondary bifurcations can be programmed in a prototype
vector field - the normal form.
To store cycles by the projection algorithm, we start with the
amplitude equations of a polar coordinate normal form, with
coupling coefficients chosen to give stable fixed points on
the axes, and transform to Cartesian coordinates. The axes of
this system of nonlinear ordinary differential equations are
then linearly transformed into desired spatial or spatio-temporal patterns by projecting the system into network coordinates - the standard basis - using the desired vectors as columns of the transformation matrix. This method of network synthesis is roughly the inverse of the usual procedure in bifurcation theory for analysis of a given physical system.
Proper choice of normal form couplings will ensure that the
axis attractors are the only attractors in the system - there
are no "spurious attractors". If symmetric normal form coefficients are chosen, then the normal form becomes a gradient
vector field. It is exactly the gradient of an explicit potential function which is therefore a strict Liapunov function
for the system. Identical normal form coefficients make the
normal form vector field equivariant under permutation of the
axes, which forces identical scale and rotation invariant
basins of attraction bounded by hyperplanes. Very complex
periodic a~tractors may be established by a kind of Fourier
synthesis as linear combinations of the simple cycles chosen
for a subset of the axes, when those are programmed to be
unstable, and a single "mixed mode" in the interior of that
subspace is made stable. Proofs and details on vectorfield
programming appear in (Baird 89).
In the general case, the network resulting from the projection
A Bifurcation Theory Approach to Programming
algorithm has fourth order correlations, but the use of restrictions on the detail of vector field programming and the
types of patterns to be stored result in network architectures
requiring only s~cond order correlations. For biological modeling, where possibly the patterns to be stored are sparse and
nearly orthogonal, the learning rule for periodic patterns
becomes a "periodic" outer product rule which is local, additive, commutative, and incremental. It reduces to the usual
Hebb-like rule for static attractors.
CYCLES
The observed physiological activity may be idealized mathe11 y as a "
1 e " , r Xj e 1(ej + wt) , J? 1 , 2 , ... ,n. S uc h a
.
ma t 1ca
cyc
cycle is ~ "periodic attractor" if it is stable. The global
amplitude r is just a scaling factor for the pattern ~ , and
the global phase w in e 1wt is a periodic scaling that scales x
by a factor between ? 1 at frequency w as t varies.
The same vector X S or "pattern" of relative amplitudes can
appear in space as a standing wave, like that seen in the
bulb, if the relative phase eS1 of each compartment (component)
is the same, eS 1
eS 1 , or as a traveling wave, like that seen
in the ~repyriform cortex. if the relative phase components of
~s form a gradient in space, eS 1+1 - 1/a e\. The traveling wave
will "sweep out" the amplitude pattern XS in time, but the
root-mean-square amplitude measured in an experiment will be
the same ~s, regardless of the phase pattern. For an arbitrary
phase vector, t~~se "simple" single frequency cycles can make
very complicated looking spatio-temporal patterns. From the
mathematical point of view, the relative phase pattern ~ is a
degree of freedom in the kind patterns that can be stored.
Patterns of uniform amplitude ~ which differed only in the
phase locking pattern ~ could be stored as well.
+, -
To store the kind of patterns seen in bulb, the amplitude
vector ~ is assumed to be parsed into equal numbers of excitatory and inhibitory components, where each class of component
has identical phase. but there is a phase difference of 60 90 degrees between the classes. The traveling wave in the
prepyriform cortex is modeled by introducing an additional
phase g~adient into both excitatory and inhibitory classes.
PROJECTION ALGORITHM
The central result of this paper is most compactly stated as
the following:
461
462
Baird
THEOREM
Any set S, s - 1,2, ... , n/2 , of cycles r S x. s e1(9js + wst) of
linearly independent vectors of relative comJonent amplitudes
xS E Rn and phases ~s E Sn, with frequencies wS E R and global
amplitudes r S E R, may be established in the vector field of
the analog fourth order network:
by some variant of the projection operation :
-1
Tij ... Emn Pim J mn P nj ,
T
ijk1?
EPA
mn im mn
p-1. p- 1
mJ
nk
p- 1
n1'
where the n x n matrix P contains the real and imaginary components [~S cos ~s , ~s sin ~S] of the complex eigenvectors
x S e 19s as columns, J is an n x n matrix of complex conjugate
eigenvalues in diagonal blocks, Amn is an n x n matrix of 2x2
blocks of repeated coefficients of the normal form equations,
and the input b i &(t) is a delta function in time that establishes an initial condition. The vector field of the dynamics
of the global amplitudes rs and phases -s is then given exactly
by the normal form equations :
rs
==
Us r s
In particular, for ask > 0 , and ass/a kS < 1 , for all sand k,
the cycles s - 1,2, ... ,n/2 are stable, and have amplitudes
rs ;; (u s/a ss )1I2, where us? 1 - "T ?
Note that there is a multiple Hopf bifurcation of codimension
n/2 at "T = 1. Since there are no approximations here, however,
the theorem is not restricted to the neighborhood of this
bifurcation, and can be discussed without further reference to
bifurcation theory. The normal form equations for drs/dt and
d_s/dt determine how r S and _s for pattern s evolve in time in
interaction with all the other patterns of the set S. This
could be thought of as the process of phase locking of the
pattern that finally emerges. The unusual power of this algorithm lies in the ability to precisely specify these ~
linear interactions. In general, determination of the modes of
the linearized system alone (li and Hopfield 89) is insufficient to say what the attractors of the nonlinear system
will be.
A Bifurcation Theory Approach to Programming
PROOF
The proof of the theorem is instructive since it is a constructive proof, and we can use it to explain the learning algorithm. We proceed by showing first that there are always fixed
points on the axes of these amplitude equations, whose stability is given by the coefficients of the nonlinear terms. Then
the network above is constructed from these equations by two
coordinate transformations. The first is from polar to Cartesian coordinates, and the second is a linear transformation
from these canonical "mode" coordinates into the standard
basis e 1, e 2, ... , eN' or "network coordinates". This second
transformation constitutes the "learning algorithm", because
it tra"nSfrirms the simple fixed points of the amplitude equations into the specific spatio-temporal memory patterns desired for the network.
Amplitude Fixed Points
Because the amplitude equations are independent of the rotation _, the fixed points of the amplitude equations characterize the asymptotic states of the underlying oscillatory
modes. The stability of these cycles is therefore given by the
stability of the fixed points of the amplitude equations. On
each axis r s ' the other components rj are zero, by definition,
rj
= rj ( uj - Ek a jk r k2 ) ? 0,
r s - rs ( Us - ass r s2 ),
and
for rj ? 0,
which leaves
rs - 0
There is an equilibrium on each axis s, at r s.(u s /a ss )1I2, as
claimed. Now the Jacobian of the amplitude equations at some
fixed point r~ has elements
J lJ
. . - - 2 a lJ
.. r~.1 r.....J ,
J 11 = u.1 - :5 a 11
.. r~.2
- ]7-i
~ a .. r~.2 .
1
lJ
J
For a fixed point r~s on axis s, J ij ? 0 , since r~i or r~j ? 0,
making J a diagonal matrix whose entries are therefore its
eigenvalues. Now J l1 ? u1 - a is r~ s2, for i /. s, and J ss ? Us :5 ass r~/. Since r~/ ? us/ass' J ss ? - 2 us' and J ii ? u i - a is
(us/ass). This gives aisfass > u1/u s as the condition for negative eigenvalues that assures the stability of r ....s . Choice of
aji/a ii ) uj/u i , for all i, j , therefore guarantees stability of
all axis fixed points.
Coordinate Transformations
We now construct the neural network from these well behaved
equations by the following transformations,
First; polar to Cartesian, (rs'-s) to (v2s-1.v2s) : Using
V 2s - 1 '" r s cos -s
v 2s = r s sin -s ,and differentiating these
463
464
Baird
gives:
V
2s - 1 ? r s cos "s
by the chain rule. Now substituting
and r s sin "s ? v 2s ,
cos tis ? v 2s - 1/r s '
gives:
v 2s
- v2s rs
+
(v 2s- l/ r s ) .. s
Entering the expressions of the normal form for rs and tis'
gives:
and since
rs
222
= v 2s- 1 + v 2s
n/2
v 2s - 1 - Us v 2s -1 - Ws v 2s + E j [v 2s -1 a sj - v 2s bsj ] (v 2j -/ + v 2/)
Similarly,
n/2
v 2s - Us v 2s + Ws v 2s-' + E j [v 2s a sj + v 2s - 1 bSj ] (v 2j _/ + v 2/)?
Setting the bsj - 0 for simplicity, choosing Us - - T + 1
to get a standard network form, and reindexing i,j-l,2, ... ,n ,
we get the Cartesian equivalent of the polar normal form equations.
n
n
Here J is a matrix containing 2x2 blocks along the diagonal of
the local couplings of the linear terms of each pair of the
previous equations v 2s -1 ' v 2s ? with - T separated out of the
diagonal terms. The matrix A has 2x2 blocks of identical coefficients a sj of the nonlinear terms from each pair.
J
1
- w,
w,
1
"-
=
1
w2
- w2
1
"
~
a'l a"
a" a 1,
a 12 a'2
a'2 a'2
a 21 a 21
a 21 a 21
a 22 a 22
a 22 a 22
.,
A Bifurcation Theory Approach to Programming
Learning Transformation - Linear Term
Second; J is the canonical form of a real matrix with complex
conjugate eigenvalues, where the conjugate pairs appear in
blocks along the diagonal as shown. The Cartesian normal form
equations describe the interaction of these linearly uncoupled
complex modes due to the coupling of the nonlinear terms. We
can interpret the normal form equations as network equations
in eigenvector (or "memory") coordinates, given by some diagonalizing transformation P, containing those eigenvectors as
its columns, so that J a p- 1 T P. Then it is clear that T may
instead be determined by the reverse projection T _ P J p- 1
back into network coordinates, if we start with desired eigenvectors and eigenvalues. We are free to choose as columns in
P, the real and imaginary vectors [X S cos 9s , XS sin 9S ] of the
cycles ~s e i9s of any linearly independent- set -S of p~tterns
to be learned. If we write the matrix expression for the projection in component form, we recover the expression given in
the theorem for Tij ,
Nonlinear Term Projection
The nonlinear terms are transformed as well, but the expression cannot be easily written in matrix form. Using the component form of the transformation,
substituting into the Cartesian normal form, gives:
Xi - (-'T+1) E j Pij (E k P- 1jk x k) + E j Pij Ek J jk (E I P-\I xl)
+ E j Pij (E k P- 1jk xk) EI Ajl (Em p-\m xm) (En p-\n x n)
Rearranging the orders of summation gives,
Xi = (-'T+1) Ek (E j Pi j P- 1jk ) x k + EI (E k E j Pij J jk P-\l) xl
+ En Em Ek
(
EI E
j
-1
-1
Pij P jk AjI P 1m
p-1
In
)
x k xm xn
Finally, performing the bracketed summations and relabeling
indices gives us the network of the theorem,
xi
=-
'T xi
+
E j T1j Xj
+
Ejkl Tijkl Xj Xk xl
with the expression for the tensor of the nonlinear term,
465
466
Baird
T ijk1 -
Emn P im Amn P
-1
mj P
-1
nk P
-1
n1
Q.E.D.
LEARNING RULE EXTENSIONS
This is the core of the mathematical story, and it may be extended in many ways. When the columns of P are orthonormal,
then p-1 ? pT, and the formula above for the linear network
coupling becomes T = pJpT. Then, for complex eigenvectors,
This is now a local, additive, incremental learning rule for
synapse ij, and the system can be truly self-organizing because the net can modify itself based on its own activity.
Between units of equal phase, or when 9i s = 9j S - 0 for a static
pattern, this reduces to the usual Hebb rule.
In a similar fashion, the learning rule for the higher order
nonlinear terms becomes a multiple periodic outer product rule
when the matrix A is chosen to have a simple form. Given our
present ignorance of the full biophysics of intracellular
processing, it is not entirely impossible that some dimensionality of the higher order weights in the mathematical network
coul~ be implemented locally within the cells of a biological
network, using the information available on the primary lines
given by the linear connections discussed above. When the A
matrix is chosen to have uniform entries Aij - c for all its
off-diagonal 2 x 2 blocks, and uniform entries Aij - c - d
for the diagonal blocks, then,
T ijk1 ?
This reduces to the multiple outer product
The network architecture generated by this learning rule is
This reduces to an architecture without higher order correlations in the case that we choose a completely uniform A matrix
(A 1j - c , for all i,j). Then
+
+
A Bifurcation Theory Approach to Programming
This network has fixed points on the axes of the normal form
as always, but the stability condition is not satisfied since
the diagonal normal form coefficients are equal, not less,
than the remaining A matrix entries. In (Baird 89) we describe
how clamped input (inspiration) can break this symmetry and
make the nearest stored pattern be the only attractor.
All of the above results hold as well for networks with sigmoids, provided their coupling is such that they have a Taylor's expansion which is equal to the above networks up to
third order. The results then hold only in the neighborhood of
the origin for which the truncated expansion is accurate. The
expected performance of such systems has been verified in
simulations.
Acknowledgements
Supported by AFOSR-87-0317. I am very grateful for the support
of Walter Freeman and invaluable assistance of Morris Hirsch.
References
B. Baird. Bifurcation Theory Methods For Programming Static
or Periodic Attractors and Their Bifurcations in Dynamic
Neural Networks. Proc. IEEE Int. Conf. Neural Networks,
San Diego, Ca.,pI-9, July(1988).
B. Baird. Bifurcation Theory Approach to Vectorfield Programming for Periodic Attractors. Proc. INNS/IEEE Int. Conf.
on Neural Networks. Washington D.C., June(1989).
W. J. Freeman & B. Baird. Relation of Olfactory EEG to Behavior: Spatial Analysis. Behavioral Neuroscience (1986).
W. J. Freeman & B. W. van Dijk. Spatial Patterns of Visual
Cortical EEG During Conditioned Reflex in a Rhesus Monkey.
Brain Research, 422, p267(1987).
C. M. Grey and W. Singer. Stimulus Specific Neuronal
Oscillations in Orientation Columns of Cat Visual Cortex.
PNAS. In Press(1988).
Z. Li & J.J. Hopfield. Modeling The Olfactory Bulb. Biological Cybernetics. Submitted(1989}.
467
| 145 |@word grey:2 rhesus:1 linearized:1 simulation:1 r:9 initial:1 contains:1 emn:2 imaginary:2 bsj:3 written:1 numerical:1 additive:3 analytic:1 v2s:3 alone:1 leaf:1 liapunov:1 xk:2 core:1 node:1 location:1 hyperplanes:1 mathematical:3 along:2 constructed:1 differential:1 hopf:2 behavioral:1 olfactory:7 expected:1 roughly:1 equivariant:1 behavior:2 brain:2 inspired:1 freeman:5 becomes:4 provided:1 bounded:1 suffice:1 underlying:1 what:1 kind:6 eigenvector:1 monkey:1 transformation:9 nj:1 guarantee:1 temporal:3 berkeley:1 act:1 ti:2 exactly:2 k2:1 unit:2 appear:4 engineering:1 local:4 modify:1 k:1 co:5 programmed:2 block:7 procedure:1 aji:2 thought:1 projection:8 get:2 cannot:1 interior:1 selection:1 storage:2 impossible:1 adient:1 restriction:1 equivalent:1 bill:1 regardless:1 rabbit:1 simplicity:1 rule:11 attraction:2 orthonormal:2 stability:8 coordinate:10 construction:1 pt:1 diego:1 exact:1 programming:11 origin:1 element:1 recognition:2 jk:7 observed:2 cycle:11 locking:2 dynamic:5 trained:1 grateful:1 basis:2 completely:1 compactly:1 easily:1 hopfield:2 cat:1 separated:1 walter:1 describe:2 neighborhood:2 choosing:1 whose:2 say:1 s:4 ability:1 transform:1 itself:1 associative:1 eigenvalue:5 net:1 inn:1 interaction:3 product:4 organizing:1 convergence:2 oscillating:1 incremental:3 coupling:6 recurrent:3 measured:1 nearest:1 ij:2 implemented:1 come:1 exploration:1 sand:1 transparent:1 biological:4 summation:2 im:2 extension:1 hold:2 normal:17 equilibrium:2 predict:2 substituting:2 polar:4 proc:2 establishes:1 always:2 ej:1 ax:6 june:1 am:1 lj:3 hidden:1 spurious:1 w:3 relation:1 transformed:2 orientation:1 animal:1 spatial:4 bifurcation:16 uc:1 field:8 equal:4 construct:1 washington:1 identical:4 nearly:1 constitutes:1 mimic:1 stimulus:1 employ:2 pim:1 relabeling:1 geometry:2 phase:14 attractor:14 n1:2 freedom:1 truly:1 chain:1 accurate:1 ambient:1 wst:1 orthogonal:1 taylor:1 desired:4 column:6 modeling:2 epa:1 ordinary:1 introducing:1 subset:1 entry:4 uniform:4 characterize:1 stored:9 varies:1 periodic:12 standing:2 off:1 synthesis:2 central:1 satisfied:1 containing:2 choose:2 possibly:1 cognitive:1 conf:2 ek:4 li:2 potential:1 coefficient:7 baird:13 tra:1 int:2 bracketed:1 idealized:1 root:1 view:1 break:1 wave:5 start:2 recover:1 complicated:1 square:1 compartment:1 cybernetics:1 submitted:1 explain:1 oscillatory:1 synaptic:1 definition:1 frequency:3 proof:4 static:6 ask:1 emerges:1 dimensionality:1 amplitude:18 back:1 higher:5 dt:2 response:1 specify:1 synapse:1 just:1 correlation:5 traveling:4 ei:3 nonlinear:9 mode:5 behaved:1 requiring:1 analytically:1 inspiration:3 entering:1 symmetric:1 i2:2 ignorance:1 amn:2 sin:4 during:3 self:1 assistance:1 l1:1 invaluable:1 rotation:2 physical:1 conditioning:1 analog:3 discussed:3 interpret:1 similarly:1 stable:4 cortex:7 j:1 own:1 reverse:1 ajl:1 store:2 claimed:1 seen:3 additional:1 determine:1 july:1 ii:2 multiple:4 full:1 rj:4 pnas:1 reduces:5 determination:1 e1:1 biophysics:2 controlled:1 variant:1 cell:1 w2:2 strict:1 hz:1 vectorfield:2 unused:1 xj:3 codimension:1 architecture:4 prototype:1 expression:5 cyc:1 proceed:1 tij:2 se:1 eigenvectors:4 clear:1 neocortex:1 locally:1 morris:1 canonical:2 inhibitory:2 delta:1 neuroscience:1 write:1 t1j:1 verified:1 inverse:1 fourth:2 oscillation:2 decision:1 scaling:2 entirely:1 activity:4 precisely:1 x2:3 fourier:1 u1:2 performing:1 department:1 combination:1 conjugate:3 em:2 biologically:1 making:1 projecting:1 invariant:1 restricted:1 equation:18 assures:1 singer:2 drs:1 unusual:1 available:1 operation:2 alternative:1 remaining:1 ensure:1 parsed:1 uj:2 classical:1 sweep:1 tensor:1 primary:2 usual:3 diagonal:8 gradient:3 subspace:1 simulated:1 capacity:1 outer:4 unstable:1 modeled:1 index:1 insufficient:1 stated:1 negative:1 design:1 proper:1 truncated:1 extended:1 looking:1 prepyriform:2 rn:1 arbitrary:1 introduced:2 pair:3 connection:1 learned:1 uncoupled:1 established:2 pattern:31 xm:2 memory:3 power:1 critical:1 force:1 diagonalizing:1 mn:3 axis:5 dijk:1 sn:1 acknowledgement:1 evolve:1 determining:1 tractor:1 relative:5 asymptotic:1 afosr:1 permutation:1 mixed:1 bulb:6 degree:2 basin:3 pij:5 story:1 storing:1 pi:2 excitatory:2 supported:1 free:1 aij:2 allow:1 differentiating:1 sparse:1 van:1 xn:1 cortical:1 made:1 san:1 employing:1 sj:3 global:4 hirsch:1 assumed:1 spatio:3 xi:4 mj:2 ca:2 rearranging:1 symmetry:1 eeg:2 expansion:2 as:5 complex:6 linearly:4 intracellular:1 s2:2 arise:1 repeated:1 neuronal:1 en:3 differed:1 hebb:2 fashion:1 explicit:1 xl:3 lie:1 clamped:1 jacobian:1 third:1 theorem:5 formula:1 specific:2 showing:1 x:3 physiological:1 sigmoids:1 commutative:2 cartesian:6 conditioned:1 nk:2 visual:4 reflex:1 ma:1 determined:1 wt:2 secondary:1 e:3 cond:1 support:1 constructive:1 es1:1 instructive:1 |
493 | 1,450 | A Non-parametric Multi-Scale Statistical
Model for Natural Images
Jeremy S. De Bonet & Paul Viola
Artificial Intelligence Laboratory
Learning & Vision Group
545 Technology Square Massachusetts Institute of Technology
Cambridge, MA 02139
EMAIL: jsd@ai.mit.edu & viola@ai.mit.edu
HOMEPAGE: http://www.ai .mit. edu/projects/lv
Abstract
The observed distribution of natural images is far from uniform.
On the contrary, real images have complex and important structure that can be exploited for image processing, recognition and
analysis. There have been many proposed approaches to the principled statistical modeling of images, but each has been limited in
either the complexity of the models or the complexity of the images. We present a non-parametric multi-scale statistical model for
images that can be used for recognition, image de-noising, and in
a "generative mode" to synthesize high quality textures.
1
Introduction
In this paper we describe a multi-scale
structure of natural images across many
it can be used to recognize novel images,
tasks is reasonably efficient, requiring no
workstation.
statistical model which can capture the
scales. Once trained on example images,
or to generate new images. Each of these
more than a few seconds or minutes on a
The statistical modeling of images is an endeavor which reaches back to the 60's
and 70's (Duda and Hart, 1973). Statistical approaches are alluring because they
provide a unified view of learning, classification and generation. To date however, a
generic, efficient and unified statistical model for natural images has yet to appear.
Nevertheless, many approaches have shown significant competence in specific areas.
Perhaps the most influential statistical model for generic images is the Markov
random field (MRF) (Geman and Geman, 1984). MRF's define a distribution over
774
J. S. D. Bonet and P. A. Viola
images that is based on simple and local interactions between pixels. Though MRF's
have been very successfully used for restoration of images, their generative properties are weak. This is due to the inability of the MRF 's to capture long range
(low frequency) interactions between pixels. Recently there has been a great deal of
interest in hierarchical models such as the Helmholtz machine (Hinton et al., 1995;
Dayan et al. , 1995). Though the Helmholtz machine can be trained to discover long
range structure, it is not easily applied to natural images.
Multi-scale wavelet models have emerged as an effective technique for modeling
realistic natural images. These techniques hypothesize that the wavelet transform
measures the underlying causes of natural images which are assumed to be statistically independent. The primary evidence for this conjecture is that the coefficients of
wavelet transformed images are uncorrelated and low in entropy (hence the success
of wavelet compression) . These insights have been used for noise reduction (Donoho
and Johnstone, 1993; Simoncelli and Adelson, 1996), and example driven texture
synthesis (Heeger and Bergen, 1995). The main drawback of wavelet algorithms is
the assumption of complete independence between coefficients. We conjecture that
in fact there is strong cross-scale dependence between the wavelet coefficients of an
image, which is consistent with observations in (De Bonet, 1997) and (Buccigrossi
and Simoncelli, 1997).
2
Multi-scale Statistical Models
Multi-scale wavelet techniques assume that images are a linear transform of a collection of statistically independent random variables: 1= W-1C, where I is an image,
W- 1 is the inverse wavelet transform, and C = {Ck} is a vector of random variable
"causes" which are assumed to be independent. The distribution of each cause Ck is
Pk ( . ), and since the Ck'S are independent it follows that: p( C) =
k Pk (Ck). Various
wavelet transforms have been developed, but all share the same type of multi-scale
structure - each row of the wavelet matrix W is a spatially localized filter that is
a shifted and scaled version of a single basis function.
n
The wavelet transform is most efficiently computed as an iterative convolution using
a bank of filters . First a "pyramid" of low frequency downsampled images is created:
Go = I , G 1 = 2 ..!-(9 ? Go), and Gi+l = 2 ..!-(9 ? Gi ), where 2..!- downsamples an
image by a factor of 2 in each dimension, ? is the convolution operation, and 9 is
a low pass filter. At each level a series offilter functions are applied: Fj = h (j!) Gj,
where the Ii 's are various types of filters. Computation of the Fj's is a linear
transformation that can thought of as a single matrix W. With careful selection
of 9 and h this matrix can be constructed so that W- 1 = W T (Simoncelli et al.,
1992)1. Where convenient we will combine the pixels of the feature images Fj(x, y)
into a single cause vector C.
The expected distribution of causes, Ck, is a function of the image classes that are
being modeled. For example it is possible to attempt to model the space of all
natural images. In that case it appears as though the most accurate Pk (.) 's are
highly kurtotic which indicates that the Ck ' S are most often zero but in rare cases
take on very large values (Donoho and Johnstone, 1993; Simoncelli and Adelson,
1996) . This is in direct contrast to the distribution of Ck'S for white-noise images which is gaussian. The difference in these distributions can be used as the basis
of noise reduction algorithms, by reducing the wavelet coefficients which are more
lComputation of the inverse wavelet transform is algorithmically similar to the computation of the forward wavelet transform.
Non-Parametric Multi-Scale Statistical Image Model
775
likely to be noise than signal.
Specific image classes can be modeled using similar methods (Heeger and Bergen,
1995)2. For a given set of input images the empirical distribution of the Ck'S is
observed. To generate a novel example of a texture a new set of causes, (}, is sampled
from the assumed independent empirical distributions Pk (.). The generated images
are computed using the inverse wavelet transform: l' = W- 1 (},. Bergen and Heeger
have used this approach to build a probabilistic model of a texture from a ::;ingle
example image. To do this they assume that textures are spatially ergodic - that
the expected distribution is not a function of position in the image. As a result the
pixels in anyone feature image, Fj(x, y), are samples from the same distribution
and can be combined 3 .
Heeger and Bergen's work is at or near the current state of the art in texture
generation. Figure 1 contains some example textures. Notice however, that this
technique is much better at generating smooth or noise-like textures than those
with well defined structure. Image structures, such as the sharp edges at the border
of the tiles in the rightmost texture can not be modeled with their approach. These
image features directly contradict the assumption that the wavelet coefficients, or
causes, of the image are independent.
For many types of natural images the coefficients of the wavelet transform are not
independent, for example images which contain long edges. While wavelets are local
both in frequency and space, a long edge is not local in frequency nor in space. As
a result the wavelet representation of such a feature requires many coefficients. The
high frequencies of the edge are captured by many small high frequency wavelets.
The long scale is captured by a number of larger low frequency wavelets. A model
which assumes these coefficients are independent can never accurately model images
which contain these non-local features. Conversely a model which captures the
conditional dependencies between coefficients will be much more effective. We chose
to approximate the joint distribution of coefficients as a chain, in which coefficients
that occur higher in the wavelet pyramid condition the distribution of coefficients
at lower levels (Le. low frequencies condition the generation of higher frequencies).
For every pixel in an image define the parent vector of that pixel:
V.... (x, y) = [ Fo0 (x, y), Fo1 (x, y), . .. ,FoN (x, y),
-nO
X
y
1
X
y)
l'{(l2"J,
l2"J),F
1 (L2"J, L2"J , ...
N
,Fl
X
(l2"J,
L2"YJ) , ...
o
x
Y
1
X J
Y)
x J' L2M
Y J)]
FM(L2MJ,l2MJ),FM(l2M
,L 2M
J , ... ,FMN ( L2M
(1)
where M is the top level of the pyramid and N is the number of features. Rather
than generating each of these coefficients independently, we define a chain across
scale. In this chain the generation of the lower levels depend on the higher levels:
p(V(x, y))
= p(VM(x, y))
x p(VM- 1 (x, y)IVM(x, y))
x p(VM-2(X, y)!VM-l (x, y), VM(x, y)) x ...
x p(Vo(x, y)IVl (x, y), ... , VM-l (x, y), VM(x, y))
(2)
2See (Zhu, Wu and Mumford, 1996) for a related but more formal model.
3Their generation process is slightly more complex than this, involving a iteration
designed to match the pixel histogram. The implementation used for generating the images
in Figure 1 incorporates this, but we do not discuss it here.
776
1. S. D. Bonet and P. A. Viola
Figure 1: Synthesis results for the Heeger and Bergen (1995) model. Top: Input
textures. BOTTOM: Synthesis results. This technique is much better at generating
fine or noisy textures then it is at generating textures which require co-occurrence
of wavelets at multiple scales.
Figure 2: Synthesis results using our technique for the input textures shown in
Figure 1 (Top).
where Yt(x , y) is the a subset of the elements of Vex , y) computed from C/. Usually
we will assume ergodicity, i.e. that p(V(x, y)) is independent of x and y. The generative process starts from the top of the pyramid, choosing values for the VM (x, y)
at all points. Once these are generated the values at the next level, VM -1 (x , y) , are
generated. The process continues until all of the wavelet coefficients are generated.
Finally the image is computed using the inverse wavelet transform.
It is important to note that this probabilistic model is not made up of a collection
of independent chains, one for each Vex , y). Parent vectors for neighboring pixels
have substantial overlap as coefficients in the higher pyramid levels (which are
Non-Parametric Multi-Scale Statistical Image Model
777
lower resolution) are shared by neighboring pixels at lower pyramid levels. Thus,
the generation of nearby pixels will be strongly dependent. In a related approach a
similar arrangement of generative chains has been termed a Markov tree (Basseville
et al., 1992).
2.1
Estimating the Conditional Distributions
The additional descriptive power of our generative model does not come without
cost. The conditional distributions that appear in (2) must be estimated from
observations. We choose to do this directly from the data in a non-parametric
fashion. Given a sample of parent vectors {8(x, y)} from an example image we
estimate the conditional distribution as a ratio of Parzen window density estimators:
p(
~(
I
x,y
)1v,M (
/+1
x,y
)) _ p(ViM(x,y))
-
'" Lx',y' R(V;M(X,y), 8r(x', y'))
.... M
'"
"'M
....M
p(V/+1(x,y))
Lx',y' R(V/+l (x, y), S/+1 (x',y'))
(3)
where Vik(x,y) is a subset of the parent vector V(x,y) that contains information
from level I to level k, and R(?) is a function of two vectors that returns maximal
values when the vectors are similar and smaller values when the vectors are dissimilar. We have explored various R(?) functions. In the results presented the
R( .) function returns a fixed constant 1/ z if all of the coefficients of the vectors are within some threshold () and zero otherwise. Given this simple definition for R(?) sampling from p(Vz(x,Y)IV;~1(X,y)) is very straightforward: find all
x', y' such that R(8#1 (x', y'), 8#1 (x, y)) = 1/ z and pick from among them to set
Vz(x,y) = SI(X',y').
3
Experiments
We have applied this approach to the problems of texture generation, texture recognition, target recognition, and signal de-noising. In each case our results are competitive with the best published approaches.
In Figure 2 we show the results of our technique on the textures from Figure 1.
For these textures we are better able to model features which are caused by a
conjunction of wavelets. This is especially striking in the rightmost texture where
the geometrical tiling is almost, but not quite, preserved. In our model, knowledge of
the joint distribution provides constraints which are critical in the overall perceived
appearance of the synthesized texture.
Using this same model, we can measure the textural similarity between a known and
novel image. We do this by measuring the likelihood of generating the parent vectors
in the novel image under the chain model of the known image. On "easy" data
sets, such as the the MeasTex Brodatz texture test suite, performance is slightly
higher than other techniques, our approach achieved 100% correct classification
compared to 97% achieved by a gaussian MRF approach (Chellappa and Chatterjee,
1985). The MeasTex lattice test suite is slightly more difficult because each
texture is actually a composition of textures containing different spatial frequencies.
Our approach achieved 97% while the best alternate method, in this case Gabor
Convolution Energy method (Fogel and Sagi, 1989) achieved 89%. Gaussian MRF's
explicitly assume that the texture is a unimodal distribution and as a result achieve
only 79% correct recognition. We also measured performance on a set of 20 types
of natural texture and compared the classification power of this model to that of
human observers (humans discriminate textures extremely accurately.) On this
1. S. D. Bonet and P. A. Viola
778
Original
Denoise Shrinkage
Shrinkage Residual
Noised
Denoise Ours
Our Residual
Figure 3: (Original) the original image; (Noised) the image corrupted with white
gaussian noise (SNR 8.9 dB); (Denoise Shrinkage) the results of de-noising using
wavelet shrinkage or coring (Donoho and Johnstone, 1993; Simoncelli and Adelson,
1996) (SNR 9.8 dB); (Shrinkage Residual) the residual error between the shrinkage
de-noised result and the original - notice that the error contains a great deal of
interpretable structure; (Denoise Ours) our de-noising approach (SNR 13.2 dB);
and (Our Residual) the residual error - these errors are much less structured.
test, humans achieved 86% accuracy, our approach achieved an accuracy of 81%,
and GMRF's achieved 68%.
A strong probabilistic model for images can be used to perform a variety of image
processing tasks including de-noising and sharpening. De-noising of an observed
image i can be performed by Monte Carlo averaging: draw a number of sample
images according to the prior density P(I), compute the likelihood of the noise for
each image P(v = (1) - 1), and then find the weighted average over these images.
The weighted average is the estimated mean over all possible ways that the image
might have been generated given the observation.
Image de-noising frequently relies on generic image models which simply enforce
image smoothness. These priors either leave a lot of residual noise or remove much
of the original image. In contrast, we construct a probability density model from
the noisy image itself. In effect we assume that the image is redundant, containing
many examples of the same visual structures, as if it were a texture. The value of
this approach is directly related to the redundancy in the image. If the redundancy
in the image is very low, then the parent structures will be everywhere different, and
the only resampled images with significant likelihood will be the original image. But
if there is some redundancy in the image - that might arise from a regular texture
or smoothly varying patch - the resampling will freely average across these similar
regions. This will have the effect of reducing noise in these images. In Figure 3 we
show results of this de-noising approach.
Non-Parametric Multi-Scale Statistical Image Model
4
779
Conclusions
We have presented a statistical model of texture which can be trained using example images. The form of the model is a conditional chain across scale on a pyramid of wavelet coefficients. The cross scale condtional distributions are estimated
non-parametrically. This is important because many of the observed conditional
distributions are complex and contain multiple modes. We believe that there are
two main weaknesses of the current approach: i) the tree on which the distributions
are defined are fixed and non-overlapping; and ii) the conditional distributions are
estimated from a small number of samples. We hope to address these limitations
in future work.
Acknowledgments
In this research, Jeremy De Bonet is supported by the DOD Multidisciplinary Research Program of the University Research Initiative, and Paul Viola by Office of
Naval Research Grant No. N00014-96-1-0311.
References
Basseville, M., Benveniste, A., Chou, K. C., Golden, S. A., Nikoukhah, R., and Will sky,
A. S. (1992). Modeling and estimation of multiresolution stochastic processes. IEEE
Transactions on Information Theory, 38(2) :766-784.
Buccigrossi, R. W. and Simoncelli, E. P. (1997). Progressive wavelet image coding based
on a conditional probability model. In Proceedings ICASSP-97, Munich, Germany.
Chellappa, R . and Chatterjee, S. (1985). Classification of textures using gaussian markov
random fields. In Proceedings of the International Joint Conference on Acoustics,
Speech and Signal Processing, volume 33, pages 959-963.
Dayan, P., Hinton, G., Neal, R., and Zemel, R. (1995). The helmholtz machine. Neural
Computation, 7:1022-1037.
De Bonet, J. S. (1997). Multiresolution sampling procedure for analysis and synthesis of
texture images. In Computer Graphics. ACM SIGGRAPH.
Donoho, D. L. and Johnstone, 1. M. (1993). Adaptation to unknown smoothness via
wavelet shrinkage. Technical report, Stanford University, Department of Statistics.
Also Tech. Report 425.
Duda, R. and Hart, P. (1973). Pattern Classification and Scene Analysis. John Wiley and
Sons.
Fogel, I. and Sagi, D. (1989). Gabor filters as texture discriminator. Biological Cybernetics,
61: 103- 113.
Geman, S. and Geman, D. (1984). Stochastic relaxation, gibbs distributions, and the
bayesian restoration of images. IEEE Transactions on Pattern Analysis and Machine
Intelligence, 6:721-741.
Heeger, D. J. and Bergen, J. R. (1995). Pyramid-based texture analysis/synthesis. In
Computer Graphics Proceedings, pages 229-238.
Hinton, G., Dayan, P., Frey, B., and Neal, R. (1995). The "wake-sleep" algorithm for
unsupervised neural networks. Science, 268:1158-116l.
Simoncelli, E. P. and Adelson, E. H. (1996). Noise removal via bayesian wavelet coring.
In IEEE Third Int'l Conf on Image Processing, Laussanne Switzerland. IEEE.
Simoncelli, E. P., Freeman, W. T., Adelson, E. H., and Heeger, D. J. (1992). Shiftable
multiscale transforms. IEEE Transactions on Information Theory, 38(2):587-607.
Zhu, S. C., Wu, Y., and Mumford, D. (1996). Filters random fields and maximum entropy(frame): To a unified theory for texture modeling. To appear in Int'l Journal
of Computer Vision.
| 1450 |@word version:1 compression:1 duda:2 pick:1 reduction:2 series:1 contains:3 ours:2 rightmost:2 current:2 si:1 yet:1 must:1 john:1 realistic:1 hypothesize:1 designed:1 interpretable:1 remove:1 resampling:1 intelligence:2 generative:5 provides:1 lx:2 constructed:1 direct:1 initiative:1 fmn:1 combine:1 expected:2 nor:1 frequently:1 multi:10 freeman:1 window:1 project:1 discover:1 underlying:1 estimating:1 homepage:1 developed:1 unified:3 transformation:1 sharpening:1 suite:2 sky:1 every:1 golden:1 scaled:1 grant:1 appear:3 local:4 textural:1 sagi:2 frey:1 might:2 chose:1 condtional:1 conversely:1 co:1 limited:1 range:2 statistically:2 acknowledgment:1 yj:1 procedure:1 area:1 empirical:2 thought:1 gabor:2 convenient:1 regular:1 downsampled:1 downsamples:1 selection:1 noising:8 www:1 yt:1 go:2 straightforward:1 independently:1 ergodic:1 resolution:1 gmrf:1 insight:1 estimator:1 target:1 synthesize:1 helmholtz:3 recognition:5 element:1 continues:1 geman:4 observed:4 bottom:1 capture:3 region:1 noised:3 principled:1 substantial:1 complexity:2 trained:3 depend:1 basis:2 easily:1 joint:3 icassp:1 siggraph:1 various:3 describe:1 effective:2 chellappa:2 monte:1 artificial:1 zemel:1 choosing:1 quite:1 emerged:1 larger:1 stanford:1 otherwise:1 statistic:1 gi:2 transform:9 noisy:2 itself:1 descriptive:1 interaction:2 maximal:1 adaptation:1 neighboring:2 date:1 achieve:1 multiresolution:2 parent:6 generating:6 brodatz:1 leave:1 measured:1 strong:2 come:1 switzerland:1 drawback:1 correct:2 filter:6 stochastic:2 human:3 require:1 biological:1 great:2 perceived:1 estimation:1 vz:2 successfully:1 weighted:2 hope:1 mit:3 gaussian:5 ck:8 rather:1 shrinkage:7 varying:1 office:1 conjunction:1 jsd:1 naval:1 laussanne:1 indicates:1 likelihood:3 ivl:1 contrast:2 tech:1 chou:1 dayan:3 bergen:6 dependent:1 transformed:1 germany:1 pixel:10 overall:1 classification:5 among:1 art:1 spatial:1 field:3 once:2 never:1 construct:1 sampling:2 progressive:1 adelson:5 unsupervised:1 future:1 report:2 few:1 recognize:1 attempt:1 interest:1 highly:1 weakness:1 chain:7 accurate:1 edge:4 tree:2 iv:1 modeling:5 kurtotic:1 measuring:1 restoration:2 lattice:1 cost:1 subset:2 rare:1 snr:3 uniform:1 parametrically:1 dod:1 shiftable:1 graphic:2 dependency:1 corrupted:1 combined:1 density:3 international:1 probabilistic:3 vm:9 synthesis:6 parzen:1 containing:2 choose:1 tile:1 conf:1 return:2 jeremy:2 de:13 coding:1 coefficient:17 int:2 caused:1 explicitly:1 performed:1 view:1 observer:1 lot:1 start:1 competitive:1 square:1 accuracy:2 efficiently:1 weak:1 bayesian:2 accurately:2 carlo:1 cybernetics:1 published:1 reach:1 email:1 definition:1 energy:1 frequency:10 workstation:1 sampled:1 massachusetts:1 knowledge:1 actually:1 back:1 appears:1 higher:5 though:3 strongly:1 ergodicity:1 until:1 bonet:7 multiscale:1 overlapping:1 mode:2 multidisciplinary:1 quality:1 perhaps:1 believe:1 effect:2 requiring:1 contain:3 hence:1 spatially:2 laboratory:1 neal:2 deal:2 white:2 complete:1 vo:1 fj:4 geometrical:1 image:78 novel:4 recently:1 volume:1 nikoukhah:1 synthesized:1 significant:2 composition:1 cambridge:1 gibbs:1 ai:3 smoothness:2 similarity:1 gj:1 driven:1 termed:1 n00014:1 success:1 exploited:1 captured:2 additional:1 freely:1 redundant:1 signal:3 ii:2 multiple:2 unimodal:1 simoncelli:8 smooth:1 technical:1 match:1 cross:2 long:5 hart:2 mrf:6 involving:1 vision:2 iteration:1 histogram:1 pyramid:8 achieved:7 preserved:1 fine:1 wake:1 db:3 contrary:1 incorporates:1 near:1 easy:1 variety:1 independence:1 fm:2 vik:1 speech:1 cause:7 transforms:2 http:1 generate:2 shifted:1 notice:2 estimated:4 algorithmically:1 l2m:3 group:1 redundancy:3 nevertheless:1 threshold:1 relaxation:1 inverse:4 everywhere:1 striking:1 almost:1 wu:2 patch:1 draw:1 vex:2 fl:1 resampled:1 sleep:1 occur:1 constraint:1 scene:1 nearby:1 anyone:1 extremely:1 conjecture:2 influential:1 structured:1 according:1 alternate:1 munich:1 department:1 across:4 slightly:3 smaller:1 son:1 discus:1 tiling:1 operation:1 hierarchical:1 generic:3 enforce:1 occurrence:1 original:6 assumes:1 top:4 build:1 especially:1 arrangement:1 mumford:2 parametric:6 primary:1 dependence:1 modeled:3 ratio:1 difficult:1 ingle:1 implementation:1 unknown:1 perform:1 observation:3 convolution:3 markov:3 viola:6 hinton:3 frame:1 sharp:1 competence:1 discriminator:1 fogel:2 acoustic:1 address:1 able:1 usually:1 pattern:2 program:1 including:1 power:2 overlap:1 critical:1 natural:10 fo1:1 residual:7 zhu:2 technology:2 created:1 prior:2 l2:6 removal:1 generation:7 limitation:1 lv:1 localized:1 coring:2 consistent:1 bank:1 benveniste:1 uncorrelated:1 share:1 row:1 supported:1 buccigrossi:2 formal:1 institute:1 johnstone:4 dimension:1 forward:1 collection:2 made:1 far:1 transaction:3 approximate:1 contradict:1 assumed:3 iterative:1 reasonably:1 complex:3 pk:4 main:2 border:1 noise:10 paul:2 arise:1 denoise:4 fashion:1 wiley:1 position:1 heeger:7 vim:1 third:1 wavelet:31 minute:1 specific:2 explored:1 evidence:1 texture:33 chatterjee:2 entropy:2 smoothly:1 simply:1 likely:1 appearance:1 visual:1 ivm:1 relies:1 acm:1 ma:1 conditional:8 endeavor:1 donoho:4 careful:1 basseville:2 shared:1 reducing:2 averaging:1 pas:1 discriminate:1 inability:1 dissimilar:1 |
494 | 1,452 | Radial Basis Functions:
a Bayesian treatment
David Barber*
Bernhard Schottky
Neural Computing Research Group
Department of Applied Mathematics and Computer Science
Aston University, Birmingham B4 7ET, U.K.
http://www.ncrg.aston.ac.uk/
{D.Barber,B.Schottky}~aston.ac.uk
Abstract
Bayesian methods have been successfully applied to regression and
classification problems in multi-layer perceptrons. We present a
novel application of Bayesian techniques to Radial Basis Function
networks by developing a Gaussian approximation to the posterior
distribution which, for fixed basis function widths, is analytic in
the parameters. The setting of regularization constants by crossvalidation is wasteful as only a single optimal parameter estimate
is retained. We treat this issue by assigning prior distributions to
these constants, which are then adapted in light of the data under
a simple re-estimation formula.
1
Introduction
Radial Basis Function networks are popular regression and classification tools[lO].
For fixed basis function centers, RBFs are linear in their parameters and can therefore be trained with simple one shot linear algebra techniques[lO]. The use of
unsupervised techniques to fix the basis function centers is, however, not generally
optimal since setting the basis function centers using density estimation on the input
data alone takes no account of the target values associated with that data. Ideally,
therefore, we should include the target values in the training procedure[7, 3, 9]. Unfortunately, allowing centers to adapt to the training targets leads to the RBF being
a nonlinear function of its parameters, and training becomes more problematic.
Most methods that perform supervised training of RBF parameters minimize the
?Present address: SNN, University of Nijmegen, Geert Grooteplein 21, Nijmegen, The
Netherlands. http://wwv.mbfys.kun.nl/snn/ email: davidb<llmbfys.kun.nl
Radial Basis Functions: A Bayesian Treatment
403
training error, or penalized training error in the case of regularized networks[7, 3, 9].
The setting of the associated regularization constants is often achieved by computationally expensive approaches such as cross-validation which search through a set
of regularization constants chosen a priori. Furthermore, much of the information
contained in such computation is discarded in favour of keeping only a single regularization constant. A single set of RBF parameters is subsequently found by minimizing the penalized training error with the determined regularization constant. In
this work, we assign prior distributions over these regularization constants, both for
the hidden to output weights and the basis function centers. Together with a noise
model, this defines an ideal Bayesian procedure in which the beliefs expressed in the
distribution of regularization constants are combined with the information in the
data to yield a posterior distribution of network parameters[6]. The beauty of this
approach is that none of the information is discarded, in contrast to cross-validation
type procedures. Bayesian techniques applied to such non-linear, non-parametric
models, however, can also be computationally extremely expensive, as predictions
require averaging over the high-dimensional posterior parameter distribution. One
approach is to use Markov chain Monte Carlo techniques to draw samples from the
posterior[8]. A simpler approach is the Laplace approximation which fits a Gaussian
distribution with mean set to a mode of the posterior, and covariance set to the
inverse Hessian evaluated at that mode. This can be viewed as a local posterior
approximation, as the form of the posterior away from the mode does not affect the
Gaussian fit. A third approach, called ensemble learning, also fits a Gaussian, but is
based on a less local fit criterion, the Kullback-Leibler divergence[4, 5]. As shown in
[1], this method can be applied successfully to multi-layer perceptrons, whereby the
KL divergence is an almost analytic quantity in the adaptable parameters. For fixed
basis function widths, the KL divergence for RBF networks is completely analytic
in the adaptable parameters, leading to a relatively fast optimization procedure.
2
Bayesian Radial Basis Function Networks
For an N dimensional input vector x, we consider RBFs that compute the linear
combination of K Gaussian basis functions,
K
f(x, m) =
L
WI
exp {-Atllx -
ctll2}
(1)
1=1
where we denote collectivel~ the ce,nters Cl ..? cKl and wei~hts w = w~ ... Wk. by
the parameter vector m = [c I , ... ,CK,Wl, ... ,WK]. We consIder the basIS functIOn
widths AI, ... Ak to be fixed although, in principle, they can also be adapted by a
similar technique to the one presented below. The data set that we wish to regress
is a set of P input-output pairs D = {xl', yJL, JL = 1 ... Pl. Assuming that the target
outputs y have been corrupted with additive Gaussian noise of variance (3-I, the
likelihood of the data is l
(2)
p(Dlm,{3) = exp (-{3ED) jZD,
where the training error is defined,
1
ED =
p
'2 L
(f(x JL , m) - yJL)2
(3)
1'=1
To discourage overfitting, we choose a prior regularizing distribution for m
p(mla) = exp (-Em(m)) jZp
lIn the following, ZD, Zp and ZF are normalising constants
(4)
D. Barber and B. Schottky
404
where we take Em(m) = ~mTAm for a matrix A of hyperparameters. More complicated regularization terms, such as those that penalize centers that move away from
specified points are easily incorporated in our formalism . For expositional clarity,
we deal here with only the simple case of a diagonal regularizer matrix A = aI.
The conditional distribution p(mID, a,.8) is then given by
p(mID,a,.8) = exp(-.8ED(m) - Em(m?/ZF
We choose to model the hyperparameters a and .8 by Gamma distributions,
p(a) ex: aa-le-o:/b
p(.8) ex: a c - 1 e-/3/d ,
(5)
(6)
where a, b, c, d are chosen constants. This completely specifies the joint posterior,
p(m,a,.BID) = p(mID,a,.8)p(a)p(.B) .
(7)
A Bayesian prediction for a new test point x is then given by the posterior average
(f(x, m?)p(In,o:,/3I D). If the centers are fixed, p(wID, a,.8) is Gaussian and computing the posterior average is trivial. However, with adaptive centers,the posterior
distribution is typically highly complex and computing this average is difficult 2 . We
describe below approaches that approximate the posterior by a simpler distribution
which can then be used to find the Bayesian predictions and error bars analytically.
3
Approximating the posterior
3.1
Laplace's method
Laplace's method is an approximation to the Bayesian procedure that fits a Gaussian
to the mode mo of p(m, ID,a,.8) by extremizing the exponent in (5)
~llmW + .8ED(m)
T =
(8)
with respect to m. The mean of the approximating distribution is then set to the
mode roo, and the covariance is taken to be the inverse Hessian around roo; this is
then used to approximately compute the posterior average. This is a local method
as no account is taken for the fit of the Gaussian away from the mode.
3.2
Kullback-Leibler method
The Kullback-Leibler divergence between the posterior p(m, a, .8ID) and an approximating distribution q(m, a,.8) is defined by
KL[q] = -
J
q(m,a,.B) In
(
p(m,a,.8ID))
q(m,a,.8)
.
(9)
K L[q] is zero only if p and q are identical, and is greater than zero otherwise. Since
in (5) ZF is unknown, we can compute the KL divergence only up to an additive
constant, L[q] = KL[q] - InZF. We seek then a posterior approximation of the
form q(m, a,.8) = Q(m)R(a)S(B) where Q(m) is Gaussian and the distributions R
and S are determined by minimization of the functional L[q][5].
We first consider optimizing L with respect to the mean m and covariance C of
the Gaussian distribution Q(m) ex: exp {-~(m - m)TC-l(m - m)}. Omitting all
constant terms and integrating out a and .8, the Q(m) dependency in Lis,
L[Q(m)] = -
J
Q(m)
[-i3ED(m) -
~Q:llmW -In Q(m)] dm + const.
(10)
2The fixed and adaptive center Bayesian approaches are contrasted more fully in [2].
Radial Basis Functions: A Bayesian Treatment
405
where
Q
=J
E= J
aR(a)da,
f3S(f3)df3
(11)
are the mean values of the hyperparameters. For Gaussian basis functions, the
remaining integration in (10) over Q(m) can be evaluated analytically, giving 3
L[Q(m)] =
~Q {tr(C) + IImW} + E(ED(m?)Q - ~ In(det C) + const.
(12)
where
(ED(m))Q
p (
= "21 ~
(y~)2 -
K
2y~ ~ sj +
J;1K)
s~1
(13)
The analytical formulae for
sj =
=
S~,
(wlexp{-'\jllxl'-CIW})Q
(14)
(WkWj exp{ -,\kllxtL - Ck 112} exp{ -'\l\lx~ - cdI 2})Q
(15)
are straightforward to compute, requiring only Gaussian integration[2]. The values
for C and m can then be found by optimizing (12).
We now turn to the functional optimisation of (9) with respect to R. Integrating
out m and f3 leaves, up to a constant,
L[R] =
JR(a){a["~W + tr~C) +t] + [K(~+l) +a-l]lna+lnR(a)}da
(16)
As the first two terms in (16) constitute the log of a Gamma distribution (6), the
functional (16) is optimized by choosing a Gamma distribution for a,
(17)
with
K(N
+ 1)
! = IIml12
s
2
The same procedure for S(f3) yields
r=
2
+ a,
+ !tr(C) + !
2
a=rs.
b'
S(f3) ex: f3 U - 1 e-/3/v
with
P
= -2
1
=
(19)
1
=
(ED(m?)Q + -d'
-:e uv,
(20)
v
where the averaged training error is given by (13). The optimization of the approximating distribution Q(m)R(a)S(f3) can then be performed using an iterative
procedure in which we first optimize (12) with respect to m and C for fixed a, 73,
and then update a and 7J according to the re-estimation formulae (18,20).
u
+ c,
(18)
After this iterative procedure has converged, we have an approximating distribution
of parameters, both for the hidden to output weights and center positions (figure
l(a?). The actual predictions are then given by the posterior average over this distribution of networks. The model averaging effect inherent in the Bayesian procedure
produces a final function potentially much more complex than that achievable by a
single network.
A significant advantage of our procedure over the Laplace procedure is that we can
lower bound model the likelihood Inp(Dlmodel) ~ -(L+lnZD+lnZp). Hence,
decreasing L increases p(Dlmodel). We can use this bound to rank different models,
leading to principled Bayesian model selection.
3( ... }Q denotes
J Q(m) ... dm
D. Barber and B. Schottky
406
Center Fluctuations
Widths
Figure 1: Regressing a surface from 40 noisy training examples. (a) The KL approximate Bayesian treatment fits 6 basis functions to the data. The posterior
distribution for the parameters gives rise to a posterior weighted average of a distribution of the 6 Gaussians. We plot here the posterior standard deviation of the
centers (center fluctuations) and the mean centers. The widths were fixed a priori
using Maximum Likelihood. (b) Fixing a basis function on each training point with
fixed widths. The hidden-output weights were determined by cross-validation of the
penalised training error.
4
Relation to non-Bayesian treatments
One non-Bayesian approach to training RBFs is to minimze the training error (3)
plus a regularizing term of the form (8) for fixed centers[7, 3, 9]. In figure l(b)
we fix a center on each training input. For fixed hyperparameters a and (3, the
optimal hidden-to-output weights can then be found by minimizing (8). To set the
hyperparameters, we iterate this procedure using cross-validation. This results in a
single estimate for the parameters mo which is then used for predictions f(x, mo).
In figure(l), both the Bayesian adaptive center and the fixed center methods have
similar performance in terms of test error on this problem. However, the parsimonious representiation of the data by the Bayesian adaptive center method may be
advantageous if interpreting the data is important.
In principle, in the Bayesian approach, there is no need to carry out a crossvalidation type procedure for the regularization parameters a, {3. After deciding
on a particular Bayesian model with suitable hyperprior constants (here a, b, c, d),
our procedure will combine these beliefs about the regularity of the RBF with the
dataset in a principled manner, returning a-posteriori probabilities for the values of
the regularization constants. Error bars on the predictions are easily calculated as
the posterior distribution quantifies our uncertainty in the parameter estimates.
One way of viewing the connection between the CV and Bayesian approaches, is to
identify the a-priori choice of CV regularization coefficients ai that one wishes to
examine as a uniform prior over the set {ad. The posterior regularizer distribution
is then a delta peak centred at that a* with minimal CV error. This delta peak
represents a loss of information regarding the performance of all the other networks
trained with ai :la*. In contrast, in our Bayesian approach we assign a continuous
prior distribution on a, which is updated according to the evidence in the data. Any
loss of information then occurs in approximating the resulting posterior distribution.
Radial Basis Functions: A Bayesian Treatment
(a) Minimum KL
407
(b) Laplace
(c) Regularized (non Bayesian)
0.8
0.8
0.8
0.&
0.&
0.8
0.'
0.'
0.'
0.2 " , , '
0.2
",
"
-0.&
\
0.2
, ,
. \, .
~ .&
"
\
'.,
:
~.&
Figure 2: Minimal KL Gaussian fit, Laplace Gaussian, and a non-Bayesian procedure on regressing with 6 Gaussian basis functions. The training points are labelled
by crosses and the target function 9 is given by the solid lines. For both (a) and (b),
the mean prediction is given by the dashed lines, and standard errors are given by
the dots. (a) Approximate Bayesian solution based on Kullback-Leibler divergence.
The regularization constant Q and inverse noise level f3 are adapted as described
in the text. (b) Laplace method based on equation (8). Both Q and f3 are set to
the mean of the hyperparameter distributions (6). The mean prediction is given
by averaging over the locally approximated posterior. Note that the error bars are
somewhat large, suggesting that the local posterior mass has been underestimated.
(c) The broken line is the Laplace solution without averaging over the posterior,
showing much greater variation than the averaged prediction in (b). The dashed line
corresponds to fixing the basis function centers at each data point, and estimating
the regularization constants a by cross-validation.
5
Demonstration
We apply the above outlined Bayesian framework to a simple one-dimensional regression problem. The function to be learned is given by
(21)
and is plotted in figure(2). The training patterns are sampled uniformly between [-4,4] and the output is corrupted with additive Gaussian noise of variance
a 2 = 0.005. The number of basis function is K = 6, giving a reasonably flexible model for this problem. In figure(2), we compare the Bayesian approaches
(a),(b) to non-Bayesian approaches(c). In this demonstration, the basis function
widths were chosen by penalised training error minimization and fixed throughout all experiments. For the Bayesian procedures, we chose hyperprior constants,
a = 2, b = 1/4, c = 4, d = 50, corresponding to mean values it = 0.5 and E= 200.
In (c), we plot a more conventional approach using cross-validation to set the regularization constant.
A useful feature of the Bayesian approaches lies in the principled theory for the error
bars. In (c), although we know the test error for each regularization constant in the
set of constants we choose to examine, we do not know any principled procedure
for using these values for error bar assessment.
408
6
D. Barber and B. Schottky
Conclusions
We have incorporated Radial Basis FUnctions within a Bayesian framework, arguing
that the selection of regularization constants by non-Bayesian methods such as
cross-validation is wasteful of the information contained in our prior beliefs and
the data set. Our framework encompasses flexible priors such as hard assigning a
basis function center to each data point or penalizing centers that wander far from
pre-assigned points. We have developed an approximation to the ideal Bayesian
procedure by fitting a Gaussian distribution to the posterior based on minimizing
the Kullback-Leibler divergence. This is an objectively better and more controlled
approximation to the Bayesian procedure than the Laplace method. FUrthermore,
the KL divergence is an analytic quantity for fixed basis function widths. This
framework also includes the automatic adaptation of regularization constants under
the influence of data and provides a rigorous lower bound on the likelihood of the
model.
Acknowledgements
We would like to thank Chris Bishop and Chris Williams for useful discussions. BS
thanks the Leverhulme Trust for support (F /250/K).
References
[1] D. Barber and C. M. Bishop. On computing the KL divergence for Bayesian Neural
Networks. Technical report, Neural Computing Research Group, Aston University,
Birmingham, 1998. See also D. Barber and C. M. Bishop These proceedings.
[2] D. Barber and B. Schottky. Bayesian Radial Basis Functions. Technical report,
Neural Computing Research Group, Aston University, Birmingham, 1998.
[3] C. M. Bishop. Improving the Generalization Properties of Radial Basis Function
Networks. Neural Computation, 4(3):579-588, 1991.
[4] G. E. Hinton and D. van Camp. Keeping neural networks simple by minimizing
the description length of the weights. In Proceedings of the Seventh Annual ACM
Workshop on Computational Learning Theory (COLT '93), 1993.
[5J D. J. C. MacKay. Developments in probabilistic modelling with neural networks ensemble learning. In Neural Networks: Artificial Intelligence and Industrial Applications. Proceedings of the 3rd Annual Symposium on Neural Networks, Nijmegan,
Netherlands, 14-15 September 1995, pages 191-198. Springer.
[6] D. J. C. MacKay. Bayesian Interpolation. Neural Computation, 4(3):415-447, 1992.
[7] J. Moody and C. J . Darken. Fast Learning in Networks of Locally-Tuned Processing
Units. Neural Computation, 1:281-294, 1989.
[8] Neal, R. M. Bayesian Learning for Neural Networks. Springer, New York, 1996.
Lecture Notes in Statistics 118.
[9J M. J . L. Orr. Regularization in the Selection of Radial Basis Function Centers. Neural
Computation, 7(3) :606-623, 1995.
[10] M. J . L. Orr. Introduction to Radial Basis Function Networks. Technical report,
Centre for Cognitive Science, Univeristy of Edinburgh, Edinburgh, ER8 9LW, U.K.,
1996.
| 1452 |@word achievable:1 advantageous:1 grooteplein:1 seek:1 r:1 covariance:3 tr:3 solid:1 shot:1 carry:1 tuned:1 expositional:1 assigning:2 additive:3 analytic:4 plot:2 hts:1 update:1 alone:1 intelligence:1 leaf:1 normalising:1 provides:1 lx:1 simpler:2 mla:1 symposium:1 combine:1 fitting:1 manner:1 mbfys:1 examine:2 multi:2 ciw:1 decreasing:1 snn:2 actual:1 becomes:1 estimating:1 mass:1 developed:1 returning:1 uk:2 unit:1 local:4 treat:1 ak:1 id:3 fluctuation:2 interpolation:1 approximately:1 plus:1 chose:1 averaged:2 arguing:1 procedure:19 pre:1 radial:12 integrating:2 inp:1 lnr:1 selection:3 influence:1 www:1 optimize:1 conventional:1 center:23 straightforward:1 williams:1 geert:1 variation:1 laplace:9 updated:1 target:5 expensive:2 approximated:1 principled:4 broken:1 ideally:1 trained:2 algebra:1 basis:29 cdi:1 completely:2 easily:2 joint:1 regularizer:2 fast:2 describe:1 monte:1 artificial:1 choosing:1 roo:2 otherwise:1 objectively:1 statistic:1 noisy:1 final:1 advantage:1 analytical:1 adaptation:1 ckl:1 description:1 yjl:2 crossvalidation:2 regularity:1 zp:1 produce:1 ac:2 fixing:2 lna:1 subsequently:1 wid:1 viewing:1 require:1 assign:2 fix:2 generalization:1 pl:1 around:1 exp:7 deciding:1 mo:3 estimation:3 birmingham:3 wl:1 successfully:2 tool:1 weighted:1 minimization:2 gaussian:18 ck:2 beauty:1 rank:1 likelihood:4 modelling:1 contrast:2 rigorous:1 industrial:1 camp:1 posteriori:1 typically:1 hidden:4 relation:1 issue:1 classification:2 flexible:2 colt:1 priori:3 exponent:1 development:1 univeristy:1 integration:2 mackay:2 f3:8 identical:1 represents:1 unsupervised:1 report:3 inherent:1 gamma:3 divergence:9 highly:1 regressing:2 nl:2 light:1 chain:1 hyperprior:2 re:2 plotted:1 minimal:2 formalism:1 ar:1 deviation:1 uniform:1 seventh:1 dependency:1 corrupted:2 combined:1 thanks:1 density:1 peak:2 probabilistic:1 together:1 moody:1 choose:3 cognitive:1 leading:2 li:1 account:2 suggesting:1 orr:2 centred:1 wk:2 includes:1 coefficient:1 ad:1 performed:1 complicated:1 rbfs:3 minimize:1 variance:2 ensemble:2 yield:2 identify:1 bayesian:40 none:1 carlo:1 converged:1 penalised:2 ed:7 email:1 regress:1 dm:2 associated:2 sampled:1 dataset:1 treatment:6 popular:1 adaptable:2 supervised:1 wei:1 evaluated:2 furthermore:2 jzp:1 trust:1 nonlinear:1 assessment:1 defines:1 mode:6 omitting:1 effect:1 requiring:1 regularization:18 analytically:2 hence:1 assigned:1 leibler:5 neal:1 deal:1 width:8 whereby:1 criterion:1 interpreting:1 extremizing:1 novel:1 functional:3 b4:1 ncrg:1 jl:2 significant:1 ai:4 cv:3 automatic:1 uv:1 outlined:1 mathematics:1 rd:1 centre:1 dot:1 surface:1 posterior:27 optimizing:2 minimum:1 greater:2 somewhat:1 dashed:2 technical:3 adapt:1 cross:8 lin:1 controlled:1 prediction:9 regression:3 optimisation:1 achieved:1 penalize:1 underestimated:1 ideal:2 bid:1 affect:1 fit:8 iterate:1 regarding:1 det:1 favour:1 hessian:2 york:1 constitute:1 generally:1 useful:2 netherlands:2 mid:3 locally:2 http:2 specifies:1 problematic:1 leverhulme:1 delta:2 zd:1 hyperparameter:1 df3:1 group:3 wasteful:2 clarity:1 penalizing:1 ce:1 schottky:6 inverse:3 uncertainty:1 almost:1 throughout:1 parsimonious:1 draw:1 layer:2 bound:3 annual:2 adapted:3 extremely:1 relatively:1 department:1 developing:1 according:2 combination:1 jr:1 em:3 wi:1 b:1 dlm:1 taken:2 computationally:2 equation:1 f3s:1 turn:1 know:2 gaussians:1 apply:1 away:3 denotes:1 remaining:1 include:1 const:2 giving:2 approximating:6 move:1 quantity:2 occurs:1 parametric:1 diagonal:1 september:1 thank:1 chris:2 barber:8 trivial:1 assuming:1 length:1 retained:1 minimizing:4 demonstration:2 kun:2 unfortunately:1 difficult:1 potentially:1 nijmegen:2 rise:1 unknown:1 perform:1 allowing:1 zf:3 darken:1 markov:1 discarded:2 hinton:1 incorporated:2 david:1 pair:1 kl:10 specified:1 optimized:1 connection:1 learned:1 address:1 bar:5 below:2 pattern:1 encompasses:1 belief:3 suitable:1 regularized:2 aston:5 text:1 prior:7 acknowledgement:1 wander:1 fully:1 loss:2 lecture:1 validation:7 principle:2 lo:2 penalized:2 keeping:2 van:1 edinburgh:2 calculated:1 adaptive:4 far:1 sj:2 approximate:3 bernhard:1 kullback:5 overfitting:1 iimw:1 search:1 iterative:2 continuous:1 quantifies:1 reasonably:1 improving:1 cl:1 discourage:1 complex:2 da:2 noise:4 hyperparameters:5 position:1 wish:2 xl:1 lie:1 lw:1 third:1 formula:3 bishop:4 showing:1 evidence:1 workshop:1 tc:1 expressed:1 contained:2 springer:2 aa:1 corresponds:1 acm:1 conditional:1 viewed:1 rbf:5 labelled:1 hard:1 determined:3 contrasted:1 uniformly:1 averaging:4 called:1 la:1 perceptrons:2 support:1 regularizing:2 ex:4 |
495 | 1,453 | Asymptotic Theory for Regularization:
One-Dimensional Linear Case
Petri Koistinen
Rolf Nevanlinna Institute, P.O. Box 4, FIN-00014 University of Helsinki,
Finland. Email: PetrLKoistinen@rnLhelsinkLfi
Abstract
The generalization ability of a neural network can sometimes be
improved dramatically by regularization. To analyze the improvement one needs more refined results than the asymptotic distribution of the weight vector. Here we study the simple case of
one-dimensional linear regression under quadratic regularization,
i.e., ridge regression. We study the random design, misspecified
case, where we derive expansions for the optimal regularization parameter and the ensuing improvement. It is possible to construct
examples where it is best to use no regularization.
1
INTRODUCTION
Suppose that we have available training data (Xl, Yd, .. 0' (Xn' Y n ) consisting of
pairs of vectors, and we try to predict Yi on the basis of Xi with a neural network
with weight vector w. One popular way of selecting w is by the criterion
1
(1)
n
-n L ?(Xi' Yi, w) + >..Q(w) = min!,
I
where the loss ?(x,y,w) is, e.g., the squared error Ily - g(x,w)11 2 , the function
g(., w) is the input/output function of the neural network, the penalty Q(w) is
a real function which takes on small values when the mapping g(o, w) is smooth
and high values when it changes rapidly, and the regularization parameter >.. is a
nonnegative scalar (which might depend on the training sample). We refer to the
setup (1) as (training with) regularization, and to the same setup with the choice
>.. = 0 as training without regularization. Regularization has been found to be very
effective for improving the generalization ability of a neural network especially when
the sample size n is of the same order of magnitude as the dimensionality of the
parameter vector w, see, e.g., the textbooks (Bishop, 1995; Ripley, 1996).
Asymptotic Theory for Regularization: One-Dimensional Linear Case
295
In this paper we deal with asymptotics in the case where the architecture of the
network is fixed but the sample size grows . To fix ideas, let us assume that the
training data is part of an Li.d. (independent, identically distributed) sequence
(X,Y);(Xl'Yl),(X2'Y2)"" of pairs of random vectors, i.e., for each i the pair
(Xi, Yi) has the same distribution as the pair (X, Y) and the collection of pairs is
independent (X and Y can be dependent) . Then we can define the (prediction) risk
of a network with weights w as the expected value
(2)
r(w) := IE:f(X, Y, w).
Let us denote the minimizer of (1) by Wn (.),) , and a minimizer of the risk r by
w*. The quantity r(w n (>.)) is the average prediction error for data independent of
the training sample. This quantity r(w n (>.)) is a random variable which describes
the generalization performance of the network: it is bounded below by r( w*) and
the more concentrated it is about r(w*), the better the performance . We will
quantify this concentration by a single number, the expected value IE:r(w n (>.)) . We
are interested in quantifying the gain (if any) in generalization for training with
versus training without regularization defined by
(3)
When regularization helps, this is positive.
However, relatively little can be said about the quantity (3) without specifying in
detail how the regularization parameter is determined. We show in the next section
that provided>' converges to zero sufficiently quickly (at the rate op(n- 1 / 2 )), then
IE: r(wn(O)) and IE: r(w n (>.)) are equal to leading order. It turns out, that the optimal
regularization parameter resides in this asymptotic regime. For this reason, delicate
analysis is required in order to get an asymptotic approximation for (3). In this
article we derive the needed asymptotic expansions only for the simplest possible
case: one-dimensional linear regression where the regularization parameter is chosen
independently of the training sample.
2
REGULARIZATION IN LINEAR REGRESSION
We now specialize the setup (1) to the case of linear regression and a quadratic
smoothness penalty, i.e. , we take f(x,y,w) = [y-x T wJ2 and Q(w) = wTRw, where
now y is scalar, x and w are vectors, and R is a symmetric, positive definite matrix.
It is well known (and easy to show) that then the minimizer of (1) is
(4)
wn (>') =
[
1
-1
n
~ ~ XiX! + >'R
]
1
n
~ ~ XiYi.
This is called the generalized ridge regression estimator, see, e.g., (Titterington,
1985); ridge regression corresponds to the choice R = I, see (Hoerl and Kennard,
1988) for a survey. Notice that (generalized) ridge regression is usually studied in
the fixed design case, where Xi:s are nonrandom. Further, it is usually assumed
that the model is correctly specified, i.e., that there exists a parameter such that
Y i = Xr w* + ?i , and such that the distribution of the noise term ?i does not depend
on Xi. In contrast, we study the random design, misspecified case.
Assuming that IE: IIXI1 2 < 00 and that IE: [XXT] is invertible, the minimizer of the
risk (2) and the risk itself can be written as
(5)
(6)
w* = A-lIE: [XY],
r(w) = r(w*)
+ (w -
with
A:=IE:[XXT]
w*f A(w - w*).
P. Koistinen
296
If Zn is a sequence of random variables, then the notation Zn = open-a) means
that n a Zn converges to zero in probability as n -+ 00 . For this notation and the
mathematical tools needed for the following proposition see, e.g., (Serfiing, 1980,
Ch. 1) or (Brockwell and Davis, 1987, Ch. 6).
Proposition 1 Suppose that IE: y4 < 00, IE: IIXII 4 < 00 and that A = IE: [X XTj is invertible. If,\ = op(n- I / 2), then both y'n(wn(O) -w*) and y'n(w n ('\) - w*) converge
in distribution to N (0, C), a normal distribution with mean zero and covariance
matrix C.
The previous proposition also generalizes to the nonlinear case (under more complicated conditions). Given this proposition, it follows (under certain additional conditions) by Taylor expansion that both IE:r(w n ('\)) - r(w*) and IEr(wn(O)) - r(w*)
admit the expansion f31 n -} + o( n -}) with the same constant f3I. Hence, in the
regime ,\ = op(n-I/2) we need to consider higher order expansions in order to
compare the performance of wn (,\) and wn(O).
3
ONE-DIMENSIONAL LINEAR REGRESSION
We now specialize the setting of the previous section to the case where x is scalar.
Also, from now on, we only consider the case where the regularization parameter
for given sample size n is deterministic; especially ,\ is not allowed to depend on
the training sample. This is necessary, since coefficients in the following type of
asymptotic expansions depend on the details of how the regularization parameter
is determined. The deterministic case is the easiest one to analyze.
We develop asymptotic expansions for the criterion
(7)
where now the regularization parameter k is deterministic and nonnegative. The
expansions we get turn out to be valid uniformly for k ~ O. We then develop
asymptotic formulas for the minimizer of I n , and also for In(O) - inf I n . The last
quantity can be interpreted as the average improvement in generalization performance gained by optimal level of regularization, when the regularization constant
is allowed to depend on n but not on the training sample.
From now on we take Q(w) = w 2 and assume that A = IEX2 = 1 (which could be
arranged by a linear change of variables). Referring back to formulas in the previous
section, we see that
(8)
r(wn(k)) - r(w*) = ern - kw*)2/(Un + 1 + k)2 =: h(Un , Vn, k),
whence In(k) = IE:h(Un , Vn , k), where we have introduced the function h (used
heavily in what follows) as well as the arithmetic means Un and Vn
_
(9)
1
n
L Vi,
n
Vn:= -
with
I
(10)
Vi
:=
XiYi - w* xl
For convenience, also define U := X2 - 1 and V := Xy - w* X2 . Notice that
U; UI, U2 , ? .. are zero mean Li.d. random variables, and that V; Vi, V2 ,. " satisfy
the same conditions. Hence Un and Vn converge to zero, and this leads to the idea
of using the Taylor expansion of h(u, v, k) about the point (u, v) = (0,0) in order
to get an expansion for In(k).
297
Asymptotic Theory for Regularization: One-Dimensional Linear Case
To outline the ideas, let Tj(u,v,k) be the degree j Taylor polynomial of (u,v) f-7
h(u, v, k) about (0,0), i.e., Tj(u, v, k) is a polynomial in u and v whose coefficients are functions of k and whose degree with respect to u and v is j. Then
IETj(Un,Vn,k) depends on n and moments of U and V. By deriving an upper
bound for the quantity IE Ih(Un , Vn , k) - Tj(Un , Vn , k)1 we get an upper bound for
the error committed in approximating In(k) by IE Tj(Un , Vn , k). It turns out that
for odd degrees j the error is of the same order of magnitude in n as for degree
j - 1. Therefore we only consider even degrees j. It also turns out that the error
bounds are uniform in k ~ 0 whenever j ~ 2. To proceed, we need to introduce
assumptions.
Assumption 1 IE IXlr
< 00
and IE IYls
< 00
for high enough rand s.
Assumption 2 Either (a) for some constant j3 > 0 almost surely
X has a density which is bounded in some neighborhood of zero.
IXI ;:::
j3 or (b)
Assumption 1 guarantees the existence of high enough moments; the values r = 20
and s = 8 are sufficient for the following proofs. E.g., if the pair (X, Y) has a
normal distribution or a distribution with compact support, then moments of all
orders exist and hence in this case assumption 1 would be satisfied. Without some
condition such as assumption 2, In(O) might fail to be meaningful or finite. The
following technical result is stated without proof.
Proposition 2 Let p
> 0 and let 0 < IE X 2 < 00. If assumption 2 holds, then
where the expectation on the left is finite (a) for n
assumption 2 (a), respectively 2 (b) holds.
~
1 (b) for n
> 2p provided that
Proposition 3 Let assumptions 1 and 2 hold. Then there exist constants no and
M such that
In(k) = JET2(Un , Vn , k) + R(n, k) where
_ _
(w*)2k 2
W*kIEUV]
-1 [IEV2
(w*)2k2JEU 2
IET2(Un , Vn , k) = (1+k)2 +n
(1+k)2 +3 (1+k)4 +4 (1+k)3
IR(n, k) I :s; Mn- 3/ 2(k
+ 1)-1, "In;::: no, k ;::: o.
PROOF SKETCH The formula for IE T 2 (Un , Vn. k) follows easily by integrating the
degree two Taylor polynomial term by term. To get the upper bound for R(n, k),
consider the residual
where we have omitted four similar terms. Using the bound
P. Koistinen
298
the Ll triangle inequality, and the Cauchy-Schwartz inequality, we get
r
IR(n, k)1 = IJE [h(Un , Vn , k) - T2(Un , Vn , k)]1
., (k+
{2(k
W'
{Ii: [(~ ~Xl)-']
+ 1)3[JE (lUn I2IVnI 4 )]l/2 + 4(w*)2k 2(k + 1)[18 IUn I6 ]l/2 ... }
By proposition 2, here 18 [(~ 2:~ X[)-4] = 0(1). Next we use the following fact, cf.
(Serfiing, 1980, Lemma B, p. 68).
Fact 1 Let {Zd be i.i.d. with 18 [Zd = 0 and with 18 IZI/v < 00 for some v
Then
v
~
2.
Applying the Cauchy-Schwartz inequality and this fact, we get, e.g., that
[18 (IUn I2 IVnI 4 )]l/2 ~ [(18 IUn I4 )1/2(E IVnI 8)1/2p/2 = 0(n- 3 / 2).
o
Going through all the terms carefully, we see that the bound holds.
Proposition 4 Let assumptions 1 and 2 hold, assume that w* :j; 0, and set
al := (18 V2
- 2w*E [UVD/(w*)2.
If al > 0, then there exists a constant ni such that for all n ~ nl the function
k ~ ET2(Un , Vn,k) has a unique minimum on [0,(0) at the point k~ admitting the
expanszon
k~ = aIn- 1 + 0(n- 2); further,
In(O) - inf{Jn(k) : k ~ O} = In(O) - In(aln- 1 ) = ar(w*)2n- 2
If a
~
+ 0(n- 5 / 2).
0, then
The proof is based on perturbation expansio!1 c~nsidering lin a
small parameter. By the previous proposition, Sn(k) := ET2 (Un , Vn , k) is the sum
of (w*)2k 2/(1 + k)2 and a term whose supremum over k ~ ko > -1 goes to zero
as n ~ 00. Here the first term has a unique minimum on (-1,00) at k = O.
Differentiating Sn we get
PROOF SKETCH
S~(k) = [2(w*)2k(k
+ 1)2 + n- 1 p2(k)]/(k + 1)5,
where P2(k) is a second degree polynomial in k. The numerator polynomial has
three roots, one of which converges to zero as n ~ 00. A regular perturbation
expansion for this root, k~ = aln- I + a2n-2 + ... , yields the stated formula for
al. This point is a minimum for all sufficiently large n; further, it is greater than
zero for all sufficiently large n if and only if al > O.
The estimate for J n (0) - inf {J n (k) : k ~ O} in the case al > 0 follows by noticing
that
In(O) - In(k) = 18 [h(Un , Vn , 0) - h(Un , Vn , k)),
where we now use a third degree Taylor expansion about (u, v, k) = (0,0,0)
h(u,v,O) - h(u,v,k) =
2w* kv - (w*)2k2 - 4w*kuv + 2(w*?k 2u + 2kv 2 - 4w*k 2v + 2(W*)2k 3
+ r(u, v, k).
Asymptotic Theory for Regularization: One-Dimensional Linear Case
299
0.2
0.18
0.16
0.14
0.12
0.1
__
__
__
__
__
__
__
0.05 0.1
0.15 0 .2 0.25 0.3 0.35 0.4
~~
o
~
~
~
~
~
~
L-~
__
0.45
~
0.5
Figure 1: Illustration of the asymptotic approximations in the situation of equation
(11) . Horizontal axis kj vertical axis .In(l?) and its asymptotic appr~ximations.
Legend: markers In(k); solid line IE T 2 (Un , Vn , k)j dashed line IET4 (Un , Vn , k).
Usin~ t~e techniques of the previous proposition, it can be shown that
IE Ir(Un , Vn , k~)1 = O(n- S/ 2 ). Integrating the Taylor polynomial and using this
estimate gives
In(O) - In(aI/n) = af(w*)2n- 2 + O(n- S / 2 ).
Finally, by the mean value theorem,
In(O) -inf{ In(k) : k
~ O} =
In(O) -In(aI/n) +
!
= In(O) -
(In(O) -
In(k)]lk=8(k~ -aI/n)
In(aI/n)
+ O(n-1)O(n-2)
where () lies between k~ and aI/n, and where we have used the fact that the indicated derivative evaluated at () is of order O(n- 1 ), as can be shown with moderate
effort.
0
Remark In the preceding we assumed that A = IEX 2 equals 1. If this is not
the case, then the formula for a1 has to be divided by A; again, if a1 > 0, then
k~ = a1n-1 + O(n- 2 ) .
If the model is correctly specified in the sense that Y = w* X + E, where E is
independent of X and IE E = 0, then V = X E and IE [UV] = O. Hence we have
a1 = IE [E 2]j(w*)2, and this is strictly positive expect in the degenerate case where
E = 0 with probability one. This means that here regularization helps provided the
regularization parameter is chosen around the value aI/n and n is large enough.
See Figure 1 for an illustration in the case
(11)
X "'" N(O, 1) ,
Y = w* X
+f,
f "'"
N(O, 1),
w* = 1,
where E and X are independent. In(k) is estimated on the basis of 1000 repetitions
of the task for n = 8. In addition to IE T2(Un , Vn , k) the function IE T4 (Un , lin, k)
is also plotted. The latter can be shown to give In(k) correctly up to order
O(n- s/ 2 (k+ 1)-3). Notice that although IE T2 (Un , Vn , k) does not give that good an
approximation for In(k), its minimizer is near the minimizer of In(k), and both of
these minimizers lie near the point al/n = 0.125 as predicted by the theory. In the
situation (11) it can actually be shown by lengthy calculations that the minimizer
of In(k) is exactly al/n for each sample size n ~ 1.
It is possible to construct cases where
X "'" Uniform (a, b),
Y = cjX + d+ Z,
a1
< O.
For instance, take
1
1
b=-(3Vs-l)
2'
4
c= -5,d= 8,
a=-
P. Koistinen
300
and Z '" N (0, a 2 ) with Z and X independent and 0 :::; a < 1.1. In such a case
regularization using a positive regularization parameter only makes matters worse;
using a properly chosen negative regularization parameter would, however, help in
this particular case. This would, however, amount to rewarding rapidly changing
functions. In the case (11) regularization using a negative value for the regularization parameter would be catastrophic.
4
DISCUSSION
We have obtained asymptotic approximations for the optimal regularization parameter in (1) and the amount of improvement (3) in the simple case of one-dimensional
linear regression when the regularization parameter is chosen independently of the
training sample. It turned out that the optimal regularization parameter is, to
leading order, given by Qln- 1 and the resulting improvement is of order O(n- 2 ).
We have also seen that if Ql < 0 then regularization only makes matters worse.
Also (Larsen and Hansen, 1994) have obtained asymptotic results for the optimal
regularization parameter in (1). They consider the case of a nonlinear network;
however, they assume that the neural network model is correctly specified.
The generalization of the present results to the nonlinear, misspecified case might
be possible using, e.g., techniques from (Bhattacharya and Ghosh, 1978). Generalization to the case where the regularization parameter is chosen on the basis of the
sample (say, by cross validation) would be desirable.
Acknowledgements
This paper was prepared while the author was visiting the Department for Statistics and Probability Theory at the Vienna University of Technology with financial
support from the Academy of Finland. I thank F. Leisch for useful discussions.
References
Bhattacharya, R. N. and Ghosh, J. K. (1978). On the validity of the formal Edgeworth expansion. The Annals of Statistics, 6(2):434-45l.
Bishop, C. M. (1995). Neural Networks for Pattern Recognition. Oxford University
Press.
Brockwell, P. J. and Davis, R. A. (1987). Time Series: Theory and Methods.
Springer series in statistics. Springer-Verlag.
Hoerl, A. E. and Kennard, R. W. (1988). Ridge regression. In Kotz, S., Johnson,
N. L., and Read, C. B., editors, Encyclopedia of Statistical Sciences. John Wiley
& Sons, Inc.
Larsen, J. and Hansen, L. K. (1994). Generalization performance of regularized
neural network models. In Vlontos, J., Whang, J.-N., and Wilson, E., editors,
Proc. of the 4th IEEE Workshop on Neural Networks for Signal Processing,
pages 42-51. IEEE Press.
Ripley, B. D. (1996). Pattern Recognition and Neural Networks. Cambridge University Press.
Serfiing, R. J. (1980). Approximation Theorems of Mathematical Statistics. John
Wiley & Sons, Inc.
Titterington, D. M. (1985). Common structure of smoothing techniques in statistics.
International Statistical Review, 53:141-170.
| 1453 |@word polynomial:6 open:1 covariance:1 solid:1 moment:3 series:2 selecting:1 wj2:1 xiyi:2 written:1 john:2 v:1 cjx:1 mathematical:2 specialize:2 introduce:1 expected:2 little:1 provided:3 bounded:2 notation:2 easiest:1 what:1 interpreted:1 aln:2 textbook:1 titterington:2 ghosh:2 nonrandom:1 guarantee:1 exactly:1 k2:1 schwartz:2 positive:4 oxford:1 yd:1 might:3 studied:1 specifying:1 unique:2 definite:1 edgeworth:1 xr:1 asymptotics:1 integrating:2 regular:1 get:8 convenience:1 risk:4 applying:1 ier:1 deterministic:3 go:1 independently:2 survey:1 estimator:1 deriving:1 financial:1 et2:2 annals:1 suppose:2 heavily:1 ixi:1 recognition:2 qln:1 ui:1 depend:5 basis:3 triangle:1 easily:1 wtrw:1 xxt:2 effective:1 neighborhood:1 refined:1 iun:3 whose:3 say:1 ability:2 statistic:5 itself:1 sequence:2 turned:1 rapidly:2 a2n:1 brockwell:2 degenerate:1 academy:1 kv:2 converges:3 help:3 derive:2 develop:2 odd:1 op:3 p2:2 predicted:1 quantify:1 fix:1 generalization:8 proposition:10 strictly:1 hold:5 sufficiently:3 around:1 normal:2 mapping:1 predict:1 whang:1 appr:1 finland:2 omitted:1 proc:1 hoerl:2 hansen:2 ain:1 repetition:1 tool:1 wilson:1 iex:1 improvement:5 ily:1 properly:1 contrast:1 sense:1 whence:1 serfiing:3 dependent:1 minimizers:1 going:1 interested:1 smoothing:1 equal:2 construct:2 kw:1 petri:1 t2:3 xtj:1 consisting:1 delicate:1 nl:1 admitting:1 tj:4 necessary:1 xy:2 taylor:6 plotted:1 instance:1 ar:1 zn:3 uniform:2 johnson:1 a1n:1 referring:1 density:1 international:1 ie:26 koistinen:4 yl:1 rewarding:1 invertible:2 quickly:1 squared:1 again:1 satisfied:1 worse:2 admit:1 derivative:1 leading:2 li:2 coefficient:2 matter:2 inc:2 satisfy:1 vi:3 depends:1 try:1 root:2 analyze:2 complicated:1 ir:3 ni:1 yield:1 whenever:1 email:1 lengthy:1 larsen:2 proof:5 gain:1 popular:1 dimensionality:1 carefully:1 actually:1 back:1 higher:1 improved:1 rand:1 izi:1 arranged:1 evaluated:1 box:1 sketch:2 horizontal:1 nonlinear:3 marker:1 leisch:1 indicated:1 grows:1 validity:1 y2:1 regularization:36 hence:4 read:1 symmetric:1 i2:1 deal:1 ll:1 numerator:1 davis:2 criterion:2 generalized:2 outline:1 ridge:5 misspecified:3 common:1 refer:1 cambridge:1 ai:6 smoothness:1 uv:1 i6:1 uvd:1 inf:4 moderate:1 kuv:1 certain:1 verlag:1 inequality:3 yi:3 seen:1 minimum:3 additional:1 greater:1 preceding:1 surely:1 converge:2 signal:1 dashed:1 arithmetic:1 ii:1 desirable:1 smooth:1 technical:1 af:1 calculation:1 cross:1 lin:2 divided:1 a1:4 prediction:2 j3:2 regression:11 ko:1 expectation:1 sometimes:1 addition:1 usin:1 legend:1 near:2 identically:1 wn:8 easy:1 enough:3 architecture:1 f3i:1 idea:3 effort:1 penalty:2 proceed:1 remark:1 dramatically:1 useful:1 amount:2 prepared:1 encyclopedia:1 concentrated:1 simplest:1 exist:2 notice:3 estimated:1 correctly:4 zd:2 four:1 ije:1 changing:1 sum:1 noticing:1 almost:1 kotz:1 vn:23 bound:6 quadratic:2 nonnegative:2 i4:1 helsinki:1 x2:3 min:1 relatively:1 ern:1 department:1 describes:1 son:2 equation:1 turn:4 fail:1 needed:2 generalizes:1 available:1 v2:2 bhattacharya:2 existence:1 jn:1 cf:1 vienna:1 especially:2 approximating:1 quantity:5 concentration:1 said:1 visiting:1 thank:1 ensuing:1 cauchy:2 reason:1 assuming:1 y4:1 illustration:2 setup:3 ql:1 stated:2 negative:2 design:3 upper:3 vertical:1 fin:1 finite:2 situation:2 committed:1 perturbation:2 introduced:1 pair:6 required:1 specified:3 nevanlinna:1 below:1 usually:2 pattern:2 regime:2 rolf:1 regularized:1 residual:1 mn:1 technology:1 ivni:2 axis:2 lk:1 sn:2 kj:1 review:1 acknowledgement:1 asymptotic:15 loss:1 expect:1 versus:1 validation:1 degree:8 sufficient:1 article:1 editor:2 last:1 formal:1 institute:1 differentiating:1 distributed:1 xix:1 f31:1 xn:1 valid:1 resides:1 author:1 collection:1 compact:1 iixi1:1 supremum:1 assumed:2 xi:5 ripley:2 un:24 improving:1 expansion:13 noise:1 allowed:2 kennard:2 je:1 wiley:2 lun:1 xl:4 lie:3 third:1 formula:5 theorem:2 bishop:2 exists:2 workshop:1 ih:1 iixii:1 gained:1 magnitude:2 t4:1 scalar:3 u2:1 springer:2 ch:2 corresponds:1 minimizer:8 quantifying:1 change:2 determined:2 uniformly:1 lemma:1 called:1 catastrophic:1 meaningful:1 support:2 latter:1 |
496 | 1,454 | Graph Matching with Hierarchical
Discrete Relaxation
Richard C. Wilson and Edwin R. Hancock
Department of Computer Science, University of York
York, YOl 5DD, UK.
Abstract
Our aim in this paper is to develop a Bayesian framework for matching hierarchical relational models. The goal is to make discrete label assignments so as to optimise a global cost function that draws
information concerning the consistency of match from different levels of the hierarchy. Our Bayesian development naturally distinguishes between intra-level and inter-level constraints. This allows
the impact of reassigning a match to be assessed not only at its
own (or peer) level ofrepresentation, but also upon its parents and
children in the hierarchy.
1
Introd uction
Hierarchical graphical structures are of critical importance in the interpretation of
sensory or perceptual data. For instance, following the influential work of Marr [6]
there has been sustained efforts at effectively organising and processing hierarchical
information in vision systems. There are a plethora of concrete examples which include pyramidal hierarchies [3] that are concerned with multi-resolution information
processing and conceptual hierarchies [4] which are concerned with processing at
different levels of abstraction. Key to the development of techniques for hierarchical
information processing is the desire to exploit not only the intra-level constraints
applying at the individual levels of representation but also inter-level constraints
operating between different levels of the hierarchy. If used effectively these interlevel constraints can be brought to bear on the interpretation of uncertain image
entities in such a way as to improve the fidelity of interpretation achieved by single
level means. Viewed as an additional information source, inter-level constraints can
be used to resolve ambiguities that would persist if single-level constraints alone
were used.
690
R. C. Wilson and E. R. Hancock
In the connectionist literature graphical structures have been widely used to represent probabilistic causation in hierarchical systems [5, 9]. Although this literature
has provided a powerful battery of techniques, they have proved to be of limited use
in practical sensory processing systems. The main reason for this is that the underpinning independence assumptions and the resulting restrictions on graph topology
are rarely realised in practice. In particular there are severe technical problems in
dealing with structures that contain loops or are not tree-like. One way to overcome
this difficulty is to edit intractable structures to produce tractable ones [8].
Our aim in this paper is to extend this discrete relaxation framework to hierarchical graphical structures. We develop a label-error process to model the violation of
both inter-level and intra-level constraints. These two sets of constraints have distinct probability distributions. Because we are concerned with directly comparing
the topology graphical structures rather than propagating causation, the resulting framework is not restricted by the topology of the hierarchy. In ,particular,
we illustrate the effectiveness of the method on amoral graphs used to represent
scene-structure in an image interpretation problem. This is a heterogeneous structure [2, 4] in which different label types and different classes of constraint operate
at different levels of abstraction. This is to be contrasted with the more familiar
pyramidal hierarchy which is effectively homogeneous [1, 3]. Since we are dealing with discrete entities inter-level information communication is via a symbolic
interpretation of the objects under consideration.
2
Hierarchical Consistency
The hierarchy consists of a number of levels, each containing objects which are
fully described by their children at the level below. Formally each level is described
by an attributed relational graph G I = (Vi, EI, Xl), Vi E L, with L being the
index-set of levels in the hierarchy; the indices t and b are used to denote the
top and bottom levels of the hierarchy respectively. According to our notation for
level i of the hierarchy, Vi is the set of nodes, EI is the set of intra-level edges
and Xl = {~~, Vu E Vi} is a set of unary attributes residing on the nodes. The
children or descendents which form the representation of an element j at a lower
level are denoted by V j . In other words, if U l - I is in Vj then there is a link in the
hierarchy between element j at level i and element u at level i-I. According to
our assumptions, the elements of Vj are drawn exclusively from Vi-I. The goal of
performing relaxation operations is to find the match between scene graph G 1 and
model graph G 2 ? At each individual level of the hierarchy this match is represented
by a mapping function p, Vi E L, where II: Vi -t Vi.
The development of a hierarchical consistency measure proceeds along a similar
line to the Single-level work of Wilson and Hancock [10]. The quantity of interest is
the MAP estimate for the mapping function I given the available unary attributes,
i.e. I = argmaxj P(jt, Vi E LIX I , Vi E L). We factorize the measurement information over the set of nodes by application of Bayes rule under the assumption of
measurement independence on the nodes. As a result
P(/, Vi E L/X l , Vi E L) =
(Xl
p,
~i E L) {II
II p(X~ll(u))}P(fI, Vi E L)
(1)
IELuEVI
The critical modelling ingredient in developing a discrete relaxation scheme from the
above MAP criterion is the joint prior for the mapping function, i.e. p(fl, Vi E L)
691
Graph Matching with Hierarchical Discrete Relaxation
Parents
Children
A
Possible mappings or children:
1,2,3 --
B
c
A,B,C
C,B,A
Figure 1: Example constrained children mappings
which represents the influence of structural information on the matching process.
The joint measurement density, p(XI, 'VI E L), on the other hand is a fixed property
of the hierarchy and can be eliminated from further consideration.
Raw perceptual information resides on the lowest level of the hierarchy. Our task
is to propagate this information upwards through the hierarchy. To commence our
development, we assume that individual levels are conditionally dependent only on
the immediately adjacent descendant and ancestor levels. This assumption allows
the factorisation of the joint probability in a manner analogous to a Markov chain
[3]. Since we wish to draw information from the bottom upwards, the factorisation commences from the highest level of labelling. The expression for the joint
probability of the hierarchical labelling is
p(fl, 'VI E L) = p(fb)
II
P(fl+Ill)
(2)
IEL,I#t
We can now focus our attention on the conditional probabilities P(fI+1lfl). These
quantities express the probability of a labelling at the level I + 1 given the previously
defined labelling at the descendant level l. We develop tractable expressions for
these probabilities by decomposing the hierarchical graph into convenient structural
units. Here we build on the idea of decomposing Single-level graphs into supercliques that was successfully exploited in our previous work [10]. Super-cliques are
the sets of nodes connected to a centre-object by intra-level edges. However, in the
hierarchical case the relational units are more complex since we must also consider
the graph-structure conveyed by inter-level edges.
We follow the philosophy adopted in the single-level case [10] by averaging the superclique probabilities to estimate the conditional matching probabilities P(fI+1lfl).
If r~ C fl denotes the current match of the super-clique centred on the object
j E ~l then we write
P(f'lfl-I) =
~l L p(r~lfl-l)
I I jEV'
(3)
In order to model this probability, we require a dictionary of constraint relations for
the corresponding graph sub-units (super-cliques) from the model graph G 2 ? The
allowed mappings between the model graph and the data graph which preserve the
topology of the graph structure at a particular level of representation are referred
692
R. C. Wilson and E R. Hancock
to as "structure preserving mappings" or SPM's. It is important to note that we
need only explore those mappings which are topologically identical to the superclique centred on object j and therefore the possible mappings of the child nodes are
heavily constrained by the mappings of their parents (Figure 1). We denote the set
of SPM's by P. Since the set P is effectively the state-space of legal matching, we
can apply the Bayes theorem to compute the conditional super-clique probability
in the following manner
p(r~I/I-l) =
2: p(r~15,/I-l)P(5Ii-l)
(4)
SEP
According to this expression, there are two distinct components to our model. The
first involves the comparison between our mapped realisation of the super-clique
from graph G 1 , i.e.
with the selected unit from graph G 2 and the mapping
from level 1- 1. Here we take the view that once we have hypothesised a particular
mapping 5 from P, the mapping P-l provides us with no further information, i.e.
p(r~ 15, /1-1) = p(r~ 15). The matched super-clique r~ is conditionally independent
given a mapping from the set of SPM's and we may write the first term as p(r~15).
In other words, this first conditional probability models only intra-level constraints.
The second term is the significant one in evaluating the impact inter-level constraints
on the labelling at the previous level. In this term the probability of the hypothesised
mapping 5 is conditioned according to the match of the child levell.
q,
All that remains now is to evaluate the conditional probabilities. Under the assumption of memoryless matching errors, the first term may be factorised over
the marginal probabilities for the assigned matches lIon the individual nodes of
the matched super-clique
given their counterparts Si belonging to the structure
preserving mapping 5. In other words,
q
p(r;15)
=
II
P('~lsi)
(5)
1'! Ef~
In order to proceed we need to specify a probability distribution for the different
matching possibilities. There are three cases. Firstly, the match Ii may be to a
dummy-node d inserted into
to raise it to the same size as 5 so as to facilitate
comparison. This process effectively models structural errors in the data-graph.
The second and third cases, relate to whether the match is correct or in error.
Assuming that dummy node insertions may be made with probability Ps and that
matching errors occur with probability Pe , then we can write down the following
distribution rule
q
if Ii = d or
1'f Ii1 = Si
otherwise
Si
= d
(6)
The second term in Equation (5) is more subtle; it represents the conditional probability of the SPM 5 given a previously determined labelling at the level below.
However, the mapping contains labels only from the current levell, not labels from
level I - 1. We can reconcile this difference by noting that selection of a particular
mapping at level I limits the number of consistent mappings allowed topologically
at the level below. In other words if one node is mapped to another at level I,
Graph Matching with Hierarchical Discrete Relaxation
693
the consistent interpretation is that the children of the nodes must match to each
other. Provided that a set of mappings is available for the child-nodes, then this
constraint can be used to model P(SljI-1). The required child-node mappings are
referred to as "Hierarchy Preserving Mappings" or HPM's . It is these hierarchical
mappings that lift the requirements for moralization in our matching scheme, since
they effectively encode potentially incestuous vertical relations. We will denote the
set of HPM's for the descendants of the SPM S as Qs and a member of this set as
Q = {qi, 'Vi E Vj}. Using this model the conditional probability P(SIfI-l) is given
by
p(SIfI-1)
L
=
P(SIQ,/I-1)P(Qll- 1 )
(7)
QEQs
Following our modelling of the intra-level probabilities, in this inter-level case assume that S is conditionally independent of 11- 1 given Q, i.e. P(SIQ, / 1- 1) =
P(SIQ)?
Traditionally, dictionary based hierarchical schemes have operated by using a labelling determined at a preceding level to prune the dictionary set by elimination
of vertically inconsistent items [4]. This approach can easily be incorporated into
our scheme by setting P(QI/I-l) equal to unity for consistent items and to zero for
those which are inconsistent. However we propose a different approach; by adopting the same kind of label distribution used in Equation 6 we can grade the SPM's
according to their consistency with the match at level 1 - 1, i.e. jI-l. The model
is developed by factorising over the child nodes qi E Q in the following manner
P(Qll- 1 ) =
II P(qih,!-1)
(8)
qiEQ
The conditional probabilities are assigned by a re-application of the distribution
rule given in Equation (6), i.e.
if dummy node match
'f
1-1
1 qi = Ii
otherwise
(9)
For the conditional probability of the SPM given the HPM Q, we adopt a simple
uniform model under the assumption that all legitimate mappings are equivalent,
i.e. P(SIQ) = P(S) = I~I'
The various simplifications can be assembled along the lines outlined in [10] to
develop a discrete update rule for matching the two hierarchical structures. The
MAP update decision depends only on the label configurations residing on levels
1 - 1, 1 and 1 + 1 together with the measurements residing on levell. Specifically,
the level 1 matching configuration satisfies the condition
II = argm!F{
f
II p(~~ljt(j)) }P(fI-lli )P(PI1 +
l
1 1)
(10)
jEV!
Here consistency of match between levels land 1 - 1 of the hierarchy is gauged by
R. C. Wilson and E. R. Hancock
694
the quantity
PUl-llf l ) =
:1 L
L
1
L
+ ks~(rL S))]
K(rD
Qs
exp
[-(keH(r LS)
K(r!-l)
exp
[- (keH (r~-l ,Q)
1 iEVI SEP
+ ks ~(r~-l, Q)) }11)
QEQs
In the above expression H (r j, S) is the "Hamming distance" which counts the number of label conflicts between the assigned match rj and the structure preserving
mapping S. This quantity measures the consistency of the matched labels. The
number of dummy nodes inserted into r j by the mapping S is denoted by ~ (r j, S).
This second quantity measures the structural compatibility of the two hierarchical
graphs. The exponential constants ke = In (l-PeMI-P,) and ks = In 1Ft, are related to the probabilities of structural errors and mis-assignment errors. Finally,
K(rj) = (1- P e ){1- Pe)lrjl is a normalisation constant. Finally, it is worth pointing out that the discrete relaxation scheme of Equation (10) can be applied at any
level in the hierarchy. In other words the process can be operated in top-down or
bottom-up modes if required.
3
Matching SAR Data
In our experimental evaluation of the discrete relaxation scheme we will focus on
the matching of perceptual groupings of line-segments in radar images. Here the
model graph is elicited from a digital map for the same area as the radar image. The
line tokens extracted from the radar data correspond to hedges in the landscape.
These are mapped as quadrilateral field boundaries in the cartographic model. To
support this application, we develop a hierarchical matching scheme based on linesegments and corner groupings. The method used to extract these features from
the radar images is explained in detail in [10]. Straight line segments extracted
from intensity ridges are organised into corner groupings. The intra-level graph
is a constrained Delaunay triangulation of the line-segments. Inter-level relations
represent the subsumption of the bottom-level line segments into corners.
The raw image data used in this study is shown in Figure 2a. The extracted linesegments are shown in Figure 2c. The map used for matching is shown in Figure
2b. The experimental matching study is based on 95 linear segments in the SAR
data and 30 segments contained in the map. However only 23 of the SAR segments
have feasible matches within the map representation. Figure 2c shows the matches
obtained by non-hierarchical means. The lines are coded as follows; the black lines
are correct matches while the grey lines are matching errors. With the same coding
scheme Figure 2d. shows the result obtained using the hierarchical method outlined
in this paper. Comparing Figures 2c and 2d it is clear that the hierarchical method
has been effective at grouping significant line structure and excluding clutter. To
give some idea of relative performance merit, in the case of the non-hierarchical
method, 20 of the 23 matchable segments are correctly identified with 75 incorrect
matches. Application of the hierarchical method gives 19 correct matches, only 17
residual clutter segments with 59 nodes correctly labelled as clutter.
4
Concl usions
We have developed graph matching technique which is tailored to hierarchical relational descriptions. The key element is this development is to quantify the match-
Graph Matching with Hierarchical Discrete Relaxation
,
a)
b)
\
"
~\
695
-
"
-
"\
'
/\{-\'\("
'\
c)
\
d)
Figure 2: Graph editing: a) Original image, b) Digital map, c) Non hierarchical
match, d) Hierarchical match.
ing consistency using the concept of hierarchy preserving mappings between two
graphs. Central to the development of this novel technique is the idea of computing
the probability of a particular node match by drawing on the topologically allowed
mappings of the child nodes in the hierarchy. Results on image data with lines and
corners as graph nodes reveal that the technique is capable of matching perceptual
groupings under moderate levels of corruption.
References
[1] F. Cohen and D. Cooper. Simple Parallel Hierarchical and Relaxation Algorithms for
Segmenting Non-Causal Markovian Random Fields. IEEE PAMI, 9, 1987, pp.195219.
[2] L. Davis and T. Henderson. Hierarchical Constraint Processes for Shape Analysis.
IEEE PAMI, 3, 1981, pp.265-277.
[3] B. Gidas. A Renormalization Group Approach to Image Processing Problems. IEEE
PAMI, 11, 1989, pp.164-180.
[4] T . Henderson. Discrete Relaxation Techniques . Oxford University Press, 1990.
[5] D.J. Spiegelhalter and S.L. Lauritzen, Sequential updating of conditional probabilities
on directed Graphical structures, Networks, 1990, 20, pp.579-605.
[6] D. Marr, Vision. W.H. Freeman and Co., San Francisco.
[7] J. Pearl, Probabilistic Reasoning in Intelligent Systems, Morgan Kaufmann, 1988.
[8] M. Meila and M. Jordan, Optimal triangulation with continuous cost functions, Advances in Neural Information Processing Systems 9, to appear 1997.
[9] P.Smyth, D. Heckerman, M.1. Jordan, Probabilistic independence networks for hidden
Markov probability models, Neural Computation, 9, 1997, pp. 227-269.
[10] R.C . Wilson and E . R. Hancock, Structural Matching by Discrete Relaxation. IEEE
PAMI, 19, 1997, pp .634- 648.
IEEE PAMI, June 1997.
| 1454 |@word grey:1 propagate:1 configuration:2 contains:1 exclusively:1 quadrilateral:1 current:2 comparing:2 si:3 must:2 shape:1 update:2 alone:1 selected:1 item:2 argm:1 provides:1 node:20 organising:1 firstly:1 along:2 descendant:3 consists:1 sustained:1 incorrect:1 manner:3 inter:9 multi:1 grade:1 freeman:1 resolve:1 provided:2 notation:1 matched:3 lowest:1 kind:1 developed:2 uk:1 unit:4 appear:1 segmenting:1 vertically:1 subsumption:1 limit:1 oxford:1 pami:5 black:1 k:3 co:1 limited:1 directed:1 practical:1 vu:1 practice:1 area:1 matching:23 convenient:1 word:5 symbolic:1 selection:1 cartographic:1 applying:1 influence:1 restriction:1 equivalent:1 map:8 attention:1 commence:1 lfl:4 l:1 resolution:1 ke:1 immediately:1 factorisation:2 legitimate:1 rule:4 q:2 marr:2 traditionally:1 analogous:1 sar:3 hierarchy:21 heavily:1 smyth:1 homogeneous:1 element:5 updating:1 persist:1 qll:2 bottom:4 lix:1 inserted:2 ft:1 connected:1 hypothesised:2 highest:1 insertion:1 battery:1 radar:4 raise:1 segment:9 upon:1 edwin:1 sep:2 joint:4 easily:1 represented:1 various:1 distinct:2 hancock:6 effective:1 lift:1 peer:1 widely:1 drawing:1 otherwise:2 propose:1 loop:1 description:1 parent:3 p:1 plethora:1 requirement:1 produce:1 object:5 illustrate:1 develop:5 propagating:1 lauritzen:1 involves:1 quantify:1 yol:1 correct:3 attribute:2 elimination:1 require:1 underpinning:1 residing:3 exp:2 mapping:28 pointing:1 dictionary:3 adopt:1 label:9 edit:1 successfully:1 brought:1 aim:2 super:7 rather:1 wilson:6 encode:1 focus:2 june:1 modelling:2 abstraction:2 dependent:1 unary:2 hidden:1 relation:3 ancestor:1 compatibility:1 fidelity:1 ill:1 denoted:2 development:6 constrained:3 marginal:1 equal:1 once:1 keh:2 field:2 eliminated:1 identical:1 represents:2 reassigning:1 connectionist:1 intelligent:1 richard:1 realisation:1 distinguishes:1 causation:2 preserve:1 individual:4 familiar:1 interest:1 normalisation:1 possibility:1 intra:8 evaluation:1 severe:1 henderson:2 violation:1 operated:2 llf:1 chain:1 edge:3 capable:1 tree:1 re:1 causal:1 uncertain:1 instance:1 markovian:1 moralization:1 assignment:2 cost:2 uniform:1 density:1 probabilistic:3 together:1 concrete:1 siq:4 ambiguity:1 central:1 containing:1 corner:4 factorised:1 centred:2 coding:1 descendent:1 vi:17 depends:1 view:1 realised:1 bayes:2 elicited:1 parallel:1 ii1:1 kaufmann:1 correspond:1 landscape:1 bayesian:2 raw:2 lli:1 worth:1 corruption:1 straight:1 pp:6 naturally:1 attributed:1 mi:1 hamming:1 proved:1 subtle:1 follow:1 specify:1 editing:1 ljt:1 hand:1 ei:2 spm:7 mode:1 reveal:1 facilitate:1 contain:1 concept:1 counterpart:1 assigned:3 memoryless:1 conditionally:3 adjacent:1 ll:1 davis:1 criterion:1 ridge:1 upwards:2 reasoning:1 image:9 consideration:2 ef:1 fi:4 novel:1 ji:1 rl:1 cohen:1 extend:1 interpretation:6 measurement:4 significant:2 rd:1 meila:1 consistency:7 outlined:2 centre:1 concl:1 operating:1 delaunay:1 own:1 triangulation:2 moderate:1 exploited:1 preserving:5 morgan:1 additional:1 preceding:1 prune:1 ii:12 rj:2 ing:1 technical:1 match:23 levell:3 concerning:1 coded:1 impact:2 qi:4 heterogeneous:1 vision:2 represent:3 tailored:1 adopting:1 achieved:1 pyramidal:2 source:1 operate:1 member:1 inconsistent:2 effectiveness:1 jordan:2 structural:6 superclique:2 noting:1 concerned:3 usions:1 independence:3 topology:4 identified:1 idea:3 whether:1 expression:4 introd:1 effort:1 york:2 proceed:1 clear:1 clutter:3 lsi:1 dummy:4 correctly:2 discrete:13 write:3 express:1 group:1 key:2 drawn:1 hpm:3 graph:27 relaxation:12 powerful:1 topologically:3 draw:2 decision:1 fl:4 simplification:1 occur:1 constraint:14 scene:2 pi1:1 performing:1 department:1 influential:1 according:5 developing:1 belonging:1 heckerman:1 unity:1 explained:1 restricted:1 gidas:1 legal:1 equation:4 previously:2 remains:1 count:1 argmaxj:1 merit:1 tractable:2 adopted:1 available:2 operation:1 decomposing:2 apply:1 hierarchical:30 original:1 top:2 denotes:1 include:1 graphical:5 exploit:1 build:1 quantity:5 distance:1 link:1 mapped:3 entity:2 reason:1 assuming:1 index:2 potentially:1 relate:1 vertical:1 markov:2 qih:1 relational:4 communication:1 incorporated:1 excluding:1 intensity:1 required:2 conflict:1 pearl:1 assembled:1 proceeds:1 below:3 lion:1 optimise:1 critical:2 difficulty:1 residual:1 scheme:8 improve:1 spiegelhalter:1 extract:1 prior:1 literature:2 relative:1 fully:1 bear:1 gauged:1 organised:1 ingredient:1 digital:2 conveyed:1 consistent:3 dd:1 factorising:1 land:1 token:1 overcome:1 boundary:1 evaluating:1 resides:1 fb:1 sensory:2 made:1 san:1 dealing:2 clique:7 global:1 conceptual:1 francisco:1 factorize:1 xi:1 continuous:1 complex:1 vj:3 main:1 reconcile:1 child:13 allowed:3 referred:2 renormalization:1 cooper:1 sub:1 wish:1 exponential:1 xl:3 pul:1 perceptual:4 pe:2 third:1 theorem:1 down:2 jt:1 grouping:5 intractable:1 uction:1 sequential:1 effectively:6 importance:1 iel:1 labelling:7 conditioned:1 explore:1 desire:1 contained:1 satisfies:1 extracted:3 hedge:1 conditional:10 goal:2 viewed:1 labelled:1 feasible:1 determined:2 specifically:1 contrasted:1 averaging:1 experimental:2 rarely:1 formally:1 support:1 commences:1 assessed:1 philosophy:1 evaluate:1 |
497 | 1,455 | Asymptotic Theory for Regularization:
One-Dimensional Linear Case
Petri Koistinen
Rolf Nevanlinna Institute, P.O. Box 4, FIN-00014 University of Helsinki,
Finland. Email: PetrLKoistinen@rnLhelsinkLfi
Abstract
The generalization ability of a neural network can sometimes be
improved dramatically by regularization. To analyze the improvement one needs more refined results than the asymptotic distribution of the weight vector. Here we study the simple case of
one-dimensional linear regression under quadratic regularization,
i.e., ridge regression. We study the random design, misspecified
case, where we derive expansions for the optimal regularization parameter and the ensuing improvement. It is possible to construct
examples where it is best to use no regularization.
1
INTRODUCTION
Suppose that we have available training data (Xl, Yd, .. 0' (Xn' Y n ) consisting of
pairs of vectors, and we try to predict Yi on the basis of Xi with a neural network
with weight vector w. One popular way of selecting w is by the criterion
1
(1)
n
-n L ?(Xi' Yi, w) + >..Q(w) = min!,
I
where the loss ?(x,y,w) is, e.g., the squared error Ily - g(x,w)11 2 , the function
g(., w) is the input/output function of the neural network, the penalty Q(w) is
a real function which takes on small values when the mapping g(o, w) is smooth
and high values when it changes rapidly, and the regularization parameter >.. is a
nonnegative scalar (which might depend on the training sample). We refer to the
setup (1) as (training with) regularization, and to the same setup with the choice
>.. = 0 as training without regularization. Regularization has been found to be very
effective for improving the generalization ability of a neural network especially when
the sample size n is of the same order of magnitude as the dimensionality of the
parameter vector w, see, e.g., the textbooks (Bishop, 1995; Ripley, 1996).
Asymptotic Theory for Regularization: One-Dimensional Linear Case
295
In this paper we deal with asymptotics in the case where the architecture of the
network is fixed but the sample size grows . To fix ideas, let us assume that the
training data is part of an Li.d. (independent, identically distributed) sequence
(X,Y);(Xl'Yl),(X2'Y2)"" of pairs of random vectors, i.e., for each i the pair
(Xi, Yi) has the same distribution as the pair (X, Y) and the collection of pairs is
independent (X and Y can be dependent) . Then we can define the (prediction) risk
of a network with weights w as the expected value
(2)
r(w) := IE:f(X, Y, w).
Let us denote the minimizer of (1) by Wn (.),) , and a minimizer of the risk r by
w*. The quantity r(w n (>.)) is the average prediction error for data independent of
the training sample. This quantity r(w n (>.)) is a random variable which describes
the generalization performance of the network: it is bounded below by r( w*) and
the more concentrated it is about r(w*), the better the performance . We will
quantify this concentration by a single number, the expected value IE:r(w n (>.)) . We
are interested in quantifying the gain (if any) in generalization for training with
versus training without regularization defined by
(3)
When regularization helps, this is positive.
However, relatively little can be said about the quantity (3) without specifying in
detail how the regularization parameter is determined. We show in the next section
that provided>' converges to zero sufficiently quickly (at the rate op(n- 1 / 2 )), then
IE: r(wn(O)) and IE: r(w n (>.)) are equal to leading order. It turns out, that the optimal
regularization parameter resides in this asymptotic regime. For this reason, delicate
analysis is required in order to get an asymptotic approximation for (3). In this
article we derive the needed asymptotic expansions only for the simplest possible
case: one-dimensional linear regression where the regularization parameter is chosen
independently of the training sample.
2
REGULARIZATION IN LINEAR REGRESSION
We now specialize the setup (1) to the case of linear regression and a quadratic
smoothness penalty, i.e. , we take f(x,y,w) = [y-x T wJ2 and Q(w) = wTRw, where
now y is scalar, x and w are vectors, and R is a symmetric, positive definite matrix.
It is well known (and easy to show) that then the minimizer of (1) is
(4)
wn (>') =
[
1
-1
n
~ ~ XiX! + >'R
]
1
n
~ ~ XiYi.
This is called the generalized ridge regression estimator, see, e.g., (Titterington,
1985); ridge regression corresponds to the choice R = I, see (Hoerl and Kennard,
1988) for a survey. Notice that (generalized) ridge regression is usually studied in
the fixed design case, where Xi:s are nonrandom. Further, it is usually assumed
that the model is correctly specified, i.e., that there exists a parameter such that
Y i = Xr w* + ?i , and such that the distribution of the noise term ?i does not depend
on Xi. In contrast, we study the random design, misspecified case.
Assuming that IE: IIXI1 2 < 00 and that IE: [XXT] is invertible, the minimizer of the
risk (2) and the risk itself can be written as
(5)
(6)
w* = A-lIE: [XY],
r(w) = r(w*)
+ (w -
with
A:=IE:[XXT]
w*f A(w - w*).
P. Koistinen
296
If Zn is a sequence of random variables, then the notation Zn = open-a) means
that n a Zn converges to zero in probability as n -+ 00 . For this notation and the
mathematical tools needed for the following proposition see, e.g., (Serfiing, 1980,
Ch. 1) or (Brockwell and Davis, 1987, Ch. 6).
Proposition 1 Suppose that IE: y4 < 00, IE: IIXII 4 < 00 and that A = IE: [X XTj is invertible. If,\ = op(n- I / 2), then both y'n(wn(O) -w*) and y'n(w n ('\) - w*) converge
in distribution to N (0, C), a normal distribution with mean zero and covariance
matrix C.
The previous proposition also generalizes to the nonlinear case (under more complicated conditions). Given this proposition, it follows (under certain additional conditions) by Taylor expansion that both IE:r(w n ('\)) - r(w*) and IEr(wn(O)) - r(w*)
admit the expansion f31 n -} + o( n -}) with the same constant f3I. Hence, in the
regime ,\ = op(n-I/2) we need to consider higher order expansions in order to
compare the performance of wn (,\) and wn(O).
3
ONE-DIMENSIONAL LINEAR REGRESSION
We now specialize the setting of the previous section to the case where x is scalar.
Also, from now on, we only consider the case where the regularization parameter
for given sample size n is deterministic; especially ,\ is not allowed to depend on
the training sample. This is necessary, since coefficients in the following type of
asymptotic expansions depend on the details of how the regularization parameter
is determined. The deterministic case is the easiest one to analyze.
We develop asymptotic expansions for the criterion
(7)
where now the regularization parameter k is deterministic and nonnegative. The
expansions we get turn out to be valid uniformly for k ~ O. We then develop
asymptotic formulas for the minimizer of I n , and also for In(O) - inf I n . The last
quantity can be interpreted as the average improvement in generalization performance gained by optimal level of regularization, when the regularization constant
is allowed to depend on n but not on the training sample.
From now on we take Q(w) = w 2 and assume that A = IEX2 = 1 (which could be
arranged by a linear change of variables). Referring back to formulas in the previous
section, we see that
(8)
r(wn(k)) - r(w*) = ern - kw*)2/(Un + 1 + k)2 =: h(Un , Vn, k),
whence In(k) = IE:h(Un , Vn , k), where we have introduced the function h (used
heavily in what follows) as well as the arithmetic means Un and Vn
_
(9)
1
n
L Vi,
n
Vn:= -
with
I
(10)
Vi
:=
XiYi - w* xl
For convenience, also define U := X2 - 1 and V := Xy - w* X2 . Notice that
U; UI, U2 , ? .. are zero mean Li.d. random variables, and that V; Vi, V2 ,. " satisfy
the same conditions. Hence Un and Vn converge to zero, and this leads to the idea
of using the Taylor expansion of h(u, v, k) about the point (u, v) = (0,0) in order
to get an expansion for In(k).
297
Asymptotic Theory for Regularization: One-Dimensional Linear Case
To outline the ideas, let Tj(u,v,k) be the degree j Taylor polynomial of (u,v) f-7
h(u, v, k) about (0,0), i.e., Tj(u, v, k) is a polynomial in u and v whose coefficients are functions of k and whose degree with respect to u and v is j. Then
IETj(Un,Vn,k) depends on n and moments of U and V. By deriving an upper
bound for the quantity IE Ih(Un , Vn , k) - Tj(Un , Vn , k)1 we get an upper bound for
the error committed in approximating In(k) by IE Tj(Un , Vn , k). It turns out that
for odd degrees j the error is of the same order of magnitude in n as for degree
j - 1. Therefore we only consider even degrees j. It also turns out that the error
bounds are uniform in k ~ 0 whenever j ~ 2. To proceed, we need to introduce
assumptions.
Assumption 1 IE IXlr
< 00
and IE IYls
< 00
for high enough rand s.
Assumption 2 Either (a) for some constant j3 > 0 almost surely
X has a density which is bounded in some neighborhood of zero.
IXI ;:::
j3 or (b)
Assumption 1 guarantees the existence of high enough moments; the values r = 20
and s = 8 are sufficient for the following proofs. E.g., if the pair (X, Y) has a
normal distribution or a distribution with compact support, then moments of all
orders exist and hence in this case assumption 1 would be satisfied. Without some
condition such as assumption 2, In(O) might fail to be meaningful or finite. The
following technical result is stated without proof.
Proposition 2 Let p
> 0 and let 0 < IE X 2 < 00. If assumption 2 holds, then
where the expectation on the left is finite (a) for n
assumption 2 (a), respectively 2 (b) holds.
~
1 (b) for n
> 2p provided that
Proposition 3 Let assumptions 1 and 2 hold. Then there exist constants no and
M such that
In(k) = JET2(Un , Vn , k) + R(n, k) where
_ _
(w*)2k 2
W*kIEUV]
-1 [IEV2
(w*)2k2JEU 2
IET2(Un , Vn , k) = (1+k)2 +n
(1+k)2 +3 (1+k)4 +4 (1+k)3
IR(n, k) I :s; Mn- 3/ 2(k
+ 1)-1, "In;::: no, k ;::: o.
PROOF SKETCH The formula for IE T 2 (Un , Vn. k) follows easily by integrating the
degree two Taylor polynomial term by term. To get the upper bound for R(n, k),
consider the residual
where we have omitted four similar terms. Using the bound
P. Koistinen
298
the Ll triangle inequality, and the Cauchy-Schwartz inequality, we get
r
IR(n, k)1 = IJE [h(Un , Vn , k) - T2(Un , Vn , k)]1
., (k+
{2(k
W'
{Ii: [(~ ~Xl)-']
+ 1)3[JE (lUn I2IVnI 4 )]l/2 + 4(w*)2k 2(k + 1)[18 IUn I6 ]l/2 ... }
By proposition 2, here 18 [(~ 2:~ X[)-4] = 0(1). Next we use the following fact, cf.
(Serfiing, 1980, Lemma B, p. 68).
Fact 1 Let {Zd be i.i.d. with 18 [Zd = 0 and with 18 IZI/v < 00 for some v
Then
v
~
2.
Applying the Cauchy-Schwartz inequality and this fact, we get, e.g., that
[18 (IUn I2 IVnI 4 )]l/2 ~ [(18 IUn I4 )1/2(E IVnI 8)1/2p/2 = 0(n- 3 / 2).
o
Going through all the terms carefully, we see that the bound holds.
Proposition 4 Let assumptions 1 and 2 hold, assume that w* :j; 0, and set
al := (18 V2
- 2w*E [UVD/(w*)2.
If al > 0, then there exists a constant ni such that for all n ~ nl the function
k ~ ET2(Un , Vn,k) has a unique minimum on [0,(0) at the point k~ admitting the
expanszon
k~ = aIn- 1 + 0(n- 2); further,
In(O) - inf{Jn(k) : k ~ O} = In(O) - In(aln- 1 ) = ar(w*)2n- 2
If a
~
+ 0(n- 5 / 2).
0, then
The proof is based on perturbation expansio!1 c~nsidering lin a
small parameter. By the previous proposition, Sn(k) := ET2 (Un , Vn , k) is the sum
of (w*)2k 2/(1 + k)2 and a term whose supremum over k ~ ko > -1 goes to zero
as n ~ 00. Here the first term has a unique minimum on (-1,00) at k = O.
Differentiating Sn we get
PROOF SKETCH
S~(k) = [2(w*)2k(k
+ 1)2 + n- 1 p2(k)]/(k + 1)5,
where P2(k) is a second degree polynomial in k. The numerator polynomial has
three roots, one of which converges to zero as n ~ 00. A regular perturbation
expansion for this root, k~ = aln- I + a2n-2 + ... , yields the stated formula for
al. This point is a minimum for all sufficiently large n; further, it is greater than
zero for all sufficiently large n if and only if al > O.
The estimate for J n (0) - inf {J n (k) : k ~ O} in the case al > 0 follows by noticing
that
In(O) - In(k) = 18 [h(Un , Vn , 0) - h(Un , Vn , k)),
where we now use a third degree Taylor expansion about (u, v, k) = (0,0,0)
h(u,v,O) - h(u,v,k) =
2w* kv - (w*)2k2 - 4w*kuv + 2(w*?k 2u + 2kv 2 - 4w*k 2v + 2(W*)2k 3
+ r(u, v, k).
Asymptotic Theory for Regularization: One-Dimensional Linear Case
299
0.2
0.18
0.16
0.14
0.12
0.1
__
__
__
__
__
__
__
0.05 0.1
0.15 0 .2 0.25 0.3 0.35 0.4
~~
o
~
~
~
~
~
~
L-~
__
0.45
~
0.5
Figure 1: Illustration of the asymptotic approximations in the situation of equation
(11) . Horizontal axis kj vertical axis .In(l?) and its asymptotic appr~ximations.
Legend: markers In(k); solid line IE T 2 (Un , Vn , k)j dashed line IET4 (Un , Vn , k).
Usin~ t~e techniques of the previous proposition, it can be shown that
IE Ir(Un , Vn , k~)1 = O(n- S/ 2 ). Integrating the Taylor polynomial and using this
estimate gives
In(O) - In(aI/n) = af(w*)2n- 2 + O(n- S / 2 ).
Finally, by the mean value theorem,
In(O) -inf{ In(k) : k
~ O} =
In(O) -In(aI/n) +
!
= In(O) -
(In(O) -
In(k)]lk=8(k~ -aI/n)
In(aI/n)
+ O(n-1)O(n-2)
where () lies between k~ and aI/n, and where we have used the fact that the indicated derivative evaluated at () is of order O(n- 1 ), as can be shown with moderate
effort.
0
Remark In the preceding we assumed that A = IEX 2 equals 1. If this is not
the case, then the formula for a1 has to be divided by A; again, if a1 > 0, then
k~ = a1n-1 + O(n- 2 ) .
If the model is correctly specified in the sense that Y = w* X + E, where E is
independent of X and IE E = 0, then V = X E and IE [UV] = O. Hence we have
a1 = IE [E 2]j(w*)2, and this is strictly positive expect in the degenerate case where
E = 0 with probability one. This means that here regularization helps provided the
regularization parameter is chosen around the value aI/n and n is large enough.
See Figure 1 for an illustration in the case
(11)
X "'" N(O, 1) ,
Y = w* X
+f,
f "'"
N(O, 1),
w* = 1,
where E and X are independent. In(k) is estimated on the basis of 1000 repetitions
of the task for n = 8. In addition to IE T2(Un , Vn , k) the function IE T4 (Un , lin, k)
is also plotted. The latter can be shown to give In(k) correctly up to order
O(n- s/ 2 (k+ 1)-3). Notice that although IE T2 (Un , Vn , k) does not give that good an
approximation for In(k), its minimizer is near the minimizer of In(k), and both of
these minimizers lie near the point al/n = 0.125 as predicted by the theory. In the
situation (11) it can actually be shown by lengthy calculations that the minimizer
of In(k) is exactly al/n for each sample size n ~ 1.
It is possible to construct cases where
X "'" Uniform (a, b),
Y = cjX + d+ Z,
a1
< O.
For instance, take
1
1
b=-(3Vs-l)
2'
4
c= -5,d= 8,
a=-
P. Koistinen
300
and Z '" N (0, a 2 ) with Z and X independent and 0 :::; a < 1.1. In such a case
regularization using a positive regularization parameter only makes matters worse;
using a properly chosen negative regularization parameter would, however, help in
this particular case. This would, however, amount to rewarding rapidly changing
functions. In the case (11) regularization using a negative value for the regularization parameter would be catastrophic.
4
DISCUSSION
We have obtained asymptotic approximations for the optimal regularization parameter in (1) and the amount of improvement (3) in the simple case of one-dimensional
linear regression when the regularization parameter is chosen independently of the
training sample. It turned out that the optimal regularization parameter is, to
leading order, given by Qln- 1 and the resulting improvement is of order O(n- 2 ).
We have also seen that if Ql < 0 then regularization only makes matters worse.
Also (Larsen and Hansen, 1994) have obtained asymptotic results for the optimal
regularization parameter in (1). They consider the case of a nonlinear network;
however, they assume that the neural network model is correctly specified.
The generalization of the present results to the nonlinear, misspecified case might
be possible using, e.g., techniques from (Bhattacharya and Ghosh, 1978). Generalization to the case where the regularization parameter is chosen on the basis of the
sample (say, by cross validation) would be desirable.
Acknowledgements
This paper was prepared while the author was visiting the Department for Statistics and Probability Theory at the Vienna University of Technology with financial
support from the Academy of Finland. I thank F. Leisch for useful discussions.
References
Bhattacharya, R. N. and Ghosh, J. K. (1978). On the validity of the formal Edgeworth expansion. The Annals of Statistics, 6(2):434-45l.
Bishop, C. M. (1995). Neural Networks for Pattern Recognition. Oxford University
Press.
Brockwell, P. J. and Davis, R. A. (1987). Time Series: Theory and Methods.
Springer series in statistics. Springer-Verlag.
Hoerl, A. E. and Kennard, R. W. (1988). Ridge regression. In Kotz, S., Johnson,
N. L., and Read, C. B., editors, Encyclopedia of Statistical Sciences. John Wiley
& Sons, Inc.
Larsen, J. and Hansen, L. K. (1994). Generalization performance of regularized
neural network models. In Vlontos, J., Whang, J.-N., and Wilson, E., editors,
Proc. of the 4th IEEE Workshop on Neural Networks for Signal Processing,
pages 42-51. IEEE Press.
Ripley, B. D. (1996). Pattern Recognition and Neural Networks. Cambridge University Press.
Serfiing, R. J. (1980). Approximation Theorems of Mathematical Statistics. John
Wiley & Sons, Inc.
Titterington, D. M. (1985). Common structure of smoothing techniques in statistics.
International Statistical Review, 53:141-170.
A General Purpose Image Processing Chip:
Orientation Detection
Ralph Etienne-Cummings and Donghui Cai
Department of Electrical Engineering
Southern Illinois University
Carbondale, IL 6290 1-6603
Abstract
A 80 x 78 pixel general purpose vision chip for spatial focal plane
processing is presented. The size and configuration of the processing
receptive field are programmable. The chip's architecture allows the
photoreceptor cells to be small and densely packed by performing all
computation on the read-out, away from the array. In addition to the
raw intensity image, the chip outputs four processed images in parallel.
Also presented is an application of the chip to line segment orientation
detection, as found in the retinal receptive fields of toads.
1 INTRODUCTION
The front-end of the biological vision system is the retina, which is a layered structure
responsible for image acquisition and pre-processing. The early processing is used to
extract spatiotemporal information which helps perception and survival. This is
accomplished with cells having feature detecting receptive fields, such as the edge
detecting center-surround spatial receptive fields of the primate and cat bipolar cells
[Spillmann, 1990]. In toads, the receptive fields of the retinal cells are even more
specialized for survival by detecting ''prey'' and "predator" (from size and orientation
filters) at this very early stage [Spi11mann, 1990].
The receptive of the retinal cells performs a convolution with the incident image in
parallel and continuous time. This has inspired many engineers to develop retinomorphic
vision systems which also imitate these parallel processing capabilities [Mead, 1989;
Camp, 1994]. While this approach is ideal for fast early processing, it is not space
efficient. That is, in realizing the receptive field within each pixel, considerable die area
is required to implement the convolution kernel. In addition, should programmability be
required, the complexity of each pixel increases drastically. The space constraints are
eliminated if the processing is performed serially during read-out. The benefits of this
approach are 1) each pixel can be as small as possible to allow high resolution imaging,
2) a single processor unit is used for the entire retina thus reducing mis-match problems,
3) programmability can be obtained with no impact on the density of imaging array, and
R. Etienne-Cummings and D. Cai
874
4) compact general purpose focal plane visual processing is realizable. The space
constrains are then transfonned into temporal restrictions since the scanning clock speed
and response time of the processing circuits must scale with the size of the array.
Dividing the array into sub-arrays which are scanned in parallel can help this problem.
Clearly this approach departs from the architecture of its biological counterpart, however,
this method capitalizes on the main advantage of silicon which is its speed. This is an
example of mixed signal neuromorphic engineering, where biological ideas are mapped
onto silicon not using direct imitation (which has been the preferred approach in the past)
but rather by realizing their essence with the best silicon architecture and computational
circuits.
This paper presents a general purpose vision chip for spatial focal plane processing. Its
architecture allows the photoreceptor cells to be small and densely packed by performing
all computation on the read-out, away from the array. Performing computation during
read-out is ideal for silicon implementation since no additional temporal over-head is
required, provided that the processing circuits are fast enough. The chip uses a single
convolution kernel, per parallel sub-array, and the scanning bit pattern to realize various
receptive fields. This is different from other focal plane image processors which am
usually restricted to hardwired convolution kernels, such as oriented 20 Gabor filters
[Camp, 1994]. In addition to the raw intensity image, the chip outputs four processed
versions per sub-array. Also presented is an application of the chip to line segment
orientation detection, as found in the retinal receptive fields of toads [Spillmann, 1990].
2 THE GENERAL PURPOSE IMAGE PROCESSING CHIP
2.1
System Overview
This chip has an 80 row by 78 column photocell array partitioned into four independent
sub-arrays, which are scanned and output in parallel, (see figure I). Each block is 40 row
by 39 column, and has its own convolution kernel and output circuit. The scanning
circuit includes three parts: virtual ground, control signal generator (CSG), and scanning
output transformer. Each block has its own virtual ground and scanning output
transformer in both x direction (horizontal) and y direction (vertical). The control signal
generator is shared among blocks.
2.2
Hardware Implementation
The photocell is composed of phototransistor, photo current amplifier, and output
control.
The phototransistor performance light transduction, while the amplifier
magnifies the photocurrent by three orders of magnitude. The output control provides
multiple copies of the amplified photocun-ent which is subsequently used for focal plane
image processing.
The phototransistor is a parasitic PNP transistor in an Nwell CMOS process. The
current amplifier uses a pair of diode connected pmosfets to obtain a logarithmic
relationship between light intensity and output current. This circuit also amplifies the
photocurrent from nanoamperes to microamperes. The photocell sends three copies of the
output currents into three independent buses. The connections from the photocell to the
buses are controlled by pass transistors, as shown in Fig. 2. The three current outputs
allow the image to be processed using mUltiple receptive field organization (convolution
kernels), while the raw image is also output. The row (column) buses provides currents
for extracting horizontally (vertically) oriented image features, while the original bus
provides the logarithmically compressed intensity image.
The scanning circuit addresses the photocell array by selecting groups of cells at one time.
Since the output of the cells are currents, virtual ground circuits are used on each bus to
mask the> I pF capacitance of the buses. The CSG, implemented with shift registers
A General Purpose Image Processing Chip: Orientation Detection
"
"
"
"
t J, -=_"ODat~tl'g<I"O''''''
I
1: ~!; ~~.+
I
:': :~ I
~'!"' I
1"
,, " :;""
o;::!' ,, : 1
......
.
-
875
J9~
.,.
.
.--
I
~~"....
i
!
I
!
!
f
W
V
lloc',I: .... , ..-.
~
m
~
1
1
V
... .
......,. ..,.
1I..... ,2orf. _
e69r>0 . . .o-Y.........
1IIocIf,.?""._
lIIIoc~' l: Drl, ......
..... . . . " . "'!J" rty
..... , 1Id,..,.Mpwy
A
r
1
f
1
-f
L
1
I
1 : >< ~;;:-i :':'< :1
- -t.". ciii-.
- -
{
~
I
,I
,
1
!
f "
:,
L
I
I
f
=
-t
*",1,,",,1, ,',,1,><,.,,',
~I.. _ .... " " " " ' 1
~
I
1
?
,
I_
t
...f~;;;i
[!J
..?..?.? ??
jir;i
c....
oA""; -
-
w
d"' ~ ;;; " "
I
~~
-
~
~
Figure 1: Block diagram of the chip.
produces signals which select photocells and control the scanning output transformer.
The scanning output transformer converts currents from all row buses into Ipe? and Icenx'
and converts currents from all row buses into lpery and Iccny. This transformation is
required to implement the various convolution kernels discussed later.
The output transformer circuits are controlled by a central CSG and a peripheral CSG.
These two generators have identical structures but different initial values. It consists of
an n-bit shift register in x direction (horizontally) and an m-bit shift register in y direction
(vertically). A feedback circuit is used to restore the scanning pattern into the x shift
register after each row scan is completed. This is repeated until all the row in each block
are scanned.
The control signals from the peripheral and central CSGs select all the cells covered by a
2D convolution mask (receptive field). The selected cells send Ixy to the original bus, Ixp
to the row bus, and Iyp to the column bus. The function of the scanning output
transformer is to identify which rows (columns) are considered as the center (Icenx or Ircny)
or periphery (Ire rx or Ipcry) of the convolution kernel, respectively. Figure 3 shows how a
3x3 convolution kernel can be constructed.
Figure 4 shows how the output transformer works for a 3x3 mask. Only row bus
transformation is shown in this example, but the same mechanism applies to the column
bus as well. The photocell array is m row by n column, and the size is 3x3. The XC (x
center) and YC (y center) come from the central CSG; while XP (x peripheral) and YP (y
peripheral) come from the peripheral CSG. After loading the CSG, the initial values of
XP and YP are both 00011...1. The initial values of XC and YC are both 10 111.. .1.
This identifies the central cell as location (2, 2). The currents from the central row
(column) are summed to form Iren ? and leeny, while all the peripheral cells are summed to
form Iperx and lpery. This is achieved by activating the switches labeled XC, YC, XP and
YP in figure 2. XPj (YP,) {i= I, 2, ... , n} controls whether the output current of one cell
R. Etienne-Cummings and D. Cai
876
YC~
IO----+----yp
,.b---~p
XC---<::A.
Jyp
hp
Original Bus--<E,..-I'----+_--+_ _
Column Bus_~_--,_-+_ _
Row
(1 ,1)
(1 ,2)
(2,1)
(2,2)
(3,1)
(3,2)
lori
(3,:l
BUS_~_ _ _-----4_ _
Figure 2: Connections between a photocell and the current buses.
Figure 3:
field .
Constructing a 3x3 receptive
goes to the row (column) bus. Since XP j (YP) is connected to the gate of a pmos
switch, a 0 in XP j (YPj) it turns on . YC j (XC j ) {i=l, 2, ... , n} controls whether a row
(column) bus connects to Icenx bus in the same way. On the other hand, the connection
from a row (column) bus to Ipcrx bus is controlled by an nmos and a pmos switch. The
connection is made if and only if YC, (XCi)' an nmos switch, is 1 and YP i (XPi), a pmos
switches, is O. The intensity image is obtained directly when XCi and YC j are both O.
Hence, lori = 1(2,2), Icenx = lrow2= 1(2,1) + 1(2,2) + 1(2,3) and Iperx = lrowl + lrow3 = 1(1,1) +
1(1,2) + 1(1,3) + 1(3,1) + 1(3,2) + 1(3,3).
The convolution kernel can be programmed to perform many image processing tasks by
loading the scanning circuit with the appropriate bit pattern. This is illustrated by
configuring the chip to perform image smoothing and edge extraction (x edge, y edge, and
20 edge), which are all computed simultaneously on read-out. It receives five inputs (lori'
Iccn ,' lperx, Iceny, Ipcry) from the scanning circuit and produces five outputs (lori' ledge.> ledgey'
Ismllllth,ledge2d). The kernel (receptive field) size is programmable from 3x3, 5x5, 7x7, 9x9
and 11 x 11 . Fig. 5 shows the 3x3 masks for this processing. Repeating the above steps
for 5x5, 7x7, 9x9, and II x 11 masks, we can get similar results.
p..
u
>- >-
c:s:l
c:s:l
c:s:l
.....
--
YPI
VCI
Yl'2
YC2
c:s:l
----
YP3
YC3
YPIoI
YCN
Figure 4: Scanning output transformer for an m row by n column photo cell array.
A General Purpose Image Processing Chip: Orientation Detection
877
1
I
1
-I
-I
-I
-1
2
-I
0
-1
0
1
1
1
2
2
2
-I
2
-I
-I
4
-I
1
1
1
-I
-I
-1
-I
2
-I
0
-I
0
(a) smooth
(b) edge_x
(c) edge-y
(d) edge_2D
Figure 5: 3x3 convolution masks for various image processing.
In general, the convolution results under different mask sizes can be expressed as follows:
I~mooth=Icen. +Ire...
Iedge.=Kld *Icen. -Ipc",
Iedgey=Kld *Iceny-Ipcry
Iedge2D=K2d *I"ri-Icen.-Iceny
Where Kid and K2d are the programmable coefficients (from 2-6 and 6-14, respectively) for
ID edge extraction and 2D edge extraction, respectively. By varying the locations of the
O's in the scanning circuits, different types of receptive fields (convolution kernels) can be
realized.
2.3
Results
The chip contains 65K transistors in a footprint of 4.6 mm x 4.7 mm. There are 80 x 78
photocells in the chip, each of which is 45.6 11m x 45 !lm and a fill factor of 15%. The
convolution kernel occupies 690.6 !lm x 102.6 11m. The power consumption of the chip
for a 3x3 (1\ x 11) receptive field, indoor light, and 5V power supply is < 2 m W (8 m W).
To capitalize on the programmability of this chip, an ND card in a Pentium 133MHz PC
is used to load the scanning circuit and to collect data. The card, which has a maximum
analog throughput of 100KHz limits the frame rate of the chip to 12 frames per second.
At this rate, five processed versions of the image is collected and displayed. The scanning
and processing circuits can operate at 10 MHz (6250 fps), however, the phototransistors
have much slower dynamics. Temporal smoothing (smear) can be observed on the scope
when the frame rate exceeds 100 fps.
The chip displays a logarithmic relationship between light intensity and output current
(unprocessed imaged) from 0.1 lux (100 nA) to 6000 lux (10 IlA). The fixed pattern
noise, defined as standard-deviation/mean, decreases abruptly from 25% in the dark to 2%
at room light (800 lux). This behavior is expected since the variation of individual pixel
current is large compared to the mean output when the mean is small. The logarithmic
response of the photocell results in high sensitivity at low light, thus increasing the
mean value sharply. Little variation is observed between chips.
The contrast sensitivity of the edge detection masks is also measured for the 3x3 and 5x5
receptive fields. Here contrast is defined as (1m .. - Imin)/(Im .. + I min ) and sensitivity is given
as a percentage of the maximum output. The measurements are performed for normal
room and bright lighting conditions. Since the two conditions corresponded to the
saturated part of the logarithmic transfer function of the photocells, then a linear
relationship between output response and contrast is expected. Figure 6 shows contrast
sensitivity plot. Figure 7 shows examples of chip's outputs. The top two images are the
raw and smoothed (5x5) images. The bottom two are the 1D edge_x (left) and 2D edge
(right) images. The pixels with positive values have been thresholded to white. The
vertical black line in the image is not visible in the edge_x image, but can be clearly seen
in the edge_2D image.
878
R. Etienne-Cummings and D. Cai
80
><
...~o
60
~
:; 40
&
::I
o
..... ??5,5 Bri!,hl
-e-5xS Normal
20
... ? ?? 3,3 Bri!,hl
_ _ 3,3 Nonnal
Contrast [%]
Figure 6: Contrast sensitivity function of
the x edge detection mask.
Figure 7: (Clockwise) Raw image. 5x5
smoothed image. edge_2D and edge_x.
3 APPLICATION: ORIENTATION DETECTION
3.1
Algorithm Overview
This vision chip can be elegantly used to measure the orientation of line segments which
fall across the receptive field of each pixel. The output of the 10 Laplacian operators,
edge_x and edge_y, shown in figure 5, can be used to detennine the orientation of edge
segments. Consider a continuous line through the origin, represented by a delta function
in 20 space by IX y-xtan()). If the origin is the center of the receptive field. the response
ofthe edge_x kernel can be computed by evaluating the convolution equation (1). where
W(x) = u(x+m)-u(x-m) is the x window over which smoothing is performed, 2m+ J is the
width of the window and 2n+ J is the number of coefficients realizing the discrete
Laplacian operator. In our case, n =m. Evaluating this equation and substituting the
origin for the pixel location yields equation (2), which indicates that the output of the 10
edge_x (edge-y) detectors have a discretized linear relationship to orientation from on to
45" (45? to 90?). At 0", the second term in equation (2) is zero. As e increase, more
terms are subtracted until all tenns are subtracted at 45?. Above 45 0 (below 45?), the
edge_x (edge-y) detectors output zero since equal numbers of positive and negative
coefficients are summed. Provided that contrast can be nonnalized. the output of the
detectors can be used to extract the orientation of the line. Clearly these responses are
even about the x- and y-axis. respectively. Hence, a second pair of edge detectors. oriented
at 45", is required to uniquely extract the angle of the line segment.
10
- - 3 7 0 Lux
8
~
..=.
0
-
,0..,.
9-
-260 Lux
- -. - - (R!) Lux
-?? .. -? - (25 Lu x
b~
6
:;
s:::I
0
o:
-o._~
4
"I
~ ? ? . ? ?.? ?
2
,,b
, 'b
. _. .
'0,
' . .. . .
o
o
20
CI' ...'
-. ... ... . ...~ .~
....:-
40
.
~ . ':
? ? . , , ; : : . :-:-' . - - ??
60
T
-0- - _ _ _
80
Angle ['-']
Figure 8: Measured orientation transfer function of edge_x detectors.
A General Purpose Image Processing Chip: Orientation Detection.
879
n
0edge_Ax,y) = [2nW(x ?m)8(y)- ...EW(x ?m)8(y?i)]*8(y-xtane)
(I)
;=}
n?
~
I
.
-I
Oedge AO.O) = 2n-[ ~(W(--)+ W(--)]
;=}
tane
fane
3.2
(2)
Results
Figure 8 shows the measured output of the edge_x detectors for various lighting
conditions as a line is rotated. The average positive outputs are plotted. As expected, the
output is maximum for bright ambients when the line is horizontal. As the line is
rotated, the output current decreases linearly and levels off at approximately 45". On the
other hand, the edge_y (not shown) begins its linear increase at 45" and maximizes at 90?.
After normalizing for brightness. the four curves are very similar (not shown).
To further demonstrate orientation detection with this chip, a character consisting of a
circle and some straight lines is presented. The intensity image of the character is shown
in figure 9(a). Figures 9(b) and 9(c) show the outputs of the edge_x and edge-y
detectors, respectively. Since a 7x7 receptive field is used in this experiment, some outer
pixels of each block are lost. The orientation selectivity of the 1D edge detectors are
clearly visible in the figures, where edge_x highlights horizontal edges and edge_y vertical
edges. Figure 9(d) shows the reported angles. A program is written which takes the two
I D edge images, finds the location of the edges from the edge_2D image, the intensity at
the edges (positive lobe) and then computes the angle of the edge segment. In figure 9(d),
the black background is chosen for locations where no edges are detected, white is used for
0? and gray for 90?.
(a)
(b)
(c)
(d)
Figure 9: Orientation detection using ID Laplacian Operators.
4 CONCLUSION
A 80x78 pixel general purpose vision chip for spatial focal plane processing has been
presented. The size and configuration of the processing receptive field are programmable.
In addition to the raw intensity image, the chip outputs four processed images in parallel.
The chip has been successfully used for compact line segment orientation detection,
which can be used in character recognition. The programmability and relatively low
power consumption makes it ideal for many visual processing tasks.
References
Camp W. and J. Van cler Spiegel, "A Silicon VLSI Optical Sensor for Pattern
Recognition, " Sensors and Actuators A, Vol. 43, No. 1-3, pp. 188-195, 1994.
Mead C. and M. Ismail (Eds.), Analog VLSI Implementation of Neural Networks,
Kluwer Academic Press, Newell, MA, 1989.
Spi11mann L. and J. Werner (Eds.), Visual Perception: The Neurophysiological
Foundations, Academic Press, San Diego, CA, 1990.
| 1455 |@word version:2 polynomial:6 loading:2 nd:1 open:1 orf:1 covariance:1 lobe:1 brightness:1 solid:1 moment:3 initial:3 configuration:2 series:2 contains:1 selecting:2 wj2:1 past:1 current:15 xiyi:2 written:2 must:1 john:2 realize:1 visible:2 plot:1 v:1 selected:1 imitate:1 plane:6 capitalizes:1 realizing:3 cjx:1 detecting:3 provides:3 ire:2 location:5 five:3 mathematical:2 constructed:1 direct:1 supply:1 fps:2 lux:6 specialize:2 consists:1 drl:1 pnp:1 introduce:1 mask:9 expected:5 behavior:1 magnifies:1 discretized:1 inspired:1 little:2 pf:1 window:2 increasing:1 provided:5 begin:1 bounded:2 notation:2 circuit:15 maximizes:1 what:1 easiest:1 interpreted:1 aln:2 textbook:1 titterington:2 ghosh:2 transformation:2 nonrandom:1 guarantee:1 temporal:3 bipolar:1 exactly:1 k2:1 schwartz:2 control:8 unit:1 configuring:1 positive:8 engineering:2 vertically:2 limit:1 io:1 oxford:1 mead:2 id:3 yd:1 approximately:1 might:3 black:2 studied:1 specifying:1 collect:1 programmed:1 unique:2 responsible:1 lost:1 block:6 definite:1 edgeworth:1 implement:2 x3:9 xr:1 footprint:1 asymptotics:1 area:1 gabor:1 pre:1 integrating:2 regular:1 ila:1 get:9 convenience:1 onto:1 layered:1 operator:3 risk:4 applying:1 transformer:8 phototransistor:3 kld:2 restriction:1 ier:1 deterministic:3 xci:2 center:5 send:1 go:2 independently:2 survey:1 resolution:1 estimator:1 array:13 deriving:1 fill:1 financial:1 et2:2 variation:2 annals:1 diego:1 suppose:2 heavily:1 rty:1 ixi:1 us:2 origin:3 logarithmically:1 recognition:4 qln:1 labeled:1 observed:2 bottom:1 electrical:1 connected:2 decrease:2 ui:1 complexity:1 constrains:1 dynamic:1 depend:5 segment:7 basis:3 triangle:1 easily:1 wtrw:1 chip:29 cat:1 various:4 represented:1 xxt:2 fast:2 effective:1 detected:1 corresponded:1 neighborhood:1 refined:1 iun:3 pmos:3 whose:3 say:1 compressed:1 ability:2 statistic:5 mooth:1 spiegel:1 itself:1 sequence:2 advantage:1 transistor:3 cai:4 turned:1 rapidly:2 a2n:1 brockwell:2 degenerate:1 amplified:1 academy:1 ismail:1 kv:2 amplifies:1 ent:1 produce:2 cmos:1 converges:3 rotated:2 help:5 derive:2 develop:3 measured:3 odd:1 op:3 p2:2 dividing:1 implemented:1 predicted:1 diode:1 come:2 quantify:1 direction:4 filter:2 subsequently:1 occupies:1 virtual:3 activating:1 fix:1 generalization:8 ao:1 proposition:10 biological:3 im:1 strictly:1 hold:5 mm:2 sufficiently:3 around:1 ground:3 normal:4 considered:1 nwell:1 mapping:1 predict:1 whang:1 appr:1 lm:2 substituting:1 scope:1 nw:1 finland:2 early:3 omitted:1 purpose:9 proc:1 csg:7 hoerl:2 hansen:2 ain:1 repetition:1 successfully:1 tool:1 clearly:4 sensor:2 rather:1 varying:1 wilson:1 iex:1 kid:1 improvement:5 ily:1 properly:1 indicates:1 contrast:8 pentium:1 realizable:1 sense:1 serfiing:3 whence:1 camp:3 dependent:1 minimizers:1 am:1 entire:1 vlsi:2 going:1 interested:1 nonnal:1 ralph:1 pixel:10 among:1 orientation:17 retinomorphic:1 smoothing:4 spatial:4 summed:3 equal:3 construct:2 field:19 having:1 extraction:3 eliminated:1 identical:1 kw:1 capitalize:1 throughput:1 donghui:1 petri:1 t2:3 retina:2 oriented:3 composed:1 simultaneously:1 densely:2 individual:1 xtj:1 consisting:2 connects:1 delicate:1 amplifier:3 detection:12 organization:1 saturated:1 nl:1 admitting:1 light:6 pc:1 tj:4 programmability:4 edge:24 necessary:1 xy:2 carbondale:1 taylor:6 circle:1 plotted:2 instance:1 column:13 ar:1 mhz:2 zn:3 werner:1 neuromorphic:1 deviation:1 uniform:2 johnson:1 a1n:1 front:1 reported:1 scanning:16 spatiotemporal:1 referring:1 density:2 international:1 sensitivity:5 xpi:1 ie:26 koistinen:4 yl:2 rewarding:1 off:1 invertible:2 quickly:1 na:1 squared:1 again:1 satisfied:1 central:5 x9:2 worse:2 admit:1 derivative:1 leading:2 yp:7 li:2 retinal:4 includes:1 coefficient:5 matter:2 inc:2 satisfy:1 register:4 vi:3 depends:1 performed:3 try:1 root:2 later:1 analyze:2 complicated:1 parallel:7 capability:1 predator:1 il:1 ir:3 ni:1 bright:2 yield:2 identify:1 ofthe:1 raw:6 lu:1 rx:1 lighting:2 straight:1 processor:2 detector:8 whenever:1 ed:2 email:1 lengthy:1 acquisition:1 pp:1 larsen:2 proof:5 mi:1 gain:1 popular:1 dimensionality:1 carefully:1 actually:1 back:1 higher:1 cummings:4 response:5 improved:1 rand:1 izi:1 arranged:1 evaluated:1 box:1 stage:1 clock:1 until:2 sketch:2 hand:2 horizontal:4 receives:1 spillmann:2 vci:1 nonlinear:3 marker:1 leisch:1 indicated:1 gray:1 grows:1 validity:1 y2:1 counterpart:1 regularization:36 hence:6 read:6 symmetric:1 imaged:1 i2:1 illustrated:1 deal:1 white:2 ll:1 numerator:1 during:2 x5:5 width:1 davis:2 essence:1 die:1 uniquely:1 ixy:1 criterion:2 generalized:2 outline:1 ridge:5 smear:1 demonstrate:1 performs:1 image:34 misspecified:3 common:1 specialized:1 overview:2 ipc:1 khz:1 discussed:1 analog:2 kluwer:1 refer:1 silicon:5 measurement:1 cambridge:1 surround:1 ai:6 smoothness:1 uv:1 focal:6 i6:1 hp:1 illinois:1 uvd:1 own:2 inf:4 moderate:1 kuv:1 periphery:1 selectivity:1 certain:1 verlag:1 inequality:3 tenns:1 yi:3 accomplished:1 seen:2 minimum:3 additional:2 greater:1 preceding:1 surely:1 converge:2 clockwise:1 signal:6 dashed:1 arithmetic:1 ii:2 desirable:1 multiple:2 exceeds:1 smooth:2 technical:1 match:1 af:1 calculation:1 cross:1 lin:2 academic:2 divided:1 a1:4 laplacian:3 controlled:3 impact:1 prediction:2 regression:11 j3:2 ko:1 vision:6 expectation:1 sometimes:1 kernel:13 achieved:1 cell:14 addition:5 background:1 diagram:1 sends:1 operate:1 usin:1 legend:1 extracting:1 near:2 ideal:3 identically:1 wn:8 easy:1 enough:4 switch:5 architecture:5 f3i:1 idea:4 shift:4 unprocessed:1 whether:2 effort:1 ipe:1 penalty:2 abruptly:1 proceed:1 remark:1 programmable:4 dramatically:1 useful:1 covered:1 j9:1 amount:2 repeating:1 prepared:1 encyclopedia:1 dark:1 concentrated:1 processed:5 hardware:1 simplest:1 exist:2 percentage:1 notice:3 estimated:1 delta:1 correctly:4 per:3 zd:2 discrete:1 vol:1 group:1 four:6 ypi:1 ije:1 changing:1 prey:1 thresholded:1 imaging:2 sum:1 convert:2 angle:4 noticing:1 almost:1 kotz:1 vn:23 bit:4 bound:6 display:1 quadratic:2 nonnegative:2 i4:1 scanned:3 constraint:1 sharply:1 helsinki:1 x2:3 ri:1 x7:3 speed:2 nmos:2 min:2 performing:3 optical:1 relatively:2 bri:2 ern:1 department:2 imin:1 peripheral:6 describes:1 across:1 son:2 character:3 partitioned:1 xpj:1 primate:1 hl:2 restricted:1 equation:5 bus:20 turn:5 fail:1 mechanism:1 needed:2 end:1 photo:2 available:1 generalizes:1 detennine:1 actuator:1 v2:2 away:2 appropriate:1 photocurrent:2 bhattacharya:2 subtracted:2 gate:1 slower:1 existence:1 jn:1 original:3 ledge:1 top:1 cf:1 completed:1 vienna:1 etienne:4 xc:5 especially:2 approximating:1 capacitance:1 quantity:5 realized:1 receptive:20 concentration:1 said:1 visiting:1 southern:1 thank:1 mapped:1 card:2 oa:1 ensuing:1 outer:1 consumption:2 cauchy:2 collected:1 reason:1 transfonned:1 assuming:1 y4:1 illustration:2 relationship:4 setup:3 ql:1 stated:2 negative:3 design:3 implementation:3 packed:2 perform:2 upper:3 vertical:4 convolution:16 i_:1 fin:1 finite:2 displayed:1 situation:2 nonnalized:1 committed:1 head:1 frame:3 perturbation:2 smoothed:2 intensity:9 introduced:1 pair:8 required:6 specified:3 nevanlinna:1 connection:4 address:1 below:2 usually:3 pattern:7 perception:2 yc:7 regime:2 indoor:1 rolf:1 program:1 power:3 serially:1 regularized:1 hardwired:1 restore:1 residual:1 mn:1 technology:1 ivni:2 jir:1 axis:3 lk:1 identifies:1 toad:3 extract:3 sn:2 kj:1 review:1 acknowledgement:1 asymptotic:15 loss:1 expect:1 highlight:1 mixed:1 k2d:2 versus:1 generator:3 validation:1 foundation:1 degree:8 incident:1 sufficient:1 xp:5 article:1 editor:2 row:17 last:1 copy:2 drastically:1 formal:1 allow:2 institute:1 fall:1 differentiating:1 distributed:1 xix:1 f31:1 benefit:1 xn:1 valid:1 feedback:1 resides:1 evaluating:2 curve:1 author:1 collection:1 made:1 computes:1 san:1 compact:3 iixi1:1 preferred:1 supremum:1 assumed:2 xi:5 ripley:2 imitation:1 un:24 continuous:2 transfer:2 ca:1 improving:1 expansion:13 constructing:1 elegantly:1 main:1 linearly:1 noise:2 allowed:2 repeated:1 kennard:2 fig:2 je:1 tl:1 transduction:1 wiley:2 lun:1 sub:4 xl:4 lie:3 third:1 ix:1 formula:5 theorem:2 departs:1 load:1 bishop:2 x:1 normalizing:1 survival:2 phototransistors:1 exists:2 workshop:1 ih:1 iixii:1 gained:1 ci:1 magnitude:3 t4:1 lori:4 logarithmic:4 neurophysiological:1 visual:3 horizontally:2 expressed:1 scalar:3 u2:1 van:1 applies:1 springer:2 ch:2 corresponds:1 minimizer:8 newell:1 ma:1 quantifying:1 room:2 shared:1 considerable:1 change:2 determined:2 uniformly:1 reducing:1 lemma:1 engineer:1 called:1 pas:1 catastrophic:1 photoreceptor:2 meaningful:1 ew:1 select:2 parasitic:1 support:2 latter:1 ciii:1 scan:1 |
498 | 1,456 | On the Separation of Signals from
Neighboring Cells in Tetrode Recordings
Maneesh Sahani, John S. Pezaris and Richard A. Andersen
maneesh@caltech.edu, pz@caltech.edu, andersen@vis.caltech.edu
Computation and Neural Systems
California Institute of Technology
216-76 Caltech, Pasadena, CA 91125 USA
Abstract
We discuss a solution to the problem of separating waveforms produced by multiple cells in an extracellular neural recording. We
take an explicitly probabilistic approach, using latent-variable models of varying sophistication to describe the distribution of waveforms produced by a single cell. The models range from a single
Gaussian distribution of waveforms for each cell to a mixture of
hidden Markov models. We stress the overall statistical structure
of the approach, allowing the details of the generative model chosen
to depend on the specific neural preparation.
1
INTRODUCTION
Much of our empirical understanding of the systems-level functioning of the brain
has come from a procedure called extracellular recording. The electrophysiologist
inserts an insulated electrode with exposed tip into the extracellular space near
one or more neuron cell bodies. Transient currents due to action potentials across
nearby cell membranes are then recorded as deflections in potential, spikes, at the
electrode tip. At an arbitrary location in gray matter, an extracellular probe is likely
to see pertubations due to firing in many nearby cells, each cell exhibiting a distinct
waveform due to the differences in current path between the cells and the electrode
tip. Commonly, the electrode is maneuvered until all the recorded deflections have
almost the same shape; the spikes are then all presumed to have arisen from a
single isolated cell. This process of cell isolation is time-consuming, and it permits
recording from only one cell at a time. IT differences in spike waveform can be
exploited to sort recorded events by cell, the experimental cost of extracellular
recording can be reduced, and data on interactions between simultaneously recorded
cells can be obtained.
Separation of Signals from Neighboring Cells
223
Many ad hoc solutions to spike sorting have been proposed and implemented, but
thus far an explicit statistical foundation, with its accompanying benefits, has
mostly been lacking. Lewicki (1994) is the exception to this rule and provides a wellfounded probabilistic approach, but uses assumptions (such as isotropic Gaussian
variability) that are not well supported in many data sets (see Fee et al (1996)).
A first step in the construction of a solution to the spike-sorting problem is the
specification of a model by which the data are taken to be generated. The model
has to be powerful enough to account for most of the variability observed in the
data, while being simple enough to allow tractable and robust inference. In this
paper we will discuss a number of models, of varying sophistication, that fall into
a general framework. We will focus on the assumptions and inferential components
that are common to these models and consider the specific models only briefly. In
particular, we will state the inference algorithms for each model without derivation
or proof; the derivations, as well as measures of performance, will appear elsewhere.
2
DATA COLLECTION
The algorithms that appear in this paper are likely to be of general applicability. They have been developed, however, with reference to data collected from the
parietal cortex of adult rhesus macaques using tetrodes (Pezaris et a11997).
The tetrode is a bundle of four individually insulated 13/lm-diameter wires twisted
together and cut so that the exposed ends lie close together. The potential on each
wire is amplified (custom electronics), low-pass filtered (9-pole Bessel filter, Ie =
6.4 kHz) to prevent aliasing, and digitized (fs between 12.8 and 20 kHz) (filters and
AjD converter from Thcker Davis Technologies). This data stream is recorded to
digital media; subsequent operations are currently performed off-line.
In preparation for inference, candidate events (where at least one cell fired) are
identified in the data stream. The signal is digitally high-pass filtered (fe = 0.05Is)
and the root-mean-square (RMS) amplitude on each channel is calculated. This
value is an upper bound on the noise power, and approaches the actual value when
the firing rates of resolvable cells are low. Epochs where the signal rises above three
times the RMS amplitude for two consecutive signals are taken to be spike events.
The signal is upsampled in the region of each such threshold crossing, and the time
of the maximal subsequent peak across all channels is determined to within onetenth of a sample. A short section is then extracted at the original Is such that this
peak time falls at a fixed position in the extracted segment. One such waveform is
extracted for each threshold crossing.
3
GENERATIVE FRAMEWORK
Our basic model is as follows. The recorded potential trace V(t) is the sum of
influences that are due to resolvable foreground cells (which have a relatively large
e~ect) and a background noise process. We write
(1)
Here, c~ is an indicator variable that takes the value 1 if the mth cell fires at time
and 0 otherwise. If cell m fires at T it adds a deflection of shape S::n (t - T) to the
recorded potential. The effect of all background neural sources, and any electrical
noise, is gathered into a single term 7](t). For a multichannel probe, such as a
tetrode, all of V(t), 7](t) and S::n(t) are vector-valued. Note that we have indexed
T
M. Sahani, 1. S. pezaris and R. A Andersen
224
8)
@)
! !
Figure 1: Schematic graph of the general framework.
the spike shapes from the mth cell by time; this allows us to model changes in the
spike waveform due to intrinsic biophysical processes (such as sodium inactivation
during a burst of spikes) as separate to the additive background process. We will
discuss models where the choice of S::n is purely stochastic, as well as models in
which both the probability of firing and the shape of the action potential depend
on the recent history of the cell.
It will be useful to rewrite (1) in terms of the event waveforms described in section 2.
At times r when no foreground cell fires all the c~ are zero. We index the remaining
times (when at least one cell fired) by i and write c!n for c~ at ri (similarly for S:n)
to obtain
(2)
This basic model is sketched, for the case of two cells, in figure 1. Circles represent stochastic variables and squares deterministic functions, while arrows indicate
conditional or functional dependence. We have not drawn nodes for 0Tl and O.
The representation chosen is similar to, and motivated by, a directed acyclic graph
(DAG) model of the generative distribution. For clarity, we have not drawn edges
that represent dependencies across time steps; the measurement V(t) depends on
many nearby values of S::n and c~, and f/(t) may be autocorrelated. We will continue to omit these edges, even when we later show connections in time between c~
and S::n .
4
INFERENCE
We have two statistical objectives. The first is model selection, which includes the
choice of the number of cells in the foreground . The second is inference: finding
good estimates for the c~ given the measured V(t) . We will have little to say on
the subject of model selection in this paper, besides making the observation that
standard techniques such as cross-validation, penalized likelihood or approximation
of the marginal likelihood (or "evidence") are all plausible approaches. We will
instead focus on the inference of the spike times.
Rather than calculating the marginalized posterior for the c~ we will find the distribution conditioned on the most probable values of the other variables. This is a
common approximation to the true posterior (compare Lewicki (1994)).
A simple property of the data allows us to estimate the most probable values of
Separation of Signals from Neighboring Cells
225
the parameters in stages; times at which at least one foreground cell fires can be
identified by a threshold, as described in section 2. We can then estimate the noise
parameters 8Tf by looking at segments of the signal with no foreground spikes, the
waveform distribution and firing time parameters 8 from the collection of spike
events, and finally the spike times c:-n and the waveforms
by a filtering process
applied to the complete data V(t) given these model parameters.
S:n
4.1
NOISE
We study the noise distribution as follows. We extract lms segments from a bandpassed recording sampled at 16 kHz from a four-channel electrode, avoiding the
foreground spikes identified as in section 2. Each segment is thus a 64-dimensional
object. We find the principal components of the ensemble of such vectors, and
construct histograms of the projections of the vectors in these directions. A few of
these histograms are shown on a log-scale in figure 2 (points), as well as a zero-mean
Gaussian fit to the distribution projected along the same axes (lines). It is clear
that the Gaussian is a reasonable description, although a slight excess in kurtosis
is visible in the higher principal components.
pc 1
2l102~.
c:
::l
8
0
10
pc 24
pc 12
pc 36
pc 48
.,
10-2
-100
0
IlV
100 -50
o
50 -20
o
20 -20
o
20 -20
o
20
Figure 2: Distribution of background noise.
The noise parameters are now seen to be the covariance of the noise, ETf (we represent it as a covariance matrix taken over the length of a spike). In general, we can
fit an autoregressive process description to the background and apply a filter that
will whiten the noise. This will prove to be quite useful during the filtering stages.
4.2
WAVEFORM PARAMETERS
We can make some general remarks about the process of inferring the parameters
of the models for S~ and c:-n. Specific models and their inference algorithms will
appear in section 5.
The models will, in general, be fit to the collection of segments extracted and aligned
as described in section 2. At other times they have no influence on the waveform
recorded. We will represent these segments by Vi, implying a connection to the
firing events ri used in (2). It should be borne in mind that the threshold-based
trigger scheme will not exactly identify all of the true ri correctly.
We will assume that each segment represents a single Sm, that is, that no two
cells fire at times close enough for their spike waveforms to overlap. This is an
unreasonable assumption; we can shore it up partially by eliminating from our collection of Vi segments that appear heuristically to contain overlaps (for example,
double-peaked waveforms). Ultimately, however, we will need to make our inference procedure robust enough that the parameters describing the model are well
estimated despite the errors in the data.
M. Sahani, 1. S. Pezaris and R. A Andersen
226
Figure 3: The mixture model for Vi.
The advantage to making this assumption is that the overall model for the distribution of the Vi becomes a mixture: a single control variable c i sets exactly one of the
c~ to 1. Vi is then drawn from the distribution of waveforms for the selected cell,
convolved with the noise. This is a formal statement of the "clustering" approach to
spike-sorting. Mixture models such as these are easy to fit using the ExpectationMaximization (EM) algorithm (Dempster et al1977). We will also consider models
with additional latent state variables, which are used to describe the distributions
of the Sm and Cm, where again EM will be of considerable utility.
The measured ensemble Vi will be incorrect on a number of counts. The threshold
may make either false positive or false negative errors in selecting ri, and some of
the identified Vi will represent overlaps. We can use heuristics to minimize such
errors, but need to account for any remaining outliers in our models. We do so by
introducing additional mixture components. Segments of noise that are incorrectly
identified as foreground events are handled by an explicit zero mixture component
whose variability is entirely due to the background noise. Overlaps are handled by
providing very broad low-probability components spanning large areas in waveform
space; clusters of overlap waveforms are likely to be diffuse and sparse.
The mixture model is sketched in figure 3. In the basic model the variables are
chosen independently for each cross-threshold event. The dynamic models discussed
below will introduce dependencies in time.
4.3
SPIKE TIMES
In our final stage of inference, we make estimates of the c~ given the V(t) and
the most-probable parameters fit in the previous two stages. This is exactly the
signal detection problem of identifying pulses (perhaps with random or else adapting
parameters) in Gaussian noise of known covariance. Solutions to this are well known
(McDonough and Whalen 1995) and easily adapted to the problem at hand (Sahani
et al1998).
Separation of Signals from Neighboring Cells
5
227
SPECIFIC MODELS
Finally, we describe examples of models that may be used within this framework.
As stated before, in this brief catalog we summarize the motivation for each, and
state without derivation or proof the algorithms for inference. The details of these
algorithms, as well as tests of performance, will appear elsewhere.
5.1
CONSTANT WAVEFORM
The simplest model is one in which we take the waveform of the mth cell to remain
unchanged and the firing probability of each cell to be constant. In this case we drop
the index r or i on the waveform shape and just write Sm (t - r i ) . We write Pm for
the probability that a given event is due to the mth cell firing. The mixture model
is then a mixture of multivariate Gaussian distributions, each with covariance E r"
mean Sm and mixture fraction Pm. The EM algorithm for such a mixture is well
known (Nowlan 1990).
Given parameters 8(n) = {S~), p~)} from the nth iteration, we find the expected
values of the e~ (called the responsibilities),
ri
= ?[e i I {Vi}
m
m
8(n)]
,
=
(n) N(Vi. S(n) E
)
pm
, m , 1)
'"'p~n) N(Vi. S~n) E )'
L.Jm
in
(3)
'm'1)
and then reestimate the parameters from the data weighted by the responsibilities.
P(n+l) =
L:r~
_i_ _ .
5.2
(4)
N'
m
REFRACTORY FIRING
A simple modification to this scheme can be used to account for the refractory period
between spikes from the same cell (Sahani et al1998). The model is similar to the
Gaussian mixture above, except that the choice of mixture component is no longer
independent for each waveform. If the waveforms arrive within a refractory period
they cannot have come from the same cell. This leads to the altered responsibilities:
IT
i
i
rm
sm = Zi
j ;(i,j)
(5)
refractory
where Z is a normalizing constant.
The M step here is identical to (4), with the responsibilities s~ replacing the r~.
5.3
STATIC MIXTURE
As we have suggested above, the waveform of the mth cell is not, in fact, unchanged
each time the cell fires. Variability in excess of the additive background noise is
introduced by changes in the biophysical properties of the cell (due to recent firing
patterns, or external modulators) as well as by background activity that may be
correlated with foreground events. We can attempt to model this variability as
giving rise to a discrete set of distinct waveforms, which are then convolved with the
previously measured noise covariance to obtain the distribution of measurements. In
effect, we are tiling an irregularly shaped distribution with a mixture of Gaussians
M. Sahani, J S. pezaris and R. A Andersen
228
of fixed shape, E1J" We obtain a hierarchical mixture distribution in which each
component corresponding to a cell is itself is a mixture of Gaussians. Given a
particular hierarchical arrangement the parameters can be fit exactly as above.
While this approach seems attractive, it suffers from the flaw that model selection
is not well defined. In particular, the hierarchical mixture is equivalent in terms of
likelihood and parameters to a single-layer, flat, mixture. To avoid this problem we
may introduce a prior requiring that the Gaussian components from a single cell
overlap, or otherwise lie close together. It is, however, difficult to avoid excessive
sensitivity to such a prior.
5.4
DYNAMICAL MIXTURE
An alternative approach is to replace the independent transitions between the components of the mixture distribution of a single cell with a dynamical process that
reflects the manner in which both firing probability and waveform shape depend
on the recent history of the cell. In this view we may construct a mixture of hidden Markov models (HMMs), one for each cell. Our earlier mixture assumption
now means that the models must be coupled so that on anyone time step at most
one makes a transition to a state corresponding to firing. This structure may be
thought of as a special case of the factorial HMM discussed by Gharamani and
Jordan (1997).
The general model is known to be intractable ..In this special case, however, the
standard forward-backward procedure for a single HMM can be modified to operate
on reponsibiIity-weighted data, where the reponsibilities are themselves calculated
during the forward phase. This is empirically found to provide an effective Estep.
The M step is then straightforward.
Acknowledgements
This work has benefited considerably from important discussions with both Bill
Bialek and Sam Roweis. John Hopfield has provided invaluable advice and mentoring to MS. We thank Jennifer Linden and Philip Sabes for useful comments on
an earlier version of the manuscript. Funding for various components of the work
has been provided by the Keck Foundation, the Sloan Center for Theoretical Neuroscience at Caltech, the Center for Neuromorphic Systems Engineering at Caltech,
and the National Institutes of Health.
References
Dempster, A. P., N. M. Laird, and D. B. Rubin (1977). J. Royal Stat. Soc. B 39,
1-38.
Fee, M. S., P. P. Mitra, and D. Kleinfeld (1996). J. Neurophys. 76(3),3823-3833.
Gharamani, Z. and M. I. Jordan (1997). Machine Learning 29, 245-275.
Lewicki, M. S. (1994). Neural Compo 6(5), 1005-1030.
McDonough, R. N. and A. D. Whalen (1995). Detection of Signals in Noise (2nd
ed.). San Diego: Academic Press.
Nowlan, S. J. (1990). In D. S. Touretzky (Ed.), Advances in Neural Information
Processing Systems 2, San Mateo, CA. Morgan Kaufmann.
Pezaris, J. S., M. Sahani, and R. A. Andersen (1997). In J. M. Bower (Ed.),
Computational Neuroscience: Trends in Research, 1997.
Sahani, M., J. S. Pezaris, and R. A. Andersen (1998). In J. M. Bower (Ed.),
Computational Neuroscience: Trends in Research, 1998.
| 1456 |@word version:1 briefly:1 eliminating:1 seems:1 nd:1 heuristically:1 rhesus:1 pulse:1 covariance:5 electronics:1 selecting:1 current:2 neurophys:1 nowlan:2 must:1 john:2 additive:2 visible:1 subsequent:2 shape:7 drop:1 implying:1 generative:3 ajd:1 selected:1 isotropic:1 short:1 compo:1 filtered:2 provides:1 node:1 location:1 burst:1 along:1 ect:1 incorrect:1 prove:1 manner:1 introduce:2 expected:1 presumed:1 themselves:1 aliasing:1 brain:1 actual:1 little:1 jm:1 becomes:1 provided:2 twisted:1 medium:1 cm:1 developed:1 finding:1 exactly:4 rm:1 control:1 omit:1 appear:5 positive:1 before:1 engineering:1 mitra:1 despite:1 firing:11 path:1 mateo:1 hmms:1 range:1 directed:1 procedure:3 area:1 empirical:1 maneesh:2 adapting:1 shore:1 inferential:1 projection:1 thought:1 upsampled:1 cannot:1 close:3 selection:3 influence:2 equivalent:1 deterministic:1 bill:1 center:2 straightforward:1 independently:1 identifying:1 rule:1 insulated:2 construction:1 trigger:1 diego:1 us:1 crossing:2 trend:2 cut:1 observed:1 electrical:1 region:1 digitally:1 dempster:2 dynamic:1 ultimately:1 depend:3 rewrite:1 segment:9 exposed:2 purely:1 easily:1 hopfield:1 various:1 autocorrelated:1 derivation:3 distinct:2 describe:3 modulators:1 effective:1 etf:1 quite:1 heuristic:1 whose:1 valued:1 plausible:1 say:1 otherwise:2 itself:1 laird:1 final:1 hoc:1 advantage:1 biophysical:2 kurtosis:1 interaction:1 maximal:1 neighboring:4 aligned:1 fired:2 amplified:1 roweis:1 mcdonough:2 description:2 electrode:5 double:1 cluster:1 keck:1 bandpassed:1 object:1 stat:1 measured:3 expectationmaximization:1 soc:1 implemented:1 come:2 indicate:1 exhibiting:1 direction:1 waveform:25 filter:3 stochastic:2 transient:1 probable:3 insert:1 accompanying:1 lm:2 consecutive:1 ilv:1 currently:1 individually:1 tf:1 weighted:2 reflects:1 gaussian:8 modified:1 rather:1 inactivation:1 avoid:2 varying:2 ax:1 focus:2 likelihood:3 inference:10 flaw:1 pasadena:1 hidden:2 mth:5 sketched:2 overall:2 special:2 marginal:1 construct:2 whalen:2 shaped:1 identical:1 represents:1 broad:1 excessive:1 foreground:8 peaked:1 richard:1 few:1 simultaneously:1 national:1 phase:1 fire:6 attempt:1 detection:2 custom:1 mixture:23 pc:5 bundle:1 edge:2 indexed:1 circle:1 isolated:1 theoretical:1 earlier:2 neuromorphic:1 cost:1 applicability:1 pole:1 introducing:1 dependency:2 considerably:1 peak:2 sensitivity:1 ie:1 probabilistic:2 off:1 tip:3 together:3 andersen:7 again:1 recorded:8 borne:1 external:1 account:3 potential:6 includes:1 matter:1 explicitly:1 sloan:1 vi:11 ad:1 stream:2 performed:1 root:1 depends:1 later:1 responsibility:4 view:1 sort:1 minimize:1 square:2 kaufmann:1 ensemble:2 gathered:1 identify:1 produced:2 history:2 reestimate:1 suffers:1 touretzky:1 ed:4 proof:2 static:1 sampled:1 amplitude:2 manuscript:1 higher:1 just:1 stage:4 until:1 hand:1 replacing:1 kleinfeld:1 perhaps:1 gray:1 usa:1 effect:2 contain:1 true:2 requiring:1 functioning:1 pezaris:7 attractive:1 during:3 davis:1 whiten:1 m:1 stress:1 complete:1 invaluable:1 funding:1 common:2 functional:1 empirically:1 khz:3 refractory:4 discussed:2 slight:1 measurement:2 dag:1 pm:3 similarly:1 specification:1 cortex:1 longer:1 add:1 posterior:2 multivariate:1 recent:3 continue:1 exploited:1 caltech:6 seen:1 morgan:1 additional:2 period:2 bessel:1 signal:11 multiple:1 academic:1 cross:2 schematic:1 basic:3 histogram:2 represent:5 arisen:1 iteration:1 cell:44 wellfounded:1 background:8 else:1 source:1 operate:1 comment:1 recording:6 subject:1 jordan:2 near:1 enough:4 easy:1 isolation:1 fit:6 zi:1 converter:1 identified:5 motivated:1 handled:2 rms:2 utility:1 f:1 action:2 remark:1 useful:3 clear:1 factorial:1 diameter:1 reduced:1 multichannel:1 simplest:1 estimated:1 neuroscience:3 correctly:1 write:4 discrete:1 four:2 threshold:6 drawn:3 clarity:1 prevent:1 backward:1 graph:2 fraction:1 sum:1 deflection:3 powerful:1 arrive:1 almost:1 reasonable:1 separation:4 fee:2 entirely:1 bound:1 layer:1 activity:1 adapted:1 ri:5 flat:1 diffuse:1 nearby:3 anyone:1 extracellular:5 relatively:1 estep:1 membrane:1 across:3 remain:1 em:3 sam:1 making:2 modification:1 outlier:1 taken:3 previously:1 jennifer:1 discus:3 describing:1 count:1 mind:1 irregularly:1 tractable:1 end:1 tiling:1 operation:1 gaussians:2 permit:1 unreasonable:1 probe:2 apply:1 hierarchical:3 alternative:1 convolved:2 original:1 remaining:2 clustering:1 marginalized:1 calculating:1 giving:1 unchanged:2 objective:1 arrangement:1 spike:19 dependence:1 bialek:1 separate:1 thank:1 separating:1 hmm:2 philip:1 collected:1 spanning:1 besides:1 length:1 index:2 providing:1 difficult:1 mostly:1 fe:1 statement:1 trace:1 negative:1 rise:2 stated:1 allowing:1 upper:1 i_:1 neuron:1 wire:2 markov:2 observation:1 sm:5 parietal:1 incorrectly:1 variability:5 looking:1 digitized:1 arbitrary:1 introduced:1 connection:2 catalog:1 california:1 macaque:1 adult:1 suggested:1 below:1 pattern:1 dynamical:2 summarize:1 royal:1 power:1 event:10 overlap:6 indicator:1 sodium:1 nth:1 scheme:2 altered:1 technology:2 brief:1 sabes:1 extract:1 coupled:1 health:1 sahani:8 epoch:1 understanding:1 prior:2 acknowledgement:1 lacking:1 filtering:2 acyclic:1 digital:1 foundation:2 validation:1 rubin:1 elsewhere:2 penalized:1 supported:1 formal:1 allow:1 institute:2 fall:2 sparse:1 benefit:1 calculated:2 transition:2 autoregressive:1 forward:2 commonly:1 collection:4 projected:1 san:2 far:1 excess:2 consuming:1 latent:2 channel:3 robust:2 ca:2 gharamani:2 arrow:1 motivation:1 noise:17 body:1 benefited:1 advice:1 tl:1 position:1 inferring:1 explicit:2 lie:2 candidate:1 bower:2 resolvable:2 specific:4 pz:1 linden:1 evidence:1 tetrode:4 intrinsic:1 normalizing:1 intractable:1 false:2 conditioned:1 sorting:3 sophistication:2 likely:3 partially:1 lewicki:3 extracted:4 conditional:1 replace:1 considerable:1 change:2 determined:1 except:1 principal:2 called:2 pas:2 experimental:1 exception:1 preparation:2 avoiding:1 correlated:1 |
499 | 1,457 | Prior Knowledge in Support Vector Kernels
Bernhard Scholkopf*t, Patrice Simard t , Alex Smola t, & Vladimir Vapnikt
* Max-Planck-Institut fur biologische Kybernetik, Tiibingen, Gennany
t GMD FIRST, Rudower Chaussee 5, 12489 Berlin, Gennany
t AT&T Research, 100 Schulz Drive, Red Bank, NJ, USA
bS@first.gmd.de
Abstract
We explore methods for incorporating prior knowledge about a problem
at hand in Support Vector learning machines. We show that both invariances under group transfonnations and prior knowledge about locality in
images can be incorporated by constructing appropriate kernel functions.
1 INTRODUCTION
When we are trying to extract regularities from data, we often have additional knowledge
about functions that we estimate. For instance, in image classification tasks, there exist
transfonnations which leave class membership invariant (e.g. local translations); moreover,
it is usually the case that images have a local structure in that not all correlations between
image regions carry equal amounts of infonnation.
The present study investigates the question how to make use of these two sources of knowledge by designing appropriate Support Vector (SV) kernel functions. We start by giving
a brief introduction to SV machines (Vapnik & Chervonenkis, 1979; Vapnik, 1995) (Sec.
2). Regarding prior knowledge about invariances, we present a method to design kernel
functions for invariant classification hyperplanes (Sec. 3). The method is applicable to
invariances under the action of differentiable local 1-parameter groups of local transfonnations, e.g. translational invariance in pattern recognition. In Sec. 4, we describe kernels
which take into account image locality by using localized receptive fields. Sec. 5 presents
experimental results on both types of kernels, followed by a discussion (Sec. 6).
2 OPTIMAL MARGIN HYPERPLANES
For linear hyperplane decision functions f(x) = sgn((w? x) + b), the VC-dimension
can be controlled by controlling the nonn of the weight vector w. Given training data
(xl,yd, ... ,(Xl,Yl), Xi E RN,Yi E {?1}, a separating hyperplane which generalizes
641
Prior Knowledge in Support Vector Kernels
well can be found by minimizing
~llwl12
subject to
Yi' ((Xi'
w) + b) ~
1 for i = 1, ... , f,
(I)
the latter being the conditions for separating the training data with a margin. Nonseparable
cases are dealt with by introducing slack variables (Cortes & Vapnik 1995), but we shall
omit this modification to simplify the exposition. All of the following also applies for the
nonseparable case.
To solve the above convex optimization problem, one introduces a Lagrangian with multipliers Qi and derives the dual form of the optimization problem: maximize
i I i
L
i=l
Qi -
'2
L
QiYiQkYk(Xi . Xk)
subject to
Qi
~ 0,
i,k=l
i
L
QiYi
= 0.
(2)
i=l
It turns out that the solution vector has an expansion in terms of training examples, W =
L:=l QiYiXi, where only those Qi corresponding to constraints (1) which are met can
become nonzero; the respective examples Xi are called Support Vectors. Substituting this
expansion for W yields the decision function
(t
f(x) = sgn
o,y,(x. Xi)
+
b) .
(3)
It can be shown that minimizing (2) corresponds to minimizing an upper bound on the VC
dimension of separating hyperplanes, or, equivalently, to maximizing the separation margin
between the two classes. In the next section, we shall depart from this and modify the dot
product used such that the minimization of (2) corresponds to enforcing transformation
invariance, while at the same time the constraints (1) still hold.
3 INVARIANT HYPERPLANES
Invariance by a self-consistency argument. We face the following problem: to express
the condition of invariance of the decision function, we already need to know its coefficients which are found only during the optimization, which in turn should already take into
account the desired invariances. As a way out of this circle, we use the following ansatz:
consider decision functions f = (sgn 0 g), where g is defined as
i
g(Xj) :=
L
QiYi(Bxj . BXi)
+ b,
(4)
i=l
with a matrix B to be determined below. This follows Vapnik (1995), who suggested to
incorporate invariances by modifying the dot product used. Any nonsingular B defines a
dot product, which can equivalently be written as (Xj . AXi), with a positive definite matrix
A = BTB.
Clearly, invariance of g under local transformations of all Xj is a sufficient condition for
the local invariance of f, which is what we are aiming for. Strictly speaking, however,
invariance of g is not necessary at points which are not Support Vectors, since these lie in
a region where (sgn 0 g) is constant - however, before training, it is hard to predict which
examples will turn out to become SVs. In the Virtual SV method (Scholkopf, Burges, &
Vapnik, 1996), a first run of the standard SV algorithm is carried out to obtain an initial SV
set; similar heuristics could be applied in the present case.
Local invariance of g for each pattern Xj under transformations of a differentiable local
I-parameter group of local transformations Lt,
~ It=o g(LtXj) = 0,
ut
(5)
B. Schllikopf, P. Simard, A. 1 Smola and V. Vapnik
642
can be approximately enforced by minimizing the regularizer
1
(88tlt=og(.c Xj) )2
eL
j=1
i
(6)
t
Note that the sum may run over labelled as well as unlabelled data, so in principle one could
also require the decision function to be invariant with respect to transformations of elements
of a test set. Moreover, we could use different transformations for different patterns.
For (4), the local invariance term (5) becomes
using the chain rule. Here, 81(B.coxj . BXi) denotes the gradient of (x? y) with respect to
x, evaluated at the point (x . y) = (B.coxj . BXi). Substituting (7) into (6), using the facts
that.co = I and 81(x, y)
YT, yields the regularizer
=
1 (i
i
8)2 i (};iYi(};kYk(Bxi' BCBT BXk)
- ' " ' " OWi(Bxi)T B8 t 1t=O .ctXj
e~
~
j=1 ~=1
where
1i
C:=-'"
e~
j=1
= '"
~
(8)
i,k=1
(8-I
)(8-I
8t t=O .ctx?
8t
J
t=O
.ctx?J
)T
(9)
We now choose B such that (8) reduces to the standard SV target function IlwW in the form
2::=1 (};iYiXi into it (cf. the quadratic term of
obtained by substituting the expansion w
(2?, utilizing the dot product chosen in (4), i.e. such that (BXi . BCB T BXk) = (f3Xi .
BXk). Assuming that the Xi span the whole space, this condition becomes BT BC B B =
B T B, or, by requiring B to be nonsingular, i.e. that no information get lost during the
preprocessing, BCB T = I. This can be satisfied by a preprocessing (whitening) matrix
=
B =C-t
(10)
(modulo a unitary matrix, which we disregard), the nonnegative square root of the inverse
of the nonnegative matrix C defined in (9). In practice, we use a matrix
C>.
o< A ~
:=
(1 - A)C + AI,
(11 )
1, instead of C. As C is nonnegative, C>. is invertible. For A = 1, we recover the
standard SV optimal hyperplane algorithm, other values of A determine the trade-off between invariance and model complexity control. It can be shown that using C>. corresponds
to using an objective function 4>(w) = (1 - A) 2:i(W' ttlt=0.c t Xi)2 + Allw112.
By choosing the preprocessing matrix B according to (10), we have obtained a formulation
of the problem where the standard SV quadratic optimization technique does in effect minimize the tangent regularizer (6): the maximum of (2), using the modified dot product as in
(4), coincides with the minimum of (6) subject to the separation conditions Yi . g(Xi) 2: I,
where 9 is defined as in (4).
Note that preprocessing with B does not affect classification speed: since (Bxj . BXi) =
(Xj . BT BXi), we can precompute BT BXi for all SVs Xi and thus obtain a machine (with
modified SVs) which is as fast as a standard SV machine (cf. (4?.
Relationship to Principal Component Analysis (PCA). Let us now provide some interpretation of (10) and (9). The tangent vectors ? tt It=o.ctxj have zero mean, thus C is a
Prior Knowledge in Support Vector Kernels
643
sample estimate of the covariance matrix of the random vector s . %t It=OLtX, s E {?1}
being a random sign. Based on this observation, we call C (9) the Tangent Covariance
Matrix of the data set {Xi: i = 1, . .. ,f} with respect to the transformations Lt.
Being positive definite,1 C can be diagonalized, C = SDS T , with an orthogonal matrix
S consisting of C's Eigenvectors and a diagonal matrix D containing the corresponding
positive Eigenvalues. Then we can compute B = C-! = SD-! ST, where D- ~ is
the diagonal matrix obtained from D by taking the inverse square roots of the diagonal
elements. Since the dot product is invariant under orthogonal transformations, we may
drop the leading S and (4) becomes
l
g(Xj) = 2:>~iYi(D-t ST Xj . D-~ sT Xi)
+ b.
( 12)
i=l
A given pattern X is thus first transformed by projecting it onto the Eigenvectors of the
tangent covariance matrix C, which are the rows of ST. The resulting feature vector is
then rescaled by dividing by the square roots of C's Eigenvalues. 2 In other words, the
directions of main variance of the random vector %t It=OLtX are scaled back, thus more
emphasis is put on features which are less variant under Lt. For example, in image analysis,
if the Lt represent translations, more emphasis is put on the relative proportions of ink in
the image rather than the positions of lines. The peA interpretation of our preprocessing
matrix suggests the possibility to regularize and reduce dimensionality by discarding part
of the features, as it is common usage when doing peA.
In the present work, the ideas described in this section have only been tested in the linear
case. More generally, SV machines use a nonlinear kernel function which can be shown
to compute a dot product in a high-dimensional space F nonlinearly related to input space
via some map <P, i.e. k(x, y) = (<fl(x) . <fl(y)). In that case, the above analysis leads to a
tangent covariance matrix C in P, and it can be shown that (12) can be evaluated in terms
of the kernel function (Scholkopf, 1997). To this end, one diagonalizes C using techniques
of kernel peA (Scholkopf, Smola, & Muller, 1996).
4
KERNELS USING LOCAL CORRELATIONS
By using a kernel k(x,y) = (x? y)d, one implicitly constructs a decision boundary in
the space of all possible products of d pixels. This may not be desirable, since in natural
images, correlations over short distances are much more reliable as features than long-range
correlations are. To take this into account, we define a kernel k~l ,d2 as follows (cf. Fig. 1):
1. compute a third image z, defined as the pixel-wise product of x and y
2. sample Z with pyramidal receptive fields of diameter p, centered at ~ 11 locations
(i,j), to obtain the values Zij
3. raise each Zij to the power d 1 , to take into account local correlations within the
range of the pyramid
4. sum ztJ over the whole image, and raise the result to the power d 2 to allow for
longe-range correlations of order d2
lIt is understood that we use C>. if C is not definite (cf. (11)). Alternatively, we can below use the
pseudoinverse.
2 As an aside, note that our goal to build invariant SV machines has thus serendipitously provided
us with an approach for an open problem in SV learning, namely the one of scaling: in SV machines,
there has so far been no way of automatically assigning different weight to different directions in input
space - in a trained SV machine, the weights of the first layer (the SV s) form a subset of the training
set. Choosing these Support Vectors from the training set only gives rather limited possibilities for
appropriately dealing with different scales in different directions of input space.
644
B. SchOlkopf, P. Simard, A. 1. Smola and V. Vapnik
~d
C.'.)
j
I
(. )d l
Figure I: Kernel utilizing local correlations in images, corresponding to a dot product in a
polynomial space which is spanned mainly by local correlations between pixels (see text).
The resulting kernel will be of order d 1 ? d 2 , however, it will not contain all possible correlations of d 1 . d 2 pixels.
5 EXPERIMENTAL RESULTS
In the experiments, we used a subset of the MNIST data base of handwritten characters
(Bottou et aI., 1994), consisting of 5000 training examples and 10000 test examples at a
resolution of 20x20 pixels, with entries in [-1, 1]. Using a linear SV machine (i.e. a separating hyperplane), we obtain a test error rate of 9.8% (training 10 binary classifiers, and
using the maximum value of 9 (cf. (4? for lO-class classification); by using a polynomial
kernel of degree 4, this drops to 4.0%. In all of the following experiments, we used degree
4 kernels of various types. The number 4 was chosen as it can be written as a product of
two integers, thus we could compare results to a kernel k~l , d 2 with d 1 = d2 = 2. For the
considered classification task, results for higher polynomial degrees are very similar.
In a series of experiments with a homogeneous polynomial kernel k(x, y) = (x? y)4, using
preprocessing with Gaussian smoothing kernels of standard deviation 0.1, 0.2, ... ,1.0, we
obtained error rates which gradually increased from 4.0% to 4.3%; thus no improvement
of this performance was possible by a simple smoothing operation. Applying the Virtual
SV method (retraining the SV machine on translated SVs; Scholkopf, Burges, & Vapnik,1996) to this problem results in an improved error rate of 2.8%. For training on the full
60000 pattern set, the Virtual SV performance is 0.8% (Scholkopf, 1997).
Invariant hyperplanes. Table 1 reports results obtained by preprocessing all patterns with
B (cf. (10?, choosing different values of ..\ (cf. (11?. In the experiments, the patterns were
first rescaled to have entries in [0,1], then B was computed, using horizontal and vertical
translations, and preprocessing was carried out; finally, the resulting patterns were scaled
back again. This was done to ensure that patterns and derivatives lie in comparable regions
of RN (note that if the pattern background level is a constant -1, then its derivative is
0). The results show that even though (9) was derived for the linear case, it can lead to
improvements in the nonlinear case (here, for a degree 4 polynomial), too.
Dimensionality reduction. The above [0, 1] scaling operation is affine rather than linear,
hence the argument leading to (12) does not hold for this case. We thus only report results
on dimensionality reduction for the case where the data is kept in [0, 1] scaling from the very
Prior Knowledge in Support Vector Kernels
645
Table I: Classification error rates for modifying the kernel k(x, y) = (X?y)4 with the invariI
ant hyperplane preprocessing matrix B).. = C~ 'i ; cf. (10) and (11). Enforcing invariance
with 0.1 < A < 1 leads to improvements over the original performance (A = 1).
A
error rate in %
0.1
4.2
0.2
3.8
0.4
3.6
0.6
3.8
Table 2: Dropping directions corresponding to smaIl Eigenvalues of C (cf. (12)) leads to
substantial improvements. AIl results given are for the case A = 0.4 (cf. Table 1); degree 4
homogeneous polynomial kernel.
principal components discarded
error rate in %
beginning on. Dropping principal components which are less important leads to substantial
improvements (Table 2); cf. the explanation foIlowing (12). The results in Table 2 are
somewhat distorted by the fact that the polynomial kernel is not translation invariant, and
performs poorly on the [0, 1] data, which becomes evident in the case where none of the
principal components are discarded. Better results have been obtained using translation
invariant kernels, e.g. Gaussian REFs (Scholkopf, 1997).
Kernels using local correlations. To exploit locality in images, we used a pyramidal
receptive field kernel k;l,d 2 with diameter p = 9 (cf. Sec. 4). For d 1 = d2 = 2, we obtained an improved error rate of 3.1%, another degree 4 kernel with only local correlations
(d l = 4, d2 = 1) led to 3.4%. Albeit significantly better than the 4.0% for the degree 4
homogeneous polynomial (the error rates on the 10000 element test set have an accuracy
of about 0.1%, cf. Bottouet aI., 1994), this is still worse than the Virtual SV resultof2.8%.
As the two methods, however, exploit different types of prior knowledge, it could be expected that combining them leads to still better performance; and indeed, this yielded the
best performance of all (2.0%).
For the purpose of benchmarking, we also ran our system on the US postal service database
of 7291 +2007 handwritten digits at a resolution of 16 x 16. In that case, we obtained the
foIlowing test error rates: SV with degree 4 polynomial kernel 4.2%, Virtual SV (same kernel) 3.5%, SV with k~,2 3.6%, Virtual SV with k~,2 3.0%. The latter compares favourably
to almost all known results on that data base, and is second only to a memory-based tangentdistance nearest neighbour classifier at 2.6% (Simard, LeCun, & Denker, 1993).
6 DISCUSSION
With its rather general class of admissible kernel functions, the SV algorithm provides ample possibilities for constructing task-specific kernels. We have considered an image classification task and used two forms of domain knowledge: first, pattern classes were required
to be locally translationaIly invariant, and second, local correlations in the images were
assumed to be more reliable than long-range correlations. The second requirement can be
seen as a more general form of prior knowledge - it can be thought of as arising partiaIly
from the fact that patterns possess a whole variety of transformations; in object recognition,
for instance, we have object rotations and deformations. Typically, these transformations
are continuous, which implies that local relationships in an image are fairly stable, whereas
global relationships are less reliable.
We have incorporated both types of domain knowledge into the SV algorithm by constructing appropriate kernel functions, leading to substantial improvements on the considered
pattern recognition tasks. Our method for constructing kernels for transformation invariant SV machines, put forward to deal with the first type of domain knowledge, so far has
646
B. SchOlkopf, P. Simard, A. 1. Smola and V. Vapnik
only been applied in the linear case, which partially explains why it only led to moderate
improvements (also, we so far only used translational invariance). It is applicable for differentiable transformations - other types, e.g. for mirror symmetry, have to be dealt with
using other techniques, e.g. Virtual SVs (Scholkopf, Burges, & Vapnik, 1996). Its main
advantages compared to the latter technique is that it does not slow down testing speed,
and that using more invariances leaves training time almost unchanged. The proposed kernels respecting locality in images led to large improvements; they are applicable not only
in image classification but in all cases where the relative importance of subsets of products
features can be specified appropriately. They do, however, slow down both training and
testing by a constant factor which depends on the specific kernel used.
Both described techniques should be directly applicable to other kernel-based methods
as SV regression (Vapnik, 1995) and kernel PCA (Scholkopf, Smola, & Muller, 1996).
Future work will include the nonlinear case (cf. our remarks in Sec. 3), the incorporation
of invariances other than translation, and the construction of kernels incorporating local
feature extractors (e.g. edge detectors) different from the pyramids described in Sec. 4.
Acknowledgements. We thank Chris Burges and Uon Bottou for parts of the code and for
helpful discussions, and Tony Bell for his remarks.
References
B. E. Boser, I .M. Guyon, and V. N. Vapnik. A training algorithm for optimal margin
classifiers. In D. Haussler, editor, Proceedings of the 5th Annual ACM Workshop on
Computational Learning Theory, pages 144-152, PittSburgh, PA, 1992. ACM Press.
L. Bottou, C. Cortes, J. S. Denker, H. Drucker, I. Guyon, L. D. Jackel, Y. LeCun, U. A.
Muller, E. Sackinger, P. Simard, and V. Vapnik. Comparison of classifier methods: a
case study in handwritten digit recognition. In Proceedings of the J2th International
Conference on Pattern Recognition and Neural Networks, Jerusalem, pages 77 - 87.
IEEE Computer Society Press, 1994.
C. Cortes and V. Vapnik. Support vector networks. Machine Learning, 20:273 - 297, 1995.
B. Scholkopf. Support Vector Learning. R. Oldenbourg Verlag, Munich, 1997. ISBN
3-486-24632-1.
B. Scholkopf, C. Burges, and V. Vapnik. Incorporating invariances in support vector learning machines. In C. von der Malsburg, W. von Seelen, J. C. Vorbriiggen, and B. Sendhoff,
editors, Artificial Neural Networks-ICANN'96, pages 47 - 52, Berlin, 1996a. Springer
Lecture Notes in Computer Science, Vol. 1112.
B. Scholkopf, A. Smola, and K.-R. MulIer. Nonlinear component analysis as a kernel
eigenvalue problem. Technical Report 44, Max-Planck-Institut fUr biologische Kybernetik, 1996b. in press (Neural Computation).
P. Simard, Y. LeCun, and J. Denker. Efficient pattern recognition using a new transformation distance. In S. 1. Hanson, J. D. Cowan, and C. L. Giles, editors, Advances in
Neural Information Processing Systems 5, pages 50-58, San Mateo, CA, 1993. Morgan
Kaufmann.
P. Simard, B. Victorri, Y. LeCun, and 1. Denker. Tangent prop - a formalism for specifying
selected invariances in an adaptive network. In J. E. Moody, S. J. Hanson, and R. P.
Lippmann, editors, Advances in Neural Information Processing Systems 4, pages 895903, San Mateo, CA, 1992. Morgan Kaufmann.
V. Vapnik. The Nature of Statistical Learning Theory. Springer Verlag, New York, 1995.
V. Vapnik and A. Chervonenkis. Theory of Pattern Recognition [in Russian}. Nauka,
Moscow, 1974. (German Translation: W. Wapnik & A. Tscherwonenkis, Theorie der
Zeichenerkennung, Akademie-Verlag, Berlin, 1979).
| 1457 |@word polynomial:9 proportion:1 retraining:1 open:1 d2:5 covariance:4 uon:1 carry:1 reduction:2 initial:1 series:1 zij:2 chervonenkis:2 bc:1 diagonalized:1 assigning:1 written:2 oldenbourg:1 seelen:1 qiyi:2 drop:2 aside:1 leaf:1 selected:1 kyk:1 xk:1 beginning:1 short:1 provides:1 mulier:1 postal:1 location:1 hyperplanes:5 become:2 scholkopf:14 indeed:1 expected:1 nonseparable:2 automatically:1 becomes:4 provided:1 moreover:2 what:1 ail:1 transformation:13 nj:1 transfonnations:3 scaled:2 classifier:4 control:1 omit:1 planck:2 positive:3 before:1 understood:1 local:19 modify:1 sd:2 service:1 kybernetik:2 aiming:1 yd:1 approximately:1 emphasis:2 mateo:2 suggests:1 specifying:1 co:1 limited:1 range:4 lecun:4 testing:2 lost:1 practice:1 definite:3 digit:2 akademie:1 bell:1 significantly:1 thought:1 word:1 get:1 onto:1 put:3 applying:1 map:1 lagrangian:1 yt:1 maximizing:1 jerusalem:1 convex:1 resolution:2 smail:1 rule:1 haussler:1 utilizing:2 spanned:1 regularize:1 his:1 controlling:1 target:1 construction:1 modulo:1 homogeneous:3 designing:1 pa:1 element:3 recognition:7 database:1 region:3 trade:1 rescaled:2 ran:1 substantial:3 complexity:1 respecting:1 rudower:1 raise:2 trained:1 translated:1 various:1 regularizer:3 fast:1 describe:1 artificial:1 choosing:3 heuristic:1 solve:1 patrice:1 advantage:1 differentiable:3 eigenvalue:4 isbn:1 product:12 combining:1 poorly:1 regularity:1 requirement:1 leave:1 object:2 nearest:1 dividing:1 implies:1 met:1 direction:4 modifying:2 pea:3 vc:2 centered:1 sgn:4 virtual:7 explains:1 require:1 strictly:1 hold:2 considered:3 predict:1 substituting:3 purpose:1 applicable:4 infonnation:1 jackel:1 tscherwonenkis:1 minimization:1 clearly:1 gaussian:2 modified:2 rather:4 og:1 gennany:2 derived:1 improvement:8 fur:2 mainly:1 helpful:1 el:1 membership:1 bt:3 typically:1 transformed:1 schulz:1 pixel:5 translational:2 classification:8 dual:1 smoothing:2 fairly:1 equal:1 field:3 construct:1 lit:1 future:1 report:3 simplify:1 neighbour:1 consisting:2 zeichenerkennung:1 possibility:3 introduces:1 ztj:1 chain:1 edge:1 necessary:1 respective:1 institut:2 orthogonal:2 desired:1 circle:1 deformation:1 instance:2 increased:1 formalism:1 giles:1 introducing:1 deviation:1 subset:3 entry:2 too:1 ilww:1 sv:28 st:4 international:1 yl:1 off:1 ansatz:1 invertible:1 b8:1 moody:1 again:1 von:2 satisfied:1 containing:1 choose:1 serendipitously:1 worse:1 simard:8 leading:3 derivative:2 account:4 de:1 sec:8 coefficient:1 depends:1 root:3 doing:1 biologische:2 red:1 start:1 recover:1 minimize:1 square:3 accuracy:1 variance:1 who:1 kaufmann:2 yield:2 nonsingular:2 ant:1 dealt:2 handwritten:3 none:1 drive:1 detector:1 knowledge:14 ut:1 dimensionality:3 back:2 higher:1 vorbriiggen:1 improved:2 formulation:1 evaluated:2 done:1 though:1 smola:7 correlation:13 hand:1 horizontal:1 favourably:1 sackinger:1 nonlinear:4 defines:1 russian:1 usage:1 effect:1 usa:1 requiring:1 multiplier:1 contain:1 hence:1 nonzero:1 bxj:2 deal:1 during:2 self:1 coincides:1 trying:1 btb:1 evident:1 tt:1 performs:1 image:17 wise:1 common:1 rotation:1 bxi:9 interpretation:2 nonn:1 ai:3 consistency:1 dot:8 stable:1 iyi:2 whitening:1 base:2 moderate:1 sendhoff:1 verlag:3 binary:1 yi:3 muller:3 der:2 seen:1 minimum:1 additional:1 somewhat:1 morgan:2 determine:1 maximize:1 full:1 desirable:1 reduces:1 technical:1 unlabelled:1 long:2 controlled:1 qi:4 variant:1 regression:1 kernel:41 represent:1 pyramid:2 background:1 whereas:1 chaussee:1 victorri:1 pyramidal:2 source:1 appropriately:2 posse:1 subject:3 cowan:1 ample:1 call:1 integer:1 unitary:1 variety:1 xj:8 affect:1 reduce:1 regarding:1 idea:1 drucker:1 pca:2 speaking:1 york:1 svs:5 action:1 remark:2 generally:1 eigenvectors:2 amount:1 locally:1 gmd:2 diameter:2 exist:1 sign:1 arising:1 shall:2 dropping:2 vol:1 express:1 group:3 kept:1 sum:2 enforced:1 run:2 inverse:2 distorted:1 almost:2 guyon:2 separation:2 decision:6 scaling:3 investigates:1 comparable:1 bound:1 fl:2 layer:1 followed:1 quadratic:2 nonnegative:3 yielded:1 annual:1 constraint:2 incorporation:1 alex:1 foilowing:2 speed:2 argument:2 span:1 munich:1 according:1 precompute:1 character:1 b:1 modification:1 projecting:1 invariant:11 gradually:1 diagonalizes:1 slack:1 turn:3 german:1 know:1 end:1 generalizes:1 operation:2 denker:4 appropriate:3 original:1 denotes:1 moscow:1 cf:14 ensure:1 include:1 tony:1 malsburg:1 exploit:2 giving:1 wapnik:1 build:1 society:1 unchanged:1 ink:1 objective:1 question:1 already:2 depart:1 receptive:3 diagonal:3 gradient:1 distance:2 thank:1 berlin:3 separating:4 chris:1 enforcing:2 assuming:1 code:1 relationship:3 minimizing:4 vladimir:1 equivalently:2 x20:1 theorie:1 design:1 upper:1 vertical:1 observation:1 discarded:2 incorporated:2 rn:2 nonlinearly:1 namely:1 required:1 specified:1 hanson:2 boser:1 suggested:1 usually:1 pattern:16 below:2 ctx:2 max:2 reliable:3 explanation:1 memory:1 power:2 natural:1 brief:1 carried:2 extract:1 text:1 prior:9 acknowledgement:1 tangent:6 relative:2 lecture:1 localized:1 degree:8 affine:1 sufficient:1 principle:1 editor:4 bank:1 translation:7 row:1 lo:1 allow:1 burges:5 face:1 taking:1 boundary:1 dimension:2 axi:1 forward:1 adaptive:1 preprocessing:9 san:2 tlt:1 far:3 tiibingen:1 lippmann:1 implicitly:1 bernhard:1 dealing:1 pseudoinverse:1 global:1 pittsburgh:1 assumed:1 xi:11 alternatively:1 continuous:1 why:1 table:6 nature:1 ca:2 symmetry:1 expansion:3 bottou:3 constructing:4 domain:3 icann:1 main:2 owi:1 whole:3 ref:1 fig:1 benchmarking:1 slow:2 position:1 xl:2 lie:2 third:1 extractor:1 admissible:1 down:2 discarding:1 specific:2 cortes:3 derives:1 incorporating:3 workshop:1 mnist:1 vapnik:17 albeit:1 importance:1 mirror:1 margin:4 locality:4 lt:4 led:3 explore:1 llwl12:1 partially:1 applies:1 springer:2 corresponds:3 acm:2 prop:1 iyixi:1 goal:1 exposition:1 labelled:1 hard:1 determined:1 hyperplane:5 principal:4 called:1 invariance:21 experimental:2 disregard:1 nauka:1 support:12 bcb:2 latter:3 incorporate:1 tested:1 |