Unnamed: 0
int64 0
7.24k
| id
int64 1
7.28k
| raw_text
stringlengths 9
124k
| vw_text
stringlengths 12
15k
|
---|---|---|---|
500 | 1,458 | Receptive field formation in natural scene
environments: comparison of single cell
learning rules
Brian S. Blais
Brown University Physics Department
Providence, Rl 02912
N.lntrator
School of Mathematical Sciences
Tel-Aviv University
Ramat-Aviv, 69978 ISRAEL
H. Shouval
Institute for Brain and Neural Systems
Brown University
Providence, Rl 02912
Leon N Cooper
Brown University Physics Department and
Institute for Brain and Neural Systems
Brown University
Providence, Rl 02912
Abstract
We study several statistically and biologically motivated learning
rules using the same visual environment, one made up of natural
scenes, and the same single cell neuronal architecture. This allows
us to concentrate on the feature extraction and neuronal coding
properties of these rules. Included in these rules are kurtosis and
skewness maximization, the quadratic form of the BCM learning
rule, and single cell ICA. Using a structure removal method, we
demonstrate that receptive fields developed using these rules depend on a small portion of the distribution. We find that the
quadratic form of the BCM rule behaves in a manner similar to a
kurtosis maximization rule when the distribution contains kurtotic
directions, although the BCM modification equations are computationally simpler.
B. S. Blais, N. Intrator, H. Shouval and L N. Cooper
424
1
INTRODUCTION
Recently several learning rules that develop simple cell-like receptive fields in a
natural image environment have been proposed (Law and Cooper, 1994; Olshausen
and Field, 1996; Bell and Sejnowski, 1997). The details of these rules are different
as well as their computational reasoning, however they all depend on statistics of
order higher than two and they all produce sparse distributions.
In what follows we investigate several specific modification functions that have
the. general properties of BCM synaptic modification functions (Bienenstock et al.,
1982), and study their feature extraction properties in a natural scene environment.
Several of the rules we consider are derived from standard statistical measures
(Kendall and Stuart, 1977), such as skewness and kurtosis, based on polynomial
moments. We compare these with the quadratic form of BCM (Intrator and Cooper,
1992), though one should note that this is not the only form that could be used.
By subjecting all of the learning rules to the same input statistics and retina/LGN
preprocessing and by studying in detail the single neuron case, we eliminate possible
network/lateral interaction effects and can examine the properties of the learning
rules themselves.
We compare the learning rules and the receptive fields they form, and introduce a
procedure for directly measuring the sparsity of the representation a neuron learns.
This gives us another way to compare the learning rules, and a more quantitative
measure of the concept of sparse'representations.
2
MOTIVATION
We use two methods for motivating the use of the particular rules. One comes
from Projection Pursuit (Friedman, 1987) and the other is Independent Component
Analysis (Comon, 1994). These methods are related, as we shall see, but they
provide two different approaches for the current work.
2.1
EXPLORATORY PROJECTION PURSUIT
Diaconis and Freedman (1984) show that for most high-dimensional clouds (of
points), most low-dimensional projections are approximately Gaussian. This finding suggests that important information in the data is conveyed in those directions
whose single dimensional projected distribution is far from Gaussian.
Intrator (1990) has shown that a BCM neuron can find structure in the input
distribution that exhibits deviation from Gaussian distribution in the form of multimodality in the projected distributions. This type of deviation is particUlarly useful
for finding clusters in high dimensional data. In the natural scene environment,
however, the structure does not seem to be contained in clusters. In this work we
show that the BCM neuron can still find interesting structure in non-clustered data.
The most common measures for deviation from Gaussian distribution are skewness
and. kurtosis which are functions of the first three and four moments of the distribution respectively. Rules based on these statistical measures satisfy the BCM
conditions proposed in Bienenstock et aI. (1982), including a threshold-based stabilization. The details of these rules and some of the qualitative features of the
stabilization are different, however. In addition, there are some learning rules,
such as the ICA rule of Bell and Sejnowski (1997) and the sparse coding algorithm
of Olshausen and Field (1995), which have been used with natural scene inputs to
produce oriented receptive fields. We do not include these in our comparison be-
RF Formation in Natural Scenes: Comparison of Single Cell Learning Rules
425
cause they are not single cell learning rules, and thus detract from our immediate
goal of comparing rules with the same input structure and neuronal architecture.
2.2
INDEPENDENT COMPONENT ANALYSIS
Recently it has been claimed that the independent components of natural scenes
are the edges found in simple cells (Bell and Sejnowski, 1997). This was achieved
through the maximization of the mutual entropy of a set of mixed signals. Others
(Hyvarinen and Oja, 1996) have claimed that maximizing kurtosis can also lead
to the separation of mixed signals into independent components. This alternate
connection between kurtosis and receptive fields leads us into a discussion of ICA.
Independent Component AnalYSis (ICA) is a statistical signal processing technique
whose goal is to express a set of random variables as a linear mixture of statistically
independent variables. The problem of ICA is then to find the transformation from
the observed mixed signals to the "unmixed" independent sources. The search
for independent components relies on the fact that a linear mixture of two nonGaussian distributions will become more Gaussian than either of them. Thus,
by seeking projections which maximize deviations from Gaussian distribution, we
recover the original (independent) signals. This explains the connection of ICA to
the framework of exploratory projection pursuit.
3
SYNAPTIC MODIFICATION RULES
In this section we outline the derivation for the learning rules in this study. Neural
activity is assumed to be a positive quantity, so for biological plausibility we denote
by c the rectified activity (T(d . m), where (T(.) is a smooth monotonic function
with a positive output (a slight negative output is also allowed). (T' denotes the
derivative of the sigmoidal. The rectification is required for all rules that depend
on odd moments because these vanish in symmetric distributions such as natural
scenes. We study the following measures(Kendall and Stuart, 1977, for review) :
Skewness 1
This measures the deviation from symmetry, and is of the form:
51 = E[c3 ]j E1.5[C2].
(1)
A maximization of this measure via gradient ascent gives
=
\.5E [c (c - E[c 3]jE[c2]) (TId]
eM
where em is defined as E[c 2 ].
"V51
=
\ .5E
eM
[c (c - E[c 3]jeM) (TId]
(2)
Skewness 2
Another skewness measure is given by
52 = E[c3 ] - E1.5[C2].
(3)
This measure requires a stabilization mechanism which we achieve by requiring that
the vector of weights, denoted by m, has norm of 1. The gradient of 52 is
"V52 = 3E [c2 - cJE[c2]] = 3E
JeM) (TId] ,II
11= 1
(4)
[c (c -
m
Kurtosis 1 Kurtosis measures deviation from Gaussian distribution mainly in the
tails of the distribution. It has the form
Kl = E[c 4 ]jE2[C2] - 3.
(5)
This measure has a gradient of the form
1
"VKl = -2E
eM
[c (c2 - E[c4]jE[c2]) (TId]
1
= -2E
eM
[c (c 2 - E[c4]je M) (TId].
(6)
426
B. S. Blais, N. Intrator. H. Shouval and L N. Cooper
Kurtosis 2
As before, there is a similar form which requires some stabilization:
K2 = E[c4 ] - 3E2[C2].
(7)
This measure has a gradient of the form
'V K2 = 4E [c3 - cE[c2]]
= 3E [c(c2 -
eM )](1'd],
II m 11= 1.
(8)
Kurtosis 2 and ICA It has been shown that kurtosis, defined as
K2
= E [c 4 ]
-
3E 2 [c 2 ]
can be used for ICA(Hyvarinen and Oja, 1996).
Thus, finding the extrema of
kurtosis of the projections enables the estimation of the independent components.
They obtain the following expression
m =
~ (E- l
[ddT] E [d(m? d)3] - 3m).
(9)
which leads to an iterative "fixed-point algorithm".
Quadratic BCM The Quadratic BCM (QBCM) measure as given in (Intrator
and Cooper, 1992) is of the form
QBCM
= !E[c3 ] 3
!E2[C2].
4
(10)
Maximizing this form using gradient ascent gives the learning rule:
(11)
4
METHODS
We use 13x13 circular patches from 12 images of natural scenes, presented to the
neuron each iteration of the learning. The natural scenes are preprocessed either
with a Difference of Gaussians (DOG) filter(Law and Cooper, 1994) or a whitening
filter (Oja, 1995; Bell and Sejnowski, 1995), which eliminates the second order correlations. The moments of the output, c, are calculated iteratively, and when it is
needed (Le. K2 and 8 2 ) we also normalize the weights at each iteration.
For Oja's fixed-point algorithm, the learning was done in batches of 1000 patterns
over which the expectation values were performed. However, the covariance matrix
was calculated over the entire set of input patterns.
5
5.1
RESULTS
RECEPTIVE FIELDS
The resulting receptive fields (RFs) formed are shown in Figure 1 for both the
DOGed and whitened images. Every learning rule developed oriented receptive
fields, though some were more sensitive to the preprocessing than others. The
additive versions of kurtosis and skewness, K2 and 8 2 respectively, developed RFs
with a higher spatial frequency, and more orientations, in the whitened environment
than in the DOGed environment.
The multiplicative versions of kurtosis and skewness, Kl and 8 1 respectively, as
well as QBCM, sampled from many orientations regardless of the preprocessing.
8 1 gives receptive fields with lower spatial frequencies than either QBCM or Kl.
427
RF Formation in Natural Scenes: Comparison of Single Cell Learning Rules
This disappears with the whitened inputs, which implies that the spatial frequency
of the RF is related to the dependence of the learning rule on the second moment.
Example receptive fields using Oja's fixed-point ICA algorithm not surprisingly
look qualitatively similar to those found using the stochastic maximization of K 2 ?
The output distributions for all of the rules appear to be double exponential. This
distribution is one which we would consider sparse, but it would be difficult to
compare the sparseness of the distributions merely on the appearance of the output
distribution alone. In order to determine the sparseness of the code, we introduce
a method for measuring it directly.
Receptive Fields from Natural Scene Input
DOGed
Whitened
Output Distribution
Output Distribution
-20
0
20
-20
0
20
-20
0
20
-20
0
20
-20
0
20
-20
0
20
o
20
fIJ ~~~I A I ?11.11 ~~~I 1\ I
~.11 i1 ~~~I A I ~.II a ~~~I A I
~t! Ii1Ii ~~~VSJ ~? ? ? ~~:I 1\ I
1\1
?LI g
Figure 1: Receptive fields using DOGed (left) and whitened (right) image input
obtained from learning rules maximizing (froni top to bottom) the Quadratic BCM
objective function, Kurtosis (multiplicative), Kurtosis (additive), Skewness (multiplicative), and Skewness (additive). Shown are three examples (left to right) from
each learning rule as well as the log of the normalized output distribution, before
the application of the rectifying sigmoid.
5.2
STRUCTURE REMOVAL: SENSITIVITY TO OUTLIERS
Learning rules which are dependent on large polynomial moments, such as
Quadratic BCM and kurtosis, tend to be sensitive to the tails of the distribution.
In the case of a sparse code the outliers, or the rare and interesting events, are what
is important. Measuring the degree to which the neurons form a sparse code can
be done in a straightforward and systematic fashion.
The procedure involves simply eliminating from the environment those patterns for
which the neuron responds strongly. These patterns tend to be the high contrast
edges, and are thus the structure found in the image. The percentage of patterns
that needs to be removed in order to cause a change in the receptive field gives a
direct measure of the sparsity of the coding. The results of this structure removal
B. S. Blais, N. Intrator, H Shouval and L N. Cooper
428
are shown in Figure 2.
For Quadratic BCM and kurtosis, one need only eliminate less than one half of a
percent of the input patterns to change the receptive field significantly. To make
this more precise, we define a normalized difference between two mean zero vectors
cos a), where a is the angle between the two vectors. This measure
as V ==
has a value of zero for identical vectors, and a maximum value of one for orthogonal
vectors.
H1 -
Also shown in Figure 2 is the normalized difference as a function of the percentage
eliminated, for the different learning rules. RF differences can be seen with as little
as a tenth of a percent, which suggests that the neuron is coding the information in
a very sparse manner. Changes of around a half a percent and above are visible as
significant orientation, phase, or spatial frequency changes. Although both skewness
and Quadratic BCM depend primarily on the third moment, QBCM behaves more
like kurtosis with regards to sparse coding.
Structure Removal for BCM, Kurtosis, and Skew
0.3
?II ? ?I! ?1'1 II
?
II rI II
BCM
BCM
?
".
Kl
S,
SI
rI)
.
,
"
;'
'
.
[I
~
~0.25
2l
e<=
~
- - 0
C>
' 0
BCM
Kurtosis 1
Skew 1
,"
0.2
CI>
~0.15
iN
~
0.1
E
0
ZO.05
---
-----
0
Figure 2: Example receptive fields (left), and normalized difference measure (right),
resulting from structure removal using QBCM, Kl, and 8 1 , The RFs show the
successive deletion of top 1% of the distribution. On the right is the normalized
difference between RFs as a function of the percentage deleted in structure removal.
The maximum possible value of the difference is 1.
6
DISCUSSION
This study attempts to compare several learning rules which have some statistical
or biological motivation, or both. For a related study discussing projection pursuit
and BCM see (Press and Lee, 1996). We have used natural scenes to gain some more
insight about the statistics underlying natural images. There are several outcomes
from this study:
? All rules used, found kurtotic distributions.
? The single cell lCA rule we considered, which used the subtractive form of kurtosis, achieved receptive fields qualitatively similar to other rules discussed.
? The Quadratic BCM and the multiplicative version of kurtosis are less sensitive
to the second moments of the distribution and produce oriented RFs even when
the data is not whitened. The subtractive versions of kurtosis and skewness
are sensitive and produces oriented RFs only after sphering the data (Friedman,
1987; Field, 1994).
RF Fonnation in Natural Scenes: Comparison of Single Cell Learning Rules
429
? Both Quadratic BCM and kurtosis are sensitive to the elimination of the upper
1/2% portion of the distribution.
The sensitivity to small portions of the
distribution represents the other side of the coin of sparse coding.
? The skew rules' sensitivity to the upper parts of the distribution is not so strong.
? Quadratic BCM learning rule, which has been advocated as a projection index
for finding multi-modality in high dimensional distribution, can find projections
emphasizing high kurtosis when no cluster structure is present in the data.
ACKNOWLEDGMENTS
This work, was supported by the Office of Naval Research, the DANA Foundation
and the National Science Foundation.
References
Bell, A. J. and Sejnowski, T. J. (1995). An information-maximisation approach to blind
separation and blind deconvolution. Neural Computation, 7(6}:1129-1159.
Bell, A. J. and Sejnowski, T. J. (1997). The independent components of natural scenes
are edge filters. Vision Research. in press.
Bienenstock, E . L., Cooper, L. N., and Munro, P. W. (1982) . Theory for the development
of neuron selectivity: orientation specificity and binocular interaction in visual cortex.
Journal of Neuroscience, 2:32-48.
Comon, P. (1994). Independent component analysis, a n'ew concept? Signal Processing,
36:287-314.
Field, D. J. (1994). What is the goal of sensory coding. Neural Computation, 6:559-601.
Friedman, J. H. (1987). Exploratory projection pursuit. Journal of the American Statistical
Association, 82:249-266.
Hyvarinen, A. and Oja, E. (1996). A fast fixed-point algorithm for independent component
analysis. Int. Journal of Neural Systems, 7(6):671-687.
Intrator, N. (1990). A neural network for feature extraction. In Touretzky, D. S. and Lippmann, R. P., editors, Advances in Neural Information Processing Systems, volume 2,
pages 719-726. Morgan Kaufmann, San Mateo, CA.
Intrator, N. and Cooper, L. N. (1992) . Objective function formulation of the BCM theory of visual cortical plasticity: Statistical connections, stability conditions. Neural
Networks, 5:3-17.
Kendall, M. and Stuart, A. (1977). The Advanced Theory of Statistics, volume 1. MacMillan Publishing, New York.
Law, C. and Cooper, L. (1994). Formation of receptive fields according to the BCM
theory in realistic visual environments. Proceedings National Academy of Sciences,
91:7797-7801.
Oja, E. (1995). The nonlinear pca learning rule and signal separation - mathematical
analysis. Technical Report A26, Helsinki University, CS and Inf. Sci. Lab.
Olshausen, B. A. and Field, D. J. (1996). Emergence of simple cell receptive field properties
by learning a sparse code for natural images. Nature, 381:607-609.
Press, W. and Lee, C. W. (1996) . Searching for optimal visual codes: Projection pursuit
analysis of the statistical structure in natural scenes. In The Neurobiology of Computation: Proceedings of the fifth annual Computation and Neural Systems conference.
Plenum Publishing Corporation.
| 1458 |@word version:4 eliminating:1 polynomial:2 norm:1 covariance:1 moment:8 contains:1 current:1 comparing:1 si:1 visible:1 realistic:1 additive:3 plasticity:1 enables:1 alone:1 half:2 unmixed:1 successive:1 sigmoidal:1 simpler:1 mathematical:2 c2:12 direct:1 become:1 qualitative:1 multimodality:1 introduce:2 manner:2 ica:9 themselves:1 examine:1 multi:1 brain:2 little:1 underlying:1 israel:1 what:3 skewness:12 developed:3 finding:4 transformation:1 extremum:1 corporation:1 quantitative:1 every:1 k2:5 appear:1 positive:2 tid:5 before:2 approximately:1 mateo:1 suggests:2 co:1 ramat:1 statistically:2 acknowledgment:1 maximisation:1 procedure:2 vsj:1 bell:6 significantly:1 projection:11 specificity:1 maximizing:3 straightforward:1 regardless:1 rule:43 insight:1 stability:1 searching:1 exploratory:3 v52:1 plenum:1 particularly:1 observed:1 cloud:1 bottom:1 removed:1 environment:9 depend:4 shouval:4 derivation:1 zo:1 fast:1 sejnowski:6 formation:4 outcome:1 whose:2 statistic:4 emergence:1 kurtosis:26 interaction:2 achieve:1 academy:1 normalize:1 cluster:3 double:1 produce:4 develop:1 odd:1 school:1 advocated:1 strong:1 c:1 jem:2 come:1 implies:1 involves:1 concentrate:1 direction:2 fij:1 filter:3 stochastic:1 stabilization:4 elimination:1 explains:1 clustered:1 brian:1 biological:2 around:1 considered:1 estimation:1 sensitive:5 gaussian:7 vkl:1 office:1 derived:1 naval:1 mainly:1 contrast:1 dependent:1 eliminate:2 entire:1 bienenstock:3 lgn:1 i1:1 orientation:4 denoted:1 development:1 spatial:4 mutual:1 field:24 extraction:3 eliminated:1 identical:1 represents:1 stuart:3 look:1 subjecting:1 others:2 report:1 primarily:1 retina:1 oriented:4 oja:7 diaconis:1 national:2 phase:1 friedman:3 attempt:1 investigate:1 circular:1 mixture:2 edge:3 orthogonal:1 kurtotic:2 measuring:3 maximization:5 deviation:6 rare:1 motivating:1 providence:3 sensitivity:3 systematic:1 physic:2 lee:2 nongaussian:1 american:1 derivative:1 li:1 coding:7 cje:1 int:1 satisfy:1 blind:2 performed:1 multiplicative:4 h1:1 lab:1 kendall:3 portion:3 recover:1 rectifying:1 formed:1 kaufmann:1 rectified:1 touretzky:1 synaptic:2 frequency:4 e2:2 sampled:1 gain:1 x13:1 higher:2 formulation:1 done:2 though:2 strongly:1 binocular:1 correlation:1 nonlinear:1 aviv:2 olshausen:3 effect:1 brown:4 concept:2 requiring:1 normalized:5 symmetric:1 iteratively:1 outline:1 demonstrate:1 percent:3 reasoning:1 image:7 recently:2 common:1 sigmoid:1 behaves:2 rl:3 volume:2 tail:2 slight:1 discussed:1 association:1 significant:1 ai:1 cortex:1 whitening:1 inf:1 claimed:2 selectivity:1 discussing:1 seen:1 morgan:1 determine:1 maximize:1 signal:7 ii:7 fonnation:1 smooth:1 technical:1 plausibility:1 e1:2 whitened:6 vision:1 expectation:1 iteration:2 achieved:2 cell:11 addition:1 source:1 modality:1 eliminates:1 ascent:2 tend:2 seem:1 architecture:2 motivated:1 expression:1 pca:1 munro:1 york:1 cause:2 useful:1 percentage:3 neuroscience:1 ddt:1 shall:1 express:1 four:1 threshold:1 deleted:1 preprocessed:1 ce:1 tenth:1 merely:1 angle:1 separation:3 patch:1 quadratic:12 annual:1 activity:2 scene:16 ri:2 helsinki:1 leon:1 sphering:1 department:2 lca:1 alternate:1 according:1 em:6 biologically:1 modification:4 comon:2 outlier:2 computationally:1 equation:1 rectification:1 skew:3 mechanism:1 needed:1 studying:1 pursuit:6 gaussians:1 intrator:8 batch:1 coin:1 original:1 denotes:1 top:2 include:1 publishing:2 a26:1 seeking:1 objective:2 quantity:1 receptive:19 dependence:1 responds:1 exhibit:1 gradient:5 lateral:1 sci:1 code:5 index:1 difficult:1 negative:1 upper:2 neuron:9 immediate:1 neurobiology:1 blais:4 precise:1 dog:1 required:1 kl:5 c3:4 connection:3 c4:3 bcm:24 deletion:1 pattern:6 sparsity:2 rf:11 including:1 event:1 natural:19 advanced:1 disappears:1 review:1 removal:6 law:3 mixed:3 interesting:2 dana:1 foundation:2 degree:1 conveyed:1 editor:1 subtractive:2 surprisingly:1 supported:1 side:1 institute:2 fifth:1 sparse:10 regard:1 calculated:2 cortical:1 sensory:1 made:1 qualitatively:2 preprocessing:3 projected:2 san:1 far:1 hyvarinen:3 lippmann:1 assumed:1 search:1 iterative:1 nature:1 ca:1 tel:1 symmetry:1 motivation:2 freedman:1 allowed:1 neuronal:3 je:3 fashion:1 cooper:11 exponential:1 vanish:1 third:1 learns:1 emphasizing:1 specific:1 deconvolution:1 ci:1 sparseness:2 entropy:1 simply:1 appearance:1 visual:5 contained:1 macmillan:1 monotonic:1 relies:1 goal:3 ii1ii:1 lntrator:1 change:4 included:1 ew:1 |
501 | 1,459 | Intrusion Detection with Neural Networks
Jake Ryan*
Department of Computer Sciences
The University of Texas at Austin
Austin, TX 78712
Department of Electrical and Computer Engineering
The University of Texas at Austin
Austin, TX 78712
raven@cs.utexas.edu
mj@orac.ece . utexas.edu
Meng-Jang Lin
Risto Miikkulainen
Department of Computer Sciences
The University of Texas at Austin
Austin, TX 78712
risto@cs.utexas.edu
Abstract
With the rapid expansion of computer networks during the past few years,
security has become a crucial issue for modern computer systems. A
good way to detect illegitimate use is through monitoring unusual user
activity. Methods of intrusion detection based on hand-coded rule sets or
predicting commands on-line are laborous to build or not very reliable.
This paper proposes a new way of applying neural networks to detect
intrusions. We believe that a user leaves a 'print' when using the system;
a neural network can be used to learn this print and identify each user
much like detectives use thumbprints to place people at crime scenes. If
a user's behavior does not match hislher print, the system administrator
can be alerted of a possible security breech. A backpropagation neural
network called NNID (Neural Network Intrusion Detector) was trained
in the identification task and tested experimentally on a system of 10
users. The system was 96% accurate in detecting unusual activity, with
7% false alarm rate. These results suggest that learning user profiles is
an effective way for detecting intrusions.
1 INTRODUCTION
Intrusion detection schemes can be classified into two categories: misuse and anomaly
intrusion detection. Misuse refers to known attacks that exploit the known vulnerabilities
of the system. Anomaly means unusual activity in general that could indicate an intrusion.
?Currently: MCI Communications Corp., 9001 N. IH 35, Austin, TX 78753; jake.ryan@mci.com.
944
1. Ryan, M-J. Lin and R. Miikkulainen
If the observed activity of a user deviates from the expected behavior, an anomaly is said
to occur.
Misuse detection can be very powerful on those attacks that have been programmed in
to the detection system. However, it is not possible to anticipate all the different attacks
that could occur, and even the attempt is laborous. Some kind of anomaly detection is
ultimately necessary. One problem with anomaly detection is that it is likely to raise many
false alarms. Unusual but legitimate use may sometimes be considered anomalous. The
challenge is to develop a model of legitimate behavior that would accept novel legitimate
use.
It is difficult to build such a model for the same reason that it is hard to build a comprehensive misuse detection system: it is not possible to anticipate aU possible variations of such
behavior. The task can be made tractable in three ways: (1) Instead of general legitimate
use, the behavior of individual users in a particular system can be modeled. The task of
characterizing regular patterns in the behavior of an individual user is an easier task than
trying to do it for aU users simultaneously. (2) The patterns of behavior can be learned
for examples of legitimate use, instead of having to describe them by hand-COding possible
behaviors. (3) Detecting an intrusion real-time, as the user is typing commands, is very
difficult because the order of commands can vary a lot. In many cases it is enough to recognize that the distribution of commands over the entire login session, or even the entire
day, differs from the usual.
The system presented in this paper, NNID (Neural Network Intrusion Detector), is based on
these three ideas. NNID is a backpropagation neural network trained to identify users based
on what commands they use during a day. The system administrator runs NNID at the end
of each day to see if the users' sessions match their normal pattern. If not, an investigation
can be launched. The NNID model is implemented in a UNIX environment and consists of
keeping logs of the commands executed, forming command histograms for each user, and
learning the users' profiles from these histograms. NNID provides an elegant solution to
off-line monitoring utilizing these user profiles. In a system of 10 users, NNID was 96%
accurate in detecting anomalous behavior (i.e. random usage patterns), with a false alarm
rate of 7%. These results show that a learning offline monitoring system such as NNID
can achieve better performance than systems that attempt to detect anomalies on-line in the
command sequences, and with computationally much less effort.
The rest of the paper outlines other approaches to intrusion detection and motivates the
NNID approach in more detail (sections 2 and 3), presents the implementation and an
evaluation on a real-world computer system (sections 4 and 5), and outlines some open
issues and avenues for future work (section 6).
2
INTRUSION DETECTION SYSTEMS
Many misuse and anomaly intrusion detection systems (lDSs) are based on the general
model proposed by Denning (1987). This model is independent of the platform, system vulnerability, and type of intrusion. It maintains a set of historical profiles for users, matches
an audit record with the appropriate profile, updates the profile whenever necessary, and reports any anomalies detected. Another component, a rule set, is used for detecting misuse.
Actual systems implement the general model with different techniques (see Frank 1994;
Mukherjee et al. 1994, for an overview). Often statistical methods are used to measure how
anomalous the behavior is, that is, how different e.g. the commands used are from normal
behavior. Such approaches require that the distribution of subjects' behavior is known.
The behavior can be represented as a rule-based model (Garvey and Lunt 1991), in terms
of predictive pattern generation (Teng et al. 1990), or using state transition analysis (Porras
Intrusion Detection with Neural Networks
945
et al. 1995). Pattern matching techniques are then used to detennine whether the sequence
of events is part of normal behavior, constitutes an anomaly, or fits the description of a
known attack.
IDSs also differ in whether they are on-line or off-line. Off-line IDSs are run periodically and they detect intrusions after-the-fact based on system logs. On-line systems are
designed to detect intrusions while they are happening, thereby allowing for quicker intervention. On-line IDSs are computationally very expensive because they require continuous
monitoring. Decisions need to be made quickly with less data and therefore they are not as
reliable.
Several IDSs that employ neural networks for on-line intrusion detection have been proposed (Debar et al. 1992; Fox et al. 1990). These systems learn to predict the next command based on a sequence of previous commands by a specific user. Through a shifting
window, the network receives the w most recent commands as its input. The network is
recurrent, that is, part of the output is fed back as the input for the next step; thus, the
network is constantly observing the new trend and "forgets" old behavior over time. The
size of the window is an important parameter: If w is too small, there will be many false
positives; if it is too big, the network may not generalize well to novel sequences. The most
recent of such systems (Debar et al. 1992) can predict the next command correctly around
80% of the time, and accept a command as predictable (among the three most likely next
commands) 90% of the time.
One problem with the on-line approach is that most of the effort goes into predicting the
order of commands. In many cases, the order does not matter much, but the distribution of
commands that are used is revealing. A possibly effective approach could therefore be to
collect statistics about the users' command usage over a period of time, such as a day, and
try to recognize the distribution of commands as legitimate or anomalous off-line. This is
the idea behind the NNID system.
3 THE NNID SYSTEM
The NNID anomaly intrusion detection system is based on identifying a legitimate user
based on the distribution of commands she or he executes. This is justifiable because
different users tend to exhibit different behavior, depending on their needs of the system.
Some use the system to send and receive e-mail only, and do not require services such as
programming and compilation. Some engage in all kinds of activities including editing,
programming, e-mail, Web browsing, and so on. However, even two users that do the same
thing may not use the same application program. For example, some may prefer the "vi"
editor to "emacs", favor "pine" over "elm" as their mail utility program, or use "gcc" more
often than "cc" to compile C programs. Also, the frequency with which a command is
used varies from user to user. The set of commands used and their frequency, therefore,
constitutes a 'print' of the user, reflecting the task performed and the choice of application
programs, and it should be possible to identify the user based on this information.
It should be noted that this approach works even if some users have aliases set up as shorthands for long commands they use frequently, because the audit log records the actual
commands executed by the system. Users' privacy is not violated, since the arguments to
a command do not need to be recorded. That is, we may know that a user sends e-mail five
times a day, but we do not need to know to whom the mail is addressed.
Building NNID for a particular computer system consists of the following three phases:
1. Collecting training data: Obtain the audit logs for each user for a period of several
days. For each day and user, form a vector that represents how often the user
executed each command.
946
1 Ryan, M-J. Un and R. Miikkulainen
as
cut
expr
ghostview
Id
man
netstat
rm
tcsh
vi
awk
cvs
fgrep
gmake
fess
mesg
nm
rsh
tee
virtex
be
date
filter
grep
look
metamail
objdump
sed
test
w
61btex
df
find
gs
Ipq
rillCdir
perl
sendmail
tgif
wc
calendar
diff
finger
gzip
Ipr
more
pgp
sh
top
whereis
cat
du
fmt
hostname
Iprm
movemail
ping
sort
tput
xbiff++
chmOd
dvips
from
id
Is
mpage
ps
strip
tr
xca1c
comsat
egrep
ftp
ifConfig
machine
mt
pwd
stty
tty
xdvi
cp
elm
gcc
Ispell
mail
mv
rcp
tail
uname
xhost
cpp
emacs
gdb
fast
make
netscape
resize
tar
vacation
xterm
Table 1: The 100 commands used to describe user behavior. The number of times the user
executed each of these commands during the day was recorded, mapped into a nonlinear scale of 11
intervals, and concatenated into a l00-dimensional input vector, representing the usage pattern for
that user for that day.
2. Training: Train the neural network to identify the user based on these command
distribution vectors.
3. Perfonnance: Let the network identify the user for each new command distribution vector. If the network's suggestion is different from the actual user, or if the
network does not have a clear suggestion, signal an anomaly.
The particular implementation of NNID and the environment where it was tested is described in the next section.
4 EXPERIMENTS
The NNID system was built and tested on a machine that serves a particular research group
at the Department of Electrical and Computer Engineering at the University of Texas at
Austin. This machine has 10 total users; some are regular users, with several other users
logging in intennittently. This platfonn was chosen for three reasons:
1. The operating system (NetBSD) provides audit trail logging for accounting purposes and this option had been enabled on this system.
2. The number of users and the total number of commands executed per day are on
an order of magnitude that is manageable. Thus, the feasibility of the approach
could be tested with real-world data without getting into scalability issues.
3. The system is relatively unknown to outsiders and the users are all known to us, so
that it is likely that the data collected on it consists of nonnal user behavior (free
of intrusions).
Data was collected on this system for 12 days, resulting in 89 user-days. Instead of trying
to optimize the selection of features (commands) for the input, we decided to simply use
a set of 100 most common commands in the logs (listed in Table 1), and let the network
figure out what infonnation was important and what superfluous. Intelligent selection of
features might improve the results some but the current approach is easy to implement and
proves the point.
In order to introduce more overlap between input vectors, and therefore better generalization, the number of times a command was used was divided into intervals. There were 11
intervals, non-linearly spaced, so that the representation is more accurate at lower frequencies where it is most important. The first interval meant the command was never used; the
second that it was used once or twice, and so on until the last interval where the command
was used more than 500 times. The intervals were represented by values from 0.0 to 1.0
in 0.1 increments. These values, one for each command, were then concatenated into a
100-dimensional command distribution vector (also called user vector below) to be used as
input to the neural network.
Intrusion Detection with Neural Networks
947
The standard three-layer backpropagation architecture was chosen for the neural network.
The idea was to get results on the most standard and general architecture so that the feasibility of the approach could be demonstrated and the results would be easily replicable.
More sophisticated architectures could be used and they would probably lead to slightly
better results. The input layer consisted of 100 units, representing the user vector; the hidden layer had 30 units and the output layer 10 units, one for each user. The network was
implemented in the PlaNet Neural Network simulator (Miyata 1991).
5
RESULTS
To avoid overtraining, several training sessions were run prior to the actual experiments to
see how many training cycles would give the highest performance. The network was trained
on 8 randomly chosen days of data (65 user vectors), and its performance was tested on the
remaining 4 days (24 vectors) after epochs 30, 50, 100,200, and 300, of which 100 gave
the best performance. Four splits of the data into training and testing sets were created by
randomly picking 8 days for training. The reSUlting four networks were tested in two tasks:
1. Identifying the user vectors of the remaining 4 days. If the activation of the output
unit representing the correct user was higher than those of all other units, and
also higher than 0.5, the identification was counted as correct. Otherwise, a false
positive was counted.
2. Identifying 100 randomly-generated user vectors. If all output units had an activation less than 0.5, the network was taken to correctly identify the vector as an
anomaly (i.e. not any of the known users in the system). Otherwise, the most
highly active output unit identifies the network's suggestion. Since all intrusions
occur under one of the 10 user accounts, there is a 111 0 chance that the suggestion
would accidentally match the compromised user account and the intrusion would
not be detected. Therefore, 1/10 of all such cases were counted as false negatives.
The second test is a suggestive measure of the accuracy of the system. It is not possible to
come up with vectors that would represent a good sampling of actual intrusions; the idea
here was to generate vectors where the values for each command were randomly drawn
from the distribution of values for that command in the entire data set. In other words, the
random test vectors had the same first-order statistics as the legitimate user vectors, but
had no higher-order correlations. Therefore they constitute a neutral but realistic sample of
unusual behavior.
All four splits led to similar results. On average, the networks rejected 63% of the random
user vectors, leading to an anomaly detection rate of 96%. They correctly identified the
legitimate user vectors 93% of the time, giving a false alarm rate of 7%.
Figure 1 shows the output of the network for one of the splits. Out of 24 legitimate user
vectors, the network identified 22. Most of the time the correct output unit is very highly
activated, indicating high certainty of identification. However, the activation of the highest
unit was below 0.5 for two of the inputs, resulting in a false alarm.
Interestingly, in all false alarms in all splits, the falsely-accused user was always the same.
A closer look at the data set revealed that there were only 3 days of data on this user. He
used the system very infrequently, and the network could not learn a proper profile for him.
While it would be easy to fix this problem by collecting more data in this case, we believe
this is a problem that would be difficult to rule out in general. No matter how much data
one collects, there may still not be enough for some extremely infrequent user. Therefore,
we believe the results obtained in this rather small data set give a realistic picture of the
performance of the NNID system.
948
1. Ryan, M-l. lin and R. Miikkulainen
O~t.,ut;':
.> 4 ?
)
2 3
...
?
:l
~t? - 4 , ?
""
6
~
0
L Z
?
~
D.itrut
. .
~
.. .
..
,
~
u
"
~~-
n..,.!. -
~"":;l
e
~~t
~t<'4'"
.
?
.
. .
~
..
-
o
?
~
?
-.
"
...
_ ?
-
~
,
t
"
e
9
?
0 1 2 3
00. .......
?
5 6
?
u
~
~<
~
,
?
.
~t
~
0
1 2
Oo.t"",
~
,
"
o::.+~"
Eo
6
9
..
"
5 6 7 8 9
<
,
Out.~, ~
"
~
0
.
?
?
g
?
?
.
"
t
?
?
??
~
~~,.,,,
?
-
??
D~t~2~'56
"
5 6 7 B 9
?
"
tI
.
??
.?
.
...
D.np;;.. '"
~
?
??
D~~L
tI
?
5 6 7 B
D.~t '" ? ?
.
'
.
?
O.lt('t-rt
..
...
??
?
,
xr~ct
..
~
.
~
Out.~, ~"O'
?
~
~Zl.'5
.
ti
"
Ou<.~.
, ,
~
~
L
?"
~
t
"
o::...~"'.'~
? s
Figure 1: User identification with the NNID Network. The output layer of NNID is shown for
each of the 24 test vectors in one of the 4 splits tested. The output units are lined up from left to
right, and their activations are represented by the size of the squares. In this split there were two false
alarms: one is displayed in the top right with activation 0.01, and one in the second row from the
bottom, second column from the left with 0.35. All the other test vectors are identified correctly with
activation higher than 0.5.
6 DISCUSSION AND FUTURE WORK
An important question is, how well does the performance of NNID scale with the number
of users? Although there are many computer systems that have no more than a dozen
users, most intrusions occur in larger systems with hundreds of users. With more users,
the network would have to make finer distinctions, and it would be difficult to maintain the
same low level of false alarms. However, the rate of detecting anomalies may not change
much, as long as the network can learn the user patterns well. Any activity that differs from
the user's normal behavior would still be detected as an anomaly.
Training the network to represent many more users may take longer and require a larger
network, but it should be possible because the user profiles share a lot of common structure, and neural networks in general are good at learning such data. Optimizing the set of
commands included in the user vector, and the size of the value intervals, might also have a
large impact on performance. It would be interesting to determine the curve of performance
Intrusion Detection with Neural Networks
949
versus the number of users, and also see how the size of the input vector and the granularity
of the value intervals affect that curve. This is the most important direction of future work.
Another important issue is, how much does a user's behavior change over time? If behavior
changes dramatically, NNID must be recalibrated often or the number of false positives
would increase. Fortunately such retraining is easy to do. Since NNID parses daily activity
of each user into a user-vector, the user profile can be updated daily. NNID could then be
retrained periodically. In the current system it takes only about 90 seconds and would not
be a great burden on the system.
7
CONCLUSION
Experimental evaluation on real-world data shows that NNID can learn to identify users
simply by what commands they use and how often, and such an identification can be used
to detect intrusions in a network computer system. The order of commands does not need
to be taken into account. NNID is easy to train and inexpensive to run because it operates
off-line on daily logs. As long as real-time detection is not required, NNID constitutes a
promising, practical approach to anomaly intrusion detection.
Acknowledgements
Special thanks to Mike Dahlin and Tom Ziaja for feedback on an earlier version of this paper, and to
Jim Bednar for help with the PlaNet simulator. This research was supported in part by DOD-ARPA
contract F30602-96-1-0313, NSF grant IRI-9504317, and the Texas Higher Education Coordinating
board grant ARP-444.
References
Debar, H., Becker, M., and Siboni, D. (1992). A neural network component for an intrusion
detection system. In Proceedings of the 1992 IEEE Computer Society Symposium on
Research in Computer Security and Privacy, 240-250.
Denning, D. E. (1987). An intrusion detection model. IEEE Transactions on Software
Engineering, SE-13:222-232.
Fox, K. L., Henning, R. R., Reed, J. H., and Simonian, R. (1990). A neural network
approach towards intrusion detection. In Proceedings of the 13th National Computer
Security Conference, 125-134.
Frank, J. (1994). Artificial intelligence and intrusion detection: Current and future directions. In Proceedings of the National 17th Computer Security Conference.
Garvey, T. D., and Lunt, T. F. (1991). Model-based intrusion detection. In Proceedings of
the 14th National Computer Security Conference.
Miyata, Y. (1991). A User's Guide to PlaNet Version 5.6 -A Toolfor Constructing, Running, and Looking in to a PDP Network. Computer Science Department, University
of Colorado, Boulder, Boulder, CO.
Mukherjee, B., Heberlein, L. T., and Levitt, K. N. (1994). Network intrusion detection.
IEEE Network, 26-41.
Porras, P. A., IIgun, K., and Kemmerer, R. A. (1995). State transition analysis: A rulebased intrusion detection approach. IEEE Transactions on Software Engineering, SE21 : 181-199.
Teng, H. S., Chen, K., and Lu, S. C. (1990). Adaptive real-time anomaly detection using inductively generated sequential patterns. In Proceedings of the 1990 IEEE Symposium
on Research in Computer Security and Privacy, 278-284.
| 1459 |@word version:2 manageable:1 retraining:1 risto:2 open:1 accounting:1 detective:1 thereby:1 tr:1 interestingly:1 past:1 current:3 com:1 activation:6 must:1 planet:3 realistic:2 periodically:2 designed:1 update:1 intelligence:1 leaf:1 record:2 detecting:6 provides:2 attack:4 five:1 become:1 symposium:2 consists:3 shorthand:1 introduce:1 privacy:3 falsely:1 expected:1 rapid:1 behavior:22 ldss:1 frequently:1 simulator:2 porras:2 actual:5 window:2 what:4 kind:2 elm:2 certainty:1 rcp:1 collecting:2 ti:3 rm:1 zl:1 unit:10 grant:2 intervention:1 positive:3 service:1 engineering:4 id:6 meng:1 ipr:1 might:2 twice:1 au:2 collect:2 compile:1 co:1 programmed:1 grep:1 decided:1 practical:1 testing:1 implement:2 differs:2 backpropagation:3 xr:1 revealing:1 matching:1 word:1 refers:1 regular:2 suggest:1 get:1 selection:2 applying:1 optimize:1 misuse:6 demonstrated:1 send:1 go:1 iri:1 identifying:3 legitimate:10 rule:4 utilizing:1 enabled:1 variation:1 increment:1 updated:1 infrequent:1 user:80 anomaly:17 programming:2 engage:1 trail:1 colorado:1 trend:1 infrequently:1 expensive:1 mukherjee:2 cut:1 observed:1 bottom:1 mike:1 quicker:1 electrical:2 awk:1 cycle:1 highest:2 environment:2 predictable:1 inductively:1 ultimately:1 trained:3 raise:1 predictive:1 logging:2 easily:1 represented:3 tx:4 finger:1 cat:1 train:2 fast:1 effective:2 describe:2 detected:3 artificial:1 larger:2 otherwise:2 calendar:1 favor:1 statistic:2 sequence:4 tput:1 date:1 achieve:1 description:1 scalability:1 getting:1 p:1 bednar:1 ftp:1 depending:1 develop:1 recurrent:1 oo:1 help:1 implemented:2 c:2 indicate:1 come:1 differ:1 direction:2 correct:3 filter:1 education:1 require:4 fix:1 generalization:1 investigation:1 anticipate:2 ryan:5 around:1 considered:1 normal:4 great:1 predict:2 pine:1 vary:1 purpose:1 currently:1 infonnation:1 utexas:3 vulnerability:2 him:1 always:1 rather:1 avoid:1 tar:1 command:43 she:1 intrusion:35 detect:6 entire:3 accept:2 hidden:1 nonnal:1 issue:4 among:1 proposes:1 platform:1 special:1 fes:1 once:1 never:1 having:1 sampling:1 represents:1 look:2 constitutes:3 future:4 report:1 np:1 intelligent:1 few:1 employ:1 modern:1 randomly:4 simultaneously:1 recognize:2 comprehensive:1 individual:2 national:3 phase:1 maintain:1 attempt:2 detection:28 highly:2 evaluation:2 sh:1 behind:1 superfluous:1 compilation:1 activated:1 accurate:3 closer:1 necessary:2 daily:3 perfonnance:1 fox:2 old:1 arpa:1 column:1 earlier:1 neutral:1 hundred:1 dod:1 too:2 varies:1 thanks:1 contract:1 off:5 picking:1 quickly:1 recorded:2 nm:1 possibly:1 administrator:2 leading:1 account:3 coding:1 matter:2 mv:1 vi:2 performed:1 try:1 lot:2 observing:1 garvey:2 sort:1 maintains:1 option:1 sed:1 square:1 accuracy:1 spaced:1 identify:7 generalize:1 identification:5 lu:1 monitoring:4 justifiable:1 cc:1 finer:1 executes:1 classified:1 overtraining:1 detector:2 ping:1 whenever:1 strip:1 rsh:1 inexpensive:1 frequency:3 ut:1 ou:1 sophisticated:1 back:1 reflecting:1 higher:5 day:17 tom:1 editing:1 rejected:1 until:1 correlation:1 hand:2 receives:1 web:1 nonlinear:1 gdb:1 believe:3 building:1 usage:3 consisted:1 during:3 noted:1 trying:2 outline:2 gzip:1 orac:1 cp:1 novel:2 common:2 mt:1 overview:1 tail:1 he:2 cv:1 session:3 had:5 longer:1 operating:1 recent:2 optimizing:1 corp:1 tee:1 fortunately:1 eo:1 determine:1 period:2 signal:1 match:4 long:3 lin:3 divided:1 coded:1 feasibility:2 impact:1 anomalous:4 df:1 histogram:2 sometimes:1 pwd:1 represent:2 receive:1 addressed:1 interval:8 sends:1 crucial:1 launched:1 rest:1 probably:1 subject:1 tend:1 elegant:1 henning:1 thing:1 granularity:1 revealed:1 split:6 enough:2 easy:4 affect:1 fit:1 gave:1 architecture:3 identified:3 idea:4 avenue:1 illegitimate:1 texas:5 whether:2 utility:1 becker:1 effort:2 constitute:1 dramatically:1 clear:1 listed:1 se:1 category:1 generate:1 nsf:1 coordinating:1 correctly:4 per:1 group:1 four:3 drawn:1 year:1 run:4 unix:1 powerful:1 place:1 decision:1 prefer:1 resize:1 layer:5 ct:1 g:1 activity:7 occur:4 rulebased:1 scene:1 software:2 wc:1 argument:1 extremely:1 relatively:1 department:5 slightly:1 boulder:2 taken:2 computationally:2 lined:1 know:2 tractable:1 fed:1 end:1 unusual:5 serf:1 detennine:1 appropriate:1 jang:1 vacation:1 top:2 remaining:2 running:1 fmt:1 exploit:1 giving:1 concatenated:2 f30602:1 build:3 prof:1 jake:2 expr:1 society:1 print:4 question:1 rt:1 usual:1 said:1 exhibit:1 mapped:1 mail:6 whom:1 collected:2 reason:2 modeled:1 reed:1 difficult:4 executed:5 frank:2 negative:1 gcc:2 implementation:2 motivates:1 proper:1 unknown:1 allowing:1 displayed:1 communication:1 looking:1 jim:1 pdp:1 tty:1 retrained:1 required:1 security:7 crime:1 learned:1 distinction:1 below:2 pattern:9 challenge:1 program:4 perl:1 built:1 reliable:2 including:1 shifting:1 event:1 overlap:1 typing:1 predicting:2 representing:3 scheme:1 improve:1 alias:1 identifies:1 picture:1 created:1 ipq:1 deviate:1 prior:1 epoch:1 acknowledgement:1 recalibrated:1 par:1 generation:1 suggestion:4 interesting:1 versus:1 editor:1 share:1 austin:8 row:1 supported:1 last:1 keeping:1 free:1 accidentally:1 offline:1 guide:1 characterizing:1 curve:2 feedback:1 login:1 world:3 transition:2 replicable:1 made:2 adaptive:1 alerted:1 historical:1 counted:3 miikkulainen:4 transaction:2 l00:1 emacs:2 suggestive:1 active:1 continuous:1 un:1 compromised:1 table:2 promising:1 mj:1 learn:5 miyata:2 expansion:1 du:1 constructing:1 linearly:1 big:1 alarm:8 profile:9 levitt:1 board:1 forgets:1 hislher:1 pgp:1 dozen:1 specific:1 lunt:2 raven:1 burden:1 ih:1 false:12 sequential:1 magnitude:1 browsing:1 chen:1 easier:1 led:1 lt:1 simply:2 likely:3 forming:1 happening:1 chance:1 constantly:1 towards:1 man:1 experimentally:1 hard:1 change:3 included:1 diff:1 operates:1 called:2 teng:2 cpp:1 ece:1 mci:2 total:2 experimental:1 arp:1 indicating:1 audit:4 people:1 meant:1 violated:1 tested:7 |
502 | 146 | 281
PERFORMANCE OF SYNTHETIC NEURAL
NETWORK CLASSIFICATION OF NOISY
RADAR SIGNALS
S. C. Ahalt
F. D. Garber
I. Jouny
A. K . Krishnamurthy
Department of Electrical Engineering
The Ohio State University, Columbus, Ohio 43210
ABSTRACT
This study evaluates the performance of the multilayer-perceptron
and the frequency-sensitive competitive learning network in identifying five commercial aircraft from radar backscatter measurements. The performance of the neural network classifiers is compared with that of the nearest-neighbor and maximum-likelihood
classifiers. Our results indicate that for this problem, the neural
network classifiers are relatively insensitive to changes in the network topology, and to the noise level in the training data. While,
for this problem, the traditional algorithms outperform these simple neural classifiers, we feel that neural networks show the potential for improved performance.
INTRODUCTION
The design of systems that identify objects based on measurements of their radar
backscatter signals has traditionally been predicated upon decision-theoretic methods of pattern recognition [1]. While it is true that these methods are characterized
by a well-defined sense of optimality, they depend on the availability of accurate
models for the statistical properties of the radar measurements.
Synthetic neural networks are an attractive alternative to this problem, since they
can learn to perform the classification from labeled training data, and do not require
knowledge of statistical models [2]. The primary objectives of this investigation are;
to establish the feasibility of using synthetic neural networks for the identification
of radar objects, and to characterize the trade-oft's between neural network and
decision-theoretic methodologies for the design of radar object identification systems.
The present study is focused on the performance evaluation of systems operating on
the received radar backscatter signals of five commercial aircraft; the Boeing 707,
727, 747, the DC-lO, and the Concord. In particular, we present results for the
multi-layer perceptron and the frequency-sensitive competitive learning (FSCL)
synthetic network models [2,3] and compare these with results for the nearestneighbor and maximum-likelihood classification algorithms.
In this paper, the performance of the classification algorithms is evaluated by means
282
Ahalt, Garber, Jouny and Krishnamurthy
of computer simulation studies; the results are compared for a number of conditions
concerning the radar environment and receiver models. The sensitivity of the neural
network classifiers, with respect to the number of layers and the number of hidden
units, is investigated. In each case, the results obtained using the synthetic neural
network classifiers are compared with those obtained using an (optimal) maximumlikelihood classifier and a (minimum-distance) nearest-neighbor classifier.
PROBLEM DESCRIPTION
The radar system is modeled as a stepped-frequency system measuring radar backscatter at 8, 11, 17, and 28 MHz. The 8-28 MHz band of frequencies was chosen to
correspond to the "resonant region" of the aircraft, i.e., frequencies with wavelengths
approximately equal to the length of the object. The four specific frequencies employed for this study were pre-selected from the database maintained at The Ohio
State University ElectroScience Laboratory compact radar range as the optimal
features among the available measurements in this band [4] .
Performance results are presented below for systems modeled as having in-phase and
quadrature measurement capability (coherent systems) and for systems modeled as
having only signal magnitude measurement capability (non coherent systems). For
coherent systems, the observation vector X = [(xI, x~), (x~, x~), (x~, x~), (xt x~)] T
represents the in-phase and quadrature components of the noisy backscatter measurements of an unknown target. The elements of X correspond to the complex
scattering coefficient whose magnitude is the square root of the measured cross
section (in units of square meters, m 2 ), and whose complex phase is that of the
measured signal at that frequency. For noncoherent systems, the observation vector X = [aI, a2, a3, a4]T consists of components which are the magnitudes of the
noisy backscatter measurements corresponding to the square root of the measured
cross section.
For the simulation experiments, it is assumed that the received signal is the result
of a superposition of the backscatter signal vector S and noise vector W which is
modeled as samples from an additive white Gaussian process.
COHERENT MEASUREMENTS
In the case of a coherent radar system, the
kth
frequency component of the obser-
vation vector is given by:
xL
= (s{ + wi),
(1)
where sL and s~ are the in-phase and quadrature components of the backscatter
signal, and wi and W~ are the in-phase and quadrature components of the sample
of the additive white Gaussian noise process at that frequency. Each of these components is modeled as a zero-mean Gaussian random variable with variance u 2 /2
Performance of Synthetic Neural Network Classification
so that the total additive noise contribution at each frequency is complex-valued
Gaussian with zero mean and variance 0'2.
During operation, the neural network classifier is presented with the observation
vector, of dimension eight, consisting of the in-phase and quadrature components
of each of the four frequency measurements;
(2)
Typically, the neural net is trained using 200 samples of the observation vector X
for each of the five commercial aircraft discussed above. The desired output vectors
are of the form
(3)
=
where di,j
1 for the desired aircraft and is 0 otherwise. Thus, for example, the
output vector di for the second aircraft is 0,1,0,0,0, with a 1 appearing in the
second position.
The structure of the neural nets used can be represented by [8, nl, ... , nh, 5], where
there are 8 input neurons, ni hidden layer neurons in the h hidden layers, and 5
output neurons.
The first experiment tested the perceptron nets of varying architectures, as shown
in Figures 1, and 2. As can be seen, there was little change in performance between
the various nets.
The effects of the signal-to-noise ratio of the data observed during the training
phase on the performance of the perceptron was also investigated. The results are
presented in Figure 3. The network showed little change in performance until a
training data SNR of 20 dB was reached.
We repeated this basic experiment using a winner-take-all network, the FSCL net
[3]. Figure 4 shows that the performance of this network is also effected minimally
by changes in network architecture.
When the FSCL net is trained with noisy data, as shown in Fig. 5, the performance decreases as the SNR of the training data increases, however, the overall
performance is still very close to the performance of the multi-layer perceptron.
Our final coherent-data experiment compared the performance of the multi-layer
perceptron, the FSCL net, a max-likelihood classifier and the nearest neighbor
classifier. The results are shown in Figure 6. For this experiment, the training data
had no superimposed noise. These results show that the max-likelihood classifier
is superior, but requires full knowledge of the noise distribution. On average, the
FSCL net performs better than the perceptron, but the nearest neighbor classifier
performs better than either of the neural network models.
283
284
Ahalt, Garber, Jouny and Krishnamurthy
100
90
E-
80
E-
- 8x5x5
----- , 8x 10x 5
------
....
~
8x2Ox 5
8x3Ox5
.......... 8x40x 5
~
70
.......
60
l
~
~
\~
\
50
~
\
t:
?
40
~
30
"
20
~, .
10
~
'\r:::..
0
-30
-25
?20
-15
?10
o
?5
5
10
15
20
SNR Idbl
Figure 1: Performance of the perceptron with different number of hidden units.
100
90
c-
80
-
----_.
-----~
70
I\~,
80
l
eai
8x1Ox5.200
8x1Ox10x5.18OO
8X1Ox10x10x5.18OO
~
50
~
\
40
\
30
\
20
\
10
~
'\
0
?30
?25
?20
?15
-10
?5
o
5
10
15
20
SNR Idbl
Figure 2: Performance of the perceptron with 1, 2 and 3 hidden layers.
Perfonnance of Synthetic Neural Network Classification
100
90
r-
80
r-
----_.
------
-
..........
70
0
___
-
t--
80
l
15
~
NoIse Free
-5 db
Odb
6db
12db
20 db
,
i\
50
\ '\
40
~
30
\\
\\
20
"
\
~\
10
.~
0
-3b
-25
-20
-15
-10
"~
o
-5
-
... .......
....
.
5
10
15
20
SNRldbl
Figure 3: Performance of the perceptron for different SNR of the training data.
100
90
,.-
80
r-
----_.
8 x 10 x5
8x2Ox5
8x30x5
8 x40x5
8x50x5
------
..........
70
~
80
l
~
e
50
~
,~
\
~
40
30
\
\
20
10
\
~
.'\k.
0
-30
-25
-20
-15
-10
-5
o
5
10
15
20
SNRldbl
Figure 4: Performance of FSCL with varying no. of hidden units.
285
286
Ahalt, Garber, Jouny and Krishnamurthy
100
90
80
:--
----_.
------
:-
-
?????? 0
?
??
.....
70
,~
60
l
50
?
40
e
Noise Free
-5 db
Odb
6db
12db
\
,
\ ......
'
..
.~
30
~....
\~'.'.
\
\
20
~~
10
'.~
0
-30
-25
-20
-15
-10
-5
~
o
.. ......
5
10
15
20
SNR Idbl
Figure 5: Performance of the FSCL network for different SNR of the training data.
100
90
80
:--
----_.
------
:--
..."..~
60
?
perceptron 8x1Ox5
max. likelihood
nearest neighbor
~
70
tg
FSCL 8x1Ox5
~
~~
50
1\,
40
~ ,,
30
~\
~\
~~
20
10
0
-30
-25
-20
-15
-10
-5
o
5
10
15
20
SNR Idbl
Figure 6: Comparison of all four classifiers for the coherent data case.
Performance of Synthetic Neural Network Classification
NONCOHERENT MEASUREMENTS
For the case of a noncoherent radar system model, the
the observation vector is given by:
kth
frequency component of
(4)
where, as before, s{ and s~ are the in-phase and quadrature components of the
backscatter signal, and wI and w~ are the in-phase and quadrature components
of the additive white Gaussian noise. Hence, while the underlying noise process
is additive Gaussian, the resultant distribution of the observation components is
Rician for the non coherent system model.
For the case of non coherent measurements, the neural network classifier is presented
with a four-dimensional observation vector whose components are the magnitudes
of the noisy measurements at each of the four frequencies;
(5)
As in the coherent case, the neural net is typically trained with 200 samples for
each of the five aircraft using exemplars of the form discussed above.
The structure of the neural nets in this experiment was [4, nl, ... ,nh, 5] and the
same training and testing procedure as in the coherent case was followed. Figure 7
shows a comparison of the performance of the neural net classifiers with both the
maximum likelihood and nearest neighbor classifiers.
As before, the max-likelihood out-performs the other classifiers, with the nearestneighbor classifier is second in performance, and the neural network classifiers perform roughly the same.
CONCLUSIONS
These experiments lead us to conclude that neural networks are good candidates
for radar classification applications. Both of the neural network learning methods
we tested have a similar performance and they are both relatively insensitive to
changes in network architecture, network topology, and to the noise level of the
training data.
Because the methods used to implement the neural networks classifiers were relatively simple, we feel that the level of performance of the neural classifiers is quite
impressive. Our ongoing research is concentrating on improving neural classifier performance by introducing more sophisticated learning algorithms such as the LVQ
algorithm proposed by Kohonen [5]. We are also investigating methods of improving
the performance of the perceptron, for example, by increasing training time.
287
288
Ahalt, Garber, Jouny and Krishnamurthy
100
90
:-
80
:--
----_.
------
-
FSCL4x20x5
perceptron 4X2Ox5
max-Okellhood
near~, ralghbor-O db
---
70
I~
\\
'\ \
60
l
!5
50
\\
\\,
I:
?
40
30
.~
20
~\,
'\\ , ,
10
,~
~ ,, .
'
0
-30
-25
-20
-15
-10
-5
0
5
10
15
20
SNR rdbl
Figure 7: Comparison of all four classifiers for the non coherent data case.
References
[1] B. Bhanu, "Automatic target recognition: State of the art survey," IEEE Transactions on Aerospace and Electronic Systems, vol. AES-22, no. 4, pp. 364-379,
July 1986.
[2] R. R. Lippmann, "An Introduction to Computing with Neural Nets," IEEE
ASSP Magazine, vol. 4, no. 2, pp. 4-22, April 1987.
[3] S. C. Ahalt, A. K. Krishnamurthy, P. Chen, and D. E. Melton, "A new competitive learning algorithm for vector quantization using neural networks," Neural
Networks, 1989. (submitted).
[4] F. D. Garber, N. F. Chamberlain, and O. Snorrason, "Time-domain and
frequency-domain feature selection for reliable radar target identification," in
Proceedings of the IEEE 1988 National Itadar Conference, pp . 79-84, Ann Arbor, MI, April 20-21, 1988.
[5] T . Kohonen, Self-Organization and Associative Memory, 2nd Ed.
Springer-Veralg, 1988.
Berlin:
| 146 |@word aircraft:7 effect:1 establish:1 true:1 indicate:1 hence:1 nd:1 objective:1 laboratory:1 simulation:2 primary:1 white:3 attractive:1 traditional:1 x5:1 during:2 self:1 rician:1 kth:2 maintained:1 distance:1 require:1 berlin:1 stepped:1 investigation:1 theoretic:2 bhanu:1 performs:3 aes:1 measured:3 length:1 modeled:5 ratio:1 ohio:3 x5x5:1 superior:1 additive:5 a2:1 winner:1 insensitive:2 nh:2 boeing:1 discussed:2 design:2 selected:1 unknown:1 perform:2 superposition:1 measurement:13 observation:7 sensitive:2 neuron:3 ai:1 automatic:1 assp:1 gaussian:6 had:1 odb:2 obser:1 noncoherent:3 dc:1 five:4 impressive:1 operating:1 varying:2 consists:1 showed:1 superimposed:1 likelihood:7 aerospace:1 coherent:12 backscatter:9 sense:1 roughly:1 multi:3 seen:1 minimum:1 below:1 typically:2 pattern:1 employed:1 little:2 hidden:6 oft:1 increasing:1 signal:10 july:1 underlying:1 full:1 overall:1 classification:8 among:1 max:5 reliable:1 memory:1 characterized:1 art:1 cross:2 equal:1 concerning:1 having:2 feasibility:1 represents:1 basic:1 multilayer:1 classifier:24 unit:4 columbus:1 meter:1 before:2 national:1 engineering:1 phase:9 consisting:1 ahalt:6 organization:1 approximately:1 db:9 minimally:1 nearestneighbor:2 evaluation:1 concord:1 nl:2 near:1 range:1 lo:1 accurate:1 testing:1 architecture:3 topology:2 free:2 implement:1 perceptron:13 perfonnance:1 procedure:1 neighbor:6 desired:2 dimension:1 pre:1 mhz:2 close:1 selection:1 measuring:1 tg:1 transaction:1 introducing:1 eai:1 compact:1 lippmann:1 snr:9 band:2 investigating:1 characterize:1 focused:1 survey:1 vation:1 receiver:1 identifying:1 outperform:1 sl:1 assumed:1 synthetic:8 conclude:1 xi:1 sensitivity:1 learn:1 traditionally:1 krishnamurthy:6 vol:2 feel:2 chamberlain:1 target:3 commercial:3 four:6 magazine:1 improving:2 investigated:2 complex:3 domain:2 element:1 recognition:2 noise:12 melton:1 labeled:1 database:1 observed:1 potential:1 repeated:1 quadrature:7 fig:1 electrical:1 availability:1 coefficient:1 region:1 resonant:1 electronic:1 decrease:1 trade:1 root:2 decision:2 position:1 xl:1 candidate:1 environment:1 reached:1 competitive:3 effected:1 layer:7 capability:2 followed:1 radar:15 trained:3 depend:1 contribution:1 square:3 ni:1 specific:1 xt:1 variance:2 upon:1 correspond:2 identify:1 a3:1 quantization:1 identification:3 represented:1 various:1 optimality:1 magnitude:4 relatively:3 department:1 submitted:1 chen:1 ed:1 wavelength:1 whose:3 garber:6 quite:1 valued:1 evaluates:1 wi:3 frequency:14 otherwise:1 pp:3 resultant:1 di:2 mi:1 springer:1 noisy:5 final:1 associative:1 concentrating:1 knowledge:2 net:12 lvq:1 ann:1 sophisticated:1 kohonen:2 scattering:1 change:5 available:1 operation:1 methodology:1 improved:1 april:2 eight:1 evaluated:1 description:1 total:1 appearing:1 arbor:1 predicated:1 alternative:1 until:1 maximumlikelihood:1 object:4 a4:1 oo:2 ongoing:1 exemplar:1 nearest:6 received:2 tested:2 |
503 | 1,460 | Self-similarity properties of natural images
ANTONIO TURIEL; GERMAN MATOt NESTOR PARGA t
Departamento de Fisica Te6rica. Universidad AutOnoma de Madrid
Cantoblanco, 28049 Madrid, Spain
and JEAN-PIERRE N ADAL?
Laboratoire de Physique Statistique de I'E. N.S. , Ecole Normale Superieure
24, rue Lhomond, F-75231 Paris Cedex OS, France
Abstract
Scale invariance is a fundamental property of ensembles of natural images [1]. Their non Gaussian properties [15, 16] are less
well understood, but they indicate the existence of a rich statistical structure. In this work we present a detailed study of the
marginal statistics of a variable related to the edges in the images.
A numerical analysis shows that it exhibits extended self-similarity
[3, 4, 5]. This is a scaling property stronger than self-similarity:
all its moments can be expressed as a power of any given moment.
More interesting, all the exponents can be predicted in terms of
a multiplicative log-Poisson process. This is the very same model
that was used very recently to predict the correct exponents of
the structure functions of turbulent flows [6]. These results allow
us to study the underlying multifractal singularities. In particular
we find that the most singular structures are one-dimensional: the
most singular manifold consists of sharp edges.
Category: Visual Processing.
1
Introduction
An important motivation for studying the statistics of natural images is its relevance
for the modeling of the visual system. In particular, the epigenetic development
? e-mail: amturiel@delta.ft.uam.es
t e-mail: matog@cab.cnea.edu.ar
+To whom correspondence should be addressed. e-mail: parga@delta.ft.uam.es
?e-mail: nadal@lps.ens.fr
'Laboratoire associe au C.N.R.S. (U.R.A. 1306), a l'ENS, et aux Universites Paris VI
et Paris VII.
837
Self-similarity PropeT1ies of Natural Images
could lead to the adaptation of visual processing to the statistical regularities in the
visual scenes [8, 9, 10, 11, 12, 13]. Most of these predictions on the development of
receptive fields have been obtained using a gaussian description of the environment
contrast statistics. However non Gaussian properties like the ones found by [15, 16]
could be important. To gain further insight into non Gaussian aspects of natural
scenes we investigate the self similarity properties of an edge type variable [14].
Scale invariance in natural images is a well-established property. In particular it
appears as a power law behaviour of the power spectrum of luminosity contrast:
S(f) ex: IfIL'! (the parameter 1] depends on the particular images that has been
included in the dataset). A more detailed analysis of the scaling properties of the
luminosity contrast was done by [15, 16]. These authors noted the possible analogy
between the statistics of natural images and turbulent flows. There is however no
model to explain the scaling behaviour that they observed.
On the other hand, a large amount of effort has been put to understand the statistics
of turbulent flows and to develop predictable models (see e.g. [17]). Qualitative
and quantitative theories of fully developed turbulence elaborate on the original
argument of Kolmogorov [2]. The cascade of energy from one scale to another is
described in terms of local energy dissipation per unit mass within a box of linear
size r. This quantity, fr, is given by:
(1)
where Vi(X) is the ith component of the velocity at point x. This variable has Sel/Similarity (SS) properties that is, there is a range of scales r (called the inertial
range) where:
(2)
here < f~ > denotes the pth moment of the energy dissipation marginal distribution.
A more general scaling relation, called Extended Self-Similarity (ESS) has been
found to be valid in a much larger scale domain. This relation reads
(3)
where p(p, q) is the ESS exponent of the pth moment with respect to the qth moment. Let us notice that if SS holds then Tp = Tqp(p, q). In the following we will
refer all the moments to < f; >.
2
The Local Edge Variance
For images the basic field is the contrast c(x), that we define as the difference
between the luminosity and its average. By analogy with the definition in eq. (1) we
will consider a variable that accumulates the value of the variation of the contrast.
We choose to study two variables, defined at position x and at scale r. The variable
fh,r(X) takes contributions from edges transverse to a horizontal segment of size r:
fh,r(X)
l1
= r
xl +r
Xl
/))2
(ac(x
-ay
dy
(4)
X'={y,X2}
A vertical variable fv,r(X) is defined similarly integrating along the vertical direction.
We will refer to the value of the derivative of the contrast along a given direction
as an edge transverse to that direction. This is justified in the sense that in the
presence of borders this derivative will take a great value, and it will almost vanish
A. Turiel, G. Mato, N. Parga and l-P. Nadal
838
if evaluated inside an almost-uniformly illuminated surface. Sharp edges will be the
maxima ofthis derivative. According to its definition, ?/,r(x) ( 1 = h, v) is the local
linear edge variance along the direction 1 at scale r. Let us remark that edges are
well known to be important in characterizing images. A recent numerical analysis
suggests that natural images are composed of statistically independent edges [18].
We have analyzed the scaling properties of the local linear edge variances in a set
of 45 images taken into a forest, of 256 x 256 pixels each (the images have been
provided to us by D. Ruderman; see [16] for technical details concerning them). An
analysis of the image resolution and of finite size effects indicates the existence of
upper and lower cut-offs. These are approximately r = 64 and r = 8, respectively.
First we show that SS holds in a range of scales r with exponents Th,p and Tv,p.
This is illustrated in Fig. (1) where the logarithm of two moments of horizontal
and vertical local edge variances are plotted as a function of In r; we see that SS
holds, but not in the whole range.
ESS holds in the whole considered range; two representative graphs are shown in
Fig. (2). The linear dependence of In < ?f,r > vs In < ?f,r > is observed in
both the horizontal (l = h) and the vertical (l = v) directions. This is similar to
what is found in turbulence, where this property has been used to obtain a more
accurate estimation of the exponents of the structure functions (see e.g. [17] and
references therein) . The exponents Ph(p, 2) and Pv(p,2), estimated with a least
squares regression, are shown in Fig. (3) as a function of p . The error bars refer to
the statistical dispersion. From figs. (1-3) one sees that the horizontal and vertical
directions have similar statistical properties. The SS exponents differ, as can be
seen in Fig(I); but, surprisingly, ESS not only holds in both directions, but it does
it with the same ESS exponents, i.e. Ph(P,2) '" Pv(p, 2).
3
ESS and multiplicative processes
Let us now consider scaling models to predict the Jrdependence of the ESS exponents Pl(p, 2). (Since ESS holds, the SS exponents Tl ,p can be obtained from the
Pl(p, 2)' s by measuring 72,2). The simplest scaling hypothesis is that, for a random
variable ?r(x) observed at the scale r (such as ?/,r(x)), its probability distribution
Pr(?r(x) = ?) can be obtained from any other scale L by
Pr(?)
= a(r~ L)
PL
(a(r~ L))
(5)
From this one derives easily that a(r, L) = [~:~~P/p (for any p) and p(p, 2) ex: p; if
SS holds, Tp ex: p: for turbulent flows this corresponds to the Kolmogorov prediction
for the SS exponents [2] . Fig (3) shows that this naive scaling is violated.
This discrepancy becomes more dramatic if eq. (5) is expressed in terms of a
normalized variable. Taking ?~ = limp -+ oo < ?~+l > / < ?~ > ( that can be shown
to be the maximum value of ?r, which in fact is finite) the new variable is defined
as ir = ?r/?~ ; 0 < ir < 1. If Pr(J) is the distribution of ir, the scaling relation
eq.(5) reads Pr(J) = PL(J) ; this identity does not hold as can be seen in Fig. (4).
A way to generalize this scaling hypothesis is to say that a is no longer a constant
as in eq. (5), but an stochastic variable. Thus, one has for Pr(J) :
(6)
This scaling relation has been first introduced in the context of turbulent flows
[6, 19, 7]. Eq. (6) is an integral representation of ESS with general (not necessarily
839
Self-similarity Properties of Natural Images
linear) exponents: once the kernel G rL is chosen, the p(p, 2)'s can be predicted.
It can also be phrased in terms of multiplicative processes [20, 21] : now ir = aiL,
where the factor a itself becomes a stochastic variable determined by the kernel
G rL (1na). Since the scale L is arbitrary (scale r can be reached from any other
scale L') the kernel must obey a composition law, GrLI ?G L' L = GrL. Consequently
ir can be obtained through a cascade of infinitesimal processes G6 == G r ,r+6r'
Specific choices of G6 define different models of ESS. The She-Leveque (SL) [6]
model corresponds to a simple process such that a is 1 with probability 1 - sand
is a constant f3 with probability s. One can see that s = ll,lF In( <~tl;? and that
this stochastic process yields a log-Poisson distribution for a [22]. It also gives ESS
with exponents p(p, q) that is expressed in terms of the parameter f3 as follows [6]:
p(p,q)
1 - f3P - (1 - (3)p
(1- (3)q
= 1- f3 q -
(7)
We can now test this models with the ESS exponents obtained with the image data
set. The resulting fit for the SL model is shown in Fig. (3). Both the vertical and
horizontal ESS exponents can be fitted with {3 = 0.50 ? 0.03.
The integral representation of ESS can also be directly tested on the probability
distributions evaluated from the data. In Fig. (4) we show the prediction for Pr (f)
obtained from PL(f) using eq. (6) , compared with the actual Pr(f).
The parameter f3 allows us to predict all the ESS exponents p(p,2). To obtain the
SS exponents 7p we need another parameter. This can be chosen e.g. as 72 or as the
asymptotic exponent ~, given by f~ ex: r-t::., r ? 1; we prefer~. As 7p = 72 p(P, 2),
then from the definition of f~ one can see that ~ = -1":!13' A least square fit of 7p
was used to determine ~, obtaining ~h = 0.4 ? 0.2 for the horizontal variable and
~v = 0.5 ? 0.2. for the vertical one.
4
Multifractal analysis
Let us now partition the image in sets of pixels with the same singularity exponent
h of the local edge variance: fr ex: rh. This defines a multifractal with dimensions
D(h) given by the Legendre transform of 7p (see e.g. [17]): D(h) = inip{ph+d-7p},
where d = 2 is the dimension of the images. We are interested in the most singular
of these manifolds; let us call Doo its dimension and h min its singularity exponent.
Since f~ is the maximum value of the variable fr, the most singular manifold
is given by the set of points where fr = f~, so h min = -~. Using again that
7p = -~ (1- {3) p(P, 2) with p(P, 2) given by the SL model, one has Doo = d- (1~13)'
From our data we obtain Doo ,h = 1.3 ? 0.3 and Doo ,v = 1.1 ? 0.3. As a result
we can say that Doo,h "" Doo,v "" 1: the most singular structures are almost onedimensional. This reflects the fact that the most singular manifold consists of sharp
edges.
5
Conclusions
We insist on the main result of this work, which is the existence of non trivial
scaling properties for the local edge variances. This property appears very similar
to the one observed in turbulence for the local energy dissipation. In fact, we have
seen that the SL model predicts all the relevant exponents and that, in particular,
it describes the scaling behaviour of the sharpest edges in the image ensemble. It
would also be interesting to have a simple generative model of images which - apart
840
A. Turiel, G. Mato, N. Parga and J-P. Nadal
from having the correct power spectrum as in [23] - would reproduce the self-similar
properties found in this work.
Acknowledgements
We are grateful to Dan Ruderman for giving us his image data base. We warmly
thank Bernard Castaing for very stimulating discussions and Zhen-Su She for a
discussion on the link between the scaling exponents and the dimension of the most
singular structure. We thank Roland Baddeley and Patrick Tabeling for fruitful
discussions. We also acknowledge Nicolas BruneI for his collaboration during the
early stages of this work. This work has been partly supported by the FrenchSpanish program "Picasso" and an E.V. grant CHRX-CT92-0063.
References
[1] Field D. J., 1. Opt. Soc. Am. 4 2379-2394 (1987).
[2] Kolmogorov, Dokl. Akad. Nauk. SSSR 30, 301-305 (1941).
[3] Benzi R., Ciliberto S., Baudet C., Ruiz Chavarria G. and Tripiccione C., Europhys. Lett. 24 275-279 (1993)
[4] Benzi, Ciliberto, Tripiccione, Baudet, Massaioli, and Succi, Phys. Rev. E 48,
R29 (1993)
[5] Benzi, Ciliberto, Baudet and Chavarria Physica D 80 385-398 (1995)
[6] She and Leveque, Phys. Rev. Lett. 72,336-339 (1994).
[7] Castaing, 1. Physique II, France 6, 105-114 (1996)
[8] Barlow H. B., in Sensory Communication (ed. Rosenblith W.) pp. 217. (M.I.T.
Press, Cambridge MA, 1961).
[9] Laughlin S. B., Z. Naturf. 36 910-912 (1981).
[10] van Hateren J.H. 1. Compo Physiology A 171157-170,1992.
[11] Atick J. J. Network 3 213-251, 1992.
[12] Olshausen B.A. and Field D. J., Nature 381, 607-609 (1996).
[13] Baddeley R., Cognitive Science, in press (1997).
[14] Turiel A., Mato G., Parga N. and Nadal J.-P., to appear in Phys. Rev. Lett.,
1998.
[15] Ruderman D. and Bialek, Phys. Rev. Lett. 73,814 (1994)
[16] Ruderman D., Network 5,517-548 (1994)
[17] Frisch V., Turbulence, Cambridge Vniv. Press (1995).
[18] Bell and Sejnowski, Vision Research 37 3327-3338 (1997).
[19] Dubrulle B., Phys. Rev. Lett. 73 959-962 (1994)
[20] Novikov, Phys. Rev. E 50, R3303 (1994)
[21] Benzi, Biferale, Crisanti, Paladin, Vergassola and Vulpiani,
Physica D 65, 352-358 (1993).
[22] She and Waymire, Phys. Rev. Lett. 74, 262-265 (1995).
[23] Ruderman D., Vision Research 37 3385-3398 (1997).
Self-similarity Properties of Natural Images
In
< ?~ >
841
b
a
..
."
..
"
Inr
Inr
Figure 1: Test of SS. We plot In < ?f r > vs. In r for p = 3 and 5; r from 8 to 64
pixels. a) horizontal direction, l = h. b) vertical direction, l = v.
In
< ?~ >
b
a
,
//
?
.
/ ."
/
/ .-/ '
""....?'
....
/ '"
,/
,
"./'"
In < ?~
>~
?
?
..", /'
/
/
././
,
//
......./...-'
.
,,, ,,--
. , .......
...",/
//
./
. ~.
, .. .
//
-'
....
' ."
."....
.-/
.,..,/
/./
./
./
//
In
?f.r
< ?~ >
Figure 2: Test of ESS. We plot In <
> vs. In < ?~, r > for p=3, 5; r from 8 to
r = 64 pixels. a) horizontal direction, l = h. b) vertical direction, l = v.
A. Turiel, G. Mato, N. Parga and J-P. Nadal
842
p(p, 2)
b
a
12
1.
p
p
Figure 3: ESS exponents p(p, 2), for the vertical and horizontal variables. a) horizontal direction, Ph (P, 2) . b) vertical direction, pv (p, 2). The solid line represents
the fit with the SL model. The best fit is obtained with (3v '" (3h '" 0.50.
P
18
16
14
12
10
8
++
+
+
+
6
+
4
+
2
++++
+
+
+
+
0
0
0.05
0.1
0.15
0.2
f
Figure 4: Verification of the validity of the integral representation of ESS, eq.(6)
with a log-Poisson kernel, for horizontal local edge variance. The largest scale is
L = 64. Starting from the histogram Pdf) (denoted with crosses), and using a
log-Poisson distribution with parameter (3 = 0.50 for the kernel GrL , eq.(6) gives
a prediction for the distribution at the scale r = 16 (squares). This has to be
compared with the direct evaluation of Pr (I) (diamonds). Similar results hold for
other pairs of scales. Although not shown in the figure, the test for vertical case is
as good as for horizontal variable.
| 1460 |@word stronger:1 dramatic:1 solid:1 moment:7 ecole:1 qth:1 must:1 numerical:2 partition:1 limp:1 plot:2 v:3 generative:1 adal:1 es:18 ith:1 compo:1 along:3 direct:1 qualitative:1 consists:2 dan:1 inside:1 insist:1 actual:1 becomes:2 spain:1 provided:1 underlying:1 mass:1 what:1 nadal:5 ail:1 developed:1 quantitative:1 unit:1 grant:1 appear:1 understood:1 local:9 accumulates:1 approximately:1 au:1 therein:1 suggests:1 range:5 statistically:1 lf:1 bell:1 cascade:2 physiology:1 statistique:1 integrating:1 turbulence:4 put:1 context:1 fruitful:1 starting:1 resolution:1 insight:1 his:2 variation:1 grl:2 hypothesis:2 velocity:1 cut:1 predicts:1 observed:4 ft:2 environment:1 predictable:1 grateful:1 segment:1 easily:1 kolmogorov:3 sejnowski:1 picasso:1 europhys:1 jean:1 larger:1 say:2 s:10 statistic:5 transform:1 itself:1 fr:5 adaptation:1 relevant:1 superieure:1 nauk:1 description:1 regularity:1 oo:1 develop:1 ac:1 novikov:1 eq:8 soc:1 predicted:2 indicate:1 differ:1 direction:13 sssr:1 correct:2 stochastic:3 sand:1 behaviour:3 opt:1 singularity:3 pl:5 physica:2 hold:9 considered:1 aux:1 great:1 predict:3 early:1 fh:2 estimation:1 largest:1 reflects:1 offs:1 gaussian:4 normale:1 sel:1 she:4 indicates:1 contrast:6 sense:1 am:1 relation:4 reproduce:1 france:2 interested:1 pixel:4 denoted:1 exponent:23 development:2 autonoma:1 marginal:2 field:4 once:1 f3:4 having:1 represents:1 discrepancy:1 composed:1 nestor:1 ciliberto:3 epigenetic:1 brunei:1 investigate:1 inr:2 evaluation:1 physique:2 analyzed:1 accurate:1 edge:17 integral:3 logarithm:1 plotted:1 fitted:1 modeling:1 ar:1 tp:2 measuring:1 crisanti:1 fundamental:1 universidad:1 na:1 again:1 choose:1 cognitive:1 derivative:3 tqp:1 de:4 vi:2 depends:1 multiplicative:3 reached:1 contribution:1 square:3 ir:5 variance:7 ensemble:2 yield:1 castaing:2 generalize:1 sharpest:1 parga:6 mato:4 explain:1 phys:7 rosenblith:1 ed:1 definition:3 infinitesimal:1 energy:4 pp:1 gain:1 dataset:1 inertial:1 appears:2 done:1 box:1 evaluated:2 stage:1 atick:1 hand:1 horizontal:12 ruderman:5 su:1 o:1 defines:1 olshausen:1 effect:1 validity:1 normalized:1 barlow:1 read:2 illustrated:1 ll:1 during:1 self:9 noted:1 benzi:4 pdf:1 ay:1 dissipation:3 l1:1 image:22 recently:1 rl:2 onedimensional:1 refer:3 composition:1 cambridge:2 turiel:5 similarly:1 similarity:9 surface:1 longer:1 base:1 patrick:1 recent:1 apart:1 multifractal:3 seen:3 determine:1 f3p:1 ii:1 technical:1 cross:1 concerning:1 roland:1 prediction:4 basic:1 regression:1 vision:2 poisson:4 histogram:1 kernel:5 justified:1 doo:6 addressed:1 laboratoire:2 singular:7 cedex:1 flow:5 call:1 presence:1 fit:4 effort:1 remark:1 antonio:1 detailed:2 amount:1 ph:4 category:1 simplest:1 sl:5 notice:1 delta:2 estimated:1 per:1 graph:1 almost:3 dy:1 scaling:14 prefer:1 illuminated:1 correspondence:1 scene:2 x2:1 phrased:1 aspect:1 argument:1 min:2 tv:1 according:1 legendre:1 describes:1 lp:1 rev:7 pr:8 g6:2 taken:1 german:1 turbulent:5 studying:1 uam:2 obey:1 pierre:1 existence:3 original:1 denotes:1 giving:1 quantity:1 receptive:1 dependence:1 bialek:1 exhibit:1 thank:2 link:1 manifold:4 mail:4 whom:1 trivial:1 akad:1 diamond:1 upper:1 vertical:12 dispersion:1 finite:2 acknowledge:1 luminosity:3 extended:2 communication:1 sharp:3 arbitrary:1 transverse:2 introduced:1 pair:1 paris:3 fv:1 established:1 bar:1 dokl:1 program:1 power:4 natural:10 zhen:1 naive:1 r29:1 acknowledgement:1 asymptotic:1 law:2 fully:1 interesting:2 analogy:2 verification:1 collaboration:1 surprisingly:1 supported:1 allow:1 understand:1 laughlin:1 characterizing:1 taking:1 van:1 dimension:4 lett:6 valid:1 rich:1 sensory:1 author:1 pth:2 spectrum:2 nature:1 nicolas:1 obtaining:1 forest:1 necessarily:1 rue:1 domain:1 main:1 rh:1 motivation:1 border:1 whole:2 fig:9 representative:1 madrid:2 en:2 elaborate:1 tl:2 lhomond:1 position:1 pv:3 baudet:3 xl:2 vanish:1 ruiz:1 specific:1 departamento:1 derives:1 ofthis:1 cab:1 vii:1 visual:4 expressed:3 corresponds:2 ma:1 stimulating:1 identity:1 consequently:1 included:1 determined:1 uniformly:1 called:2 bernard:1 invariance:2 e:2 partly:1 relevance:1 violated:1 hateren:1 baddeley:2 tested:1 ex:5 |
504 | 1,461 | Refractoriness and Neural Precision
Michael J. Berry n and Markus Meister
Molecular and Cellular Biology Department
Harvard University
Cambridge, MA 02138
Abstract
The relationship between a neuron's refractory period and the precision of
its response to identical stimuli was investigated. We constructed a model of
a spiking neuron that combines probabilistic firing with a refractory period.
For realistic refractoriness, the model closely reproduced both the average
firing rate and the response precision of a retinal ganglion cell. The model is
based on a "free" firing rate, which exists in the absence of refractoriness.
This function may be a better description of a spiking neuron's response
than the peri-stimulus time histogram.
1 INTRODUCTION
The response of neurons to repeated stimuli is intrinsically noisy. In order to take this
trial-to-trial variability into account, the response of a spiking neuron is often described
by an instantaneous probability for generating an action potential. The response
variability of such a model is determined by Poisson counting statistics; in particular, the
variance in the spike count is equal to the mean spike count for any time bin (Rieke,
1997). However, recent experiments have found far greater precision in the vertebrate
retina (Berry, 1997) and the HI interneuron in the fly visual system (de Ruyter, 1997). In
both cases, the neurons exhibited sharp transitions between silence and nearly maximal
firing. When a neuron is firing near its maximum rate, refractoriness causes spikes to
become more regularly spaced than for a Poisson process with the same firing rate. Thus,
we asked the question: does the refractory period play an important role in a neuron's
response precision under these stimulus conditions?
2 FIRING EVENTS IN RETINAL GANGLION CELLS
We addressed the role of refractoriness in the precision of light responses for retinal
ganglion cells.
2.1 RECORDING AND STIMULATION
Experiments were performed on the larval tiger salamander. The retina was isolated from
the eye and superfused with oxygenated Ringer's solution. Action potentials from retinal
Refractoriness and Neural Precision
111
ganglion cells were recorded extracellularly with a multi-electrode array, and their spike
times measured relative to the beginning of each stimulus repeat (Meister, 1994).
Spatially uniform white light was projected from a computer monitor onto the
photoreceptor layer. The intensity was flickered by choosing a new value at random from
a Gaussian distribution (mean J, standard deviation oJ) every 30 ms. The mean light level
(J= 4'10-3 W/m2) corresponded to photopic (daylight) vision. Contrast C is defined here
as the temporal standard deviation of the light intensity divided by the mean, C = 01/ I.
Recordings extended over 60 repeats of a 60-sec segment of random flicker.
The qualitative features of ganglion cell responses to random flicker stimulation at 35 %
contrast are seen in Fig. 1. First, spike trains had extensive periods in which no spikes
were seen in 60 repeated trials. Many spike trains were sparse, in that the silent periods
covered a large fraction of the total stimulus time. Second, during periods of firing, the
peri-stimulus time histogram (PSTH) rose from zero to the maximum firing rate
(-200 Hz) on a time scale comparable to the time interval between spikes (-10 ms). We
have argued that these responses are better viewed as a set of discrete firing "events" than
as a continuously varying firing rate (Berry, 1997). In general, the firing events were
bursts containing more than one spike (Fig. IB). Identifiable firing events were seen
across cell types; similar results were also found in the rabbit retina (Berry, 1997).
>.
'..-wc
Q)
..c
2
A
1
0
60
~
?c
40
~'
If:
N
I
, I
'I~~
~
.
?' .I
1
\
~
,.~.
\>.~
~-(-
/..'1
'~*':
>f.
0
!
:
) ..0:
~~,{
20
.'J:
I'r
'~
.~'
:~
~~
I-
.- 300
I'
W?
8
,It:
C
Q)
ni
a:
0
43.4
43.5
43.6
43.7
43.8
Time (5)
Figure 1: Response of a salamander ganglion cell to random flicker stimulation.
(A) Stimulus intensity in units of the mean for a O.5-s segment, (B) spike rasters
from 60 trials, and (C) the firing rate r(t).
2.2 FIRING EVENT PRECISION
Discrete episodes of ganglion cell firing were recognized from the PSTH as a contiguous
period of firing bounded by periods of complete silence. To provide a consistent
demarcation of firing events, we drew the boundaries of a firing event at minima v in the
PSTH that were significantly lower than neighboring maxima PI and P2' such that
~ PIP2 ~ ? with 95 % confidence (Berry, 1997). With these boundaries defined, every
spike in each trial was assigned to exactly one firing event.
Iv
M. J Berry and M. Meister
112
Measurements of both timing and number precision can be obtained if the spike train is
parsed into such firing events. For each firing event i, we accumulated the distribution of
spike times across trials and calculated several statistics: the average time Tj of the first
spike in the event and its standard deviation OTj across trials, which quantified the
temporal jitter of the first spike; similarly, the average number N j of spikes in the event
and its variance ONj 2 across trials, which quantified the precision of spike number. In
trials that contained zero spikes for event i, no contribution was made to Tj or OTj , while
a value of zero was included in the calculation of Nj and ONj 2 .
For the ganglion cell shown in Fig. 1, the temporal jitter oT of the first spike in an event
was very small (1 to 10 ms). Thus, repeated trials of the same stimulus typically elicit
action potentials with a timing uncertainty of a few milliseconds. The temporal jitter of
all firing events was distilled into a single number Tby taking the median o"er all events.
The variance ON 2 in the spike count was remarkably low as well: it often approached the
lower bound imposed by the fact that individual trials necessarily produce integer spike
counts. Because ON 2 ? N for all events, ganglion cell spike trains cannot be completely
characterized by their firing rate (Berry, 1997). The spike number precision of a cell was
average variance over events and dividing by the average
assessed by comp.utin,fo.
spike count: F =(ON- Jj(N). This quantity, also known as the Fano factor, has a value
of one for a Poisson process with no refractoriness.
tHe
3 PROBABILISTIC MODELS OF A SPIKE TRAIN
We start by reviewing one of the simplest probabilistic models of a spike train, the
inhomogeneous Poisson model. Here, the measured spike times {t j } are used to estimate
the instantaneous rate r(t) of spike generation during a time Lit . This can be written
fonnallyas
where M is the number of repeated stimulus trials and e( x) is the Heaviside function
x~O}
1
()
ex=
o
.
x<O
We can randomly generate a sequence of spike trains from a set of random numbers
between zero and one: {aj } with a j E (0,1]. If there is a spike at time tj , then the next
spike time tj+1 is found by numerically solving the equation
1,+,
-Ina;+! =
Jr(t)dt
.
t,
3.1 INCLUDING AN ABSOLUTE REFRACTORY PERIOD
In order to add refractoriness to the Poisson spike-generator, we expressed the firing rate
as the product of a "free" firing rate q(t) , which obtains when the neuron is not
refractory, and a recovery function w(t), which describes how the neuron recovers from
refractoriness (Johnson, 1983; Miller, 1985). When the recovery function is zero, spiking
is not possible; and when it is one, spiking is not affected. The modified rule for
selecting spikes then becomes
I,... ,
-lna j +1 =
Jq(t)w(t-t;)dt
.
"
For an absolute refractory period of time J1, the weight function is zero for times between
o and J1 and one otherwise
Refractoriness and Neural Precision
113
w{t;,u) = 1- B{t )B(,u - t)
Because the refractory period may exclude spiking in a given time bin, the probability of
firing a spike when not prevented by the refractory period is higher than predicted by
r( t). This free firing rate q( t ; ,u) can be estimated by excluding trials where the neuron is
unable to fire due to refractoriness
The sum is restricted to spike times ti nearest to the time bin on a given trial. This
restriction follows from the assumption that the recovery function only depends on the
time since the last action potential. Notice that this new probability obeys the inequality
q( t ) ~ p( t) and also that it depends upon the refractory period ,u.
-N
I
4.5
Q)
~
a:
t - + - -i- -1- -f -
-f" - ~- - .. - -. - - - --
?
4.3
OJ
c
?c
u::
?
4.1
1.00
lJ...
.... 0.75
0
( ,)
ctS
0.50
0
c
ctS
u.. 0.25
? ?
? ?
----------------------.--.-~
0.00
-
-
5
.,..
4
=:
3
e!
2
(J)
E
....
Q)
::s
8.
1
~
0
E
?_____________
? ? ? ? _?___? __ S __ ?
o
1
2
3
4
__ ,.. __?
5
Refractory Period (ms)
Figure 2: Results for model spike trains with an absolute refractory period.
(A) Mean firing rate averaged over a 60-s segment (circles), (B) Fano factor
F, a measure of spike number precision in an event (triangles), and (C)
temporal jitter 't'(diamonds) plotted versus the absolute refractory period ,u.
Shown in dotted in each panel is the value for the real data.
With this definition of the free firing rate, we can now generate spike trains with the same
first order statistics (i.e., the average firing rate) for a range of values of the refractory
period ,u. For each value of ,u, we can then compare the second order statistics (i.e., the
precision) of the model spike trains to the real data. To this end, the free rate q( t) was
M. 1 Berry and M. Meister
114
calculated for a 60-s segment of the response to random flicker of the salamander
ganglion cell shown in Fig. 1. Then, q(t) was used to generate 60 spike trains. Firing
events were identified in the set of model spike trains, and their precision was calculated.
Finally, this procedure was repeated 10 times for each value ofthe refractory period.
Figure 2A plots the firing rate (circles) generated by the model, averaged over the entire
60-s segment of random flicker with error bars equal to the standard deviation of the rate
among the 10 repeated sets. The firing rate of the model matches the actual firing rate for
the real ganglion cell (dashed) up to refractory periods of J1 :: 4 ms, although the
deviation for larger refractory periods is still quite small. For large enough values of the
absolute refractory period, there will be inter-spike intervals in the real data that are
shorter than J1. In this case, the free firing rate q( t) cannot be enhanced enough to match
the observed firing rate.
While the mean firing rate is approximately constant for refractory periods up to 5 ms, the
precision changes dramatically. Figure 2B shows that the Fano factor F (triangles) has the
expected value of 1 for no refractory period, but drops to ~ 0.2 for the largest refractory
period. In Fig. 2e, the temporal jitter 'l" (diamonds) also decreases as refractoriness is
added, although the effect is not as large as for the precision of spike number. The
sharpening of temporal precision is due to the fact that the probability q( t) rises more
steeply than r(t) (see Fig. 4), so that the first spike occurs over a narrower range oftimes.
The number precision of the model matches the real data for J1 = 4 to 4.5 ms and the
timing precision matches for ~ :: 4 ms. Therefore, a probabilistic spike generator with an
absolute refractory period can match both the average firing rate and the precision of a
retinal ganglion cell's spike train with roughly the same value of one free parameter.
3.2 USING A RELATIVE REFRACTORY PERIOD
Salamander ganglion cells typically have a relative refractory period that lasts beyond
their absolute refractory period. This can be seen in Fig. 3A from the distribution of interspike intervals P{L1) for the ganglion cell shown above - the absolute refractory period
lasts for only 2 ms, while relative refractoriness extends to ~ 5 ms. We can include the
effects of relative refractoriness by using weight values in w( t) that are between zero and
one. Figure 3 illustrates a parameter-free method for determining this weight function. If
there were no refractoriness and a neuron had a constant firing rate q, then the inter-spike
interval distribution would drop exponentially. This behavior is seen from the curve fit in
Fig. 3A for intervals in the range 5 to 10 ms. The recovery function w(t) can then be
found from the inter-spike interval distribution (Berry, 1998)
Notice in Fig. 3B, that the recovery function w(t) is zero out to 3 ms, rises almost
linearly between 3 and 5 ms, and then reaches unity beyond 5 ms.
Using the weight function shown in Fig. 3B, the free firing rate q(t) was calculated and
10 sets of 60 spike trains were generated. The results, summarized in Table 1, give very
close agreement with the real data:
Table 1: Results for a Relative Refractory Period
QUANTITY
REAL DATA
MODEL
STD. DEV.
Firing Rate
Timing Precision 'l"
Number Precision F
4.43 Hz
3.20ms
0.250
4.44 Hz
2.95 ms
0.266
0.017 Hz
0.09ms
0.004
Refractoriness and Neural Precision
115
Thus, a Poisson spike generator with a relative refractory period reproduces the measured
precision. A similar test, performed over a population of ganglion cells, also yielded close
agreement (Berry, 1998).
1000
>u
c:
A
0
0
0000
0
100
0
Q)
:::J
0
CT
....
Q)
u.
10
~OOO
0
0
000 0
0
1
0
2
4
6
8
10
12
10
12
Inter-Spike Interval (ms)
c:
0
U
c:
1.0
B
:::J
U.
~
0.5
Q)
>
0
u
Q)
0.0
a::
0
2
4
6
8
Time (ms)
Figure 3: Determination of the relative refractory period. (A) The interspike interval distribution (diamonds) is fit by an exponential curve (solid),
resulting in (B) the recovery function.
Not only is the average firing rate well-matched by the model, but the firing rate in each
time bin is also very similar. Figure 4A compares the firing rate for the real neuron to that
generated by the model. The mean-squared error between the two is 4 %, while the
counting noise, estimated as the variance of the standard error divided by the variance of
r{t) , is also 4 %. Thus, the agreement is limited by the finite number of repeated trials.
Figure 4B compares the free firing rate q( t) to the observed rate firing r{ t). q( t) is equal
to r(t) at the beginning of a firing event, but becomes much larger after several spikes
have occurred. In addition, q(t) is generally smoother than r(t), because there is a
greater enhancement in q(t) at times following a peak in r(t).
In summary, the free firing rate q( t) can be calculated from the raw spike train with no
more computational difficulty than r{t), and thus can be used for any spiking neuron.
Furthermore, q(t) has some advantages over r(t): 1) in conjunction with a refractory
spike-generator, it produces the correct response precision; 2) it does not saturate at high
firing rates, so that it can continue to distinguish gradations in the neuron's response.
Thus, q( t) may prove useful for constructing models of the input-output relationship of a
spiking neuron (Berry, 1998).
Acknowledgments
We would like to thank Mike DeWeese for many useful conversations. One of us, MJB,
acknowledges the support of the National Eye Institute. The other, MM, acknowledges
the support of the National Science Foundation.
Figure 4: Illustration of the free fIring rate. (A) The observed fIring rate r(t)
for real data (solid) is compared to that from the model (dotted). (B) The
free rate q(t) (thick) is shown on the same scale as r(t) (thin). All rates
used a time bin of 0.25 ms and boxcar smoothing 'over 9 bins.
References
Berry, M. J., D. K. Warland, and M. Meister, The Structure and Precision ofRetinal
Spike Trains. PNAS, USA, 1997.94: pp. 5411-5416.
Berry II, M. J. and Markus Meister, Refractoriness and Neural Precision. 1. Neurosci.,
1998. in press.
De Ruyter van Steveninck, R. R., G. D. Lewen, S. P. Strong, R. Koberle, and W. Bialek,
Reliability and Variability in Neural Spike Trains. Science, 1997.275: pp. 1805-1808.
Johnson, D. H. and A. Swami, The Transmission of Signals by Auditory-Nerve Fiber
Discharge Patterns. J. Acoust. Soc. Am., 1983. 74: pp. 493-501.
Meister, M., J. Pine, and D. A. Baylor, Multi-Neuronal Signals from the Retina:
Acquisition and Analysis. 1. Neurosci. Methods, 1994.51: pp. 95-106.
Miller, M . 1. Algorithms for Removing Recovery-Related Distortion gtom Auditory-Nerve
Discharge Patterns. J. Acoust. Soc. Am., 1985.77: pp. 1452-1464.
Rieke, F., D. K. Warland, R. R. de Ruyter van Steveninck, and W. Bialek, Spikes:
Exploring the Neural Code . 1997, Cambridge, MA: MIT Press.
| 1461 |@word trial:15 solid:2 selecting:1 written:1 realistic:1 j1:5 interspike:2 plot:1 drop:2 beginning:2 psth:3 burst:1 constructed:1 become:1 qualitative:1 prove:1 combine:1 inter:4 expected:1 roughly:1 behavior:1 multi:2 actual:1 vertebrate:1 becomes:2 bounded:1 matched:1 panel:1 acoust:2 sharpening:1 nj:1 temporal:7 every:2 ti:1 exactly:1 unit:1 timing:4 firing:52 approximately:1 quantified:2 limited:1 range:3 obeys:1 averaged:2 steveninck:2 acknowledgment:1 procedure:1 elicit:1 significantly:1 confidence:1 onto:1 cannot:2 close:2 gradation:1 restriction:1 imposed:1 rabbit:1 recovery:7 m2:1 rule:1 array:1 population:1 rieke:2 discharge:2 enhanced:1 play:1 agreement:3 harvard:1 std:1 observed:3 role:2 mike:1 fly:1 episode:1 decrease:1 rose:1 asked:1 reviewing:1 segment:5 solving:1 swami:1 upon:1 completely:1 triangle:2 fiber:1 train:17 approached:1 corresponded:1 choosing:1 quite:1 larger:2 distortion:1 otherwise:1 statistic:4 noisy:1 reproduced:1 sequence:1 advantage:1 maximal:1 product:1 neighboring:1 description:1 electrode:1 enhancement:1 transmission:1 produce:2 generating:1 measured:3 nearest:1 strong:1 dividing:1 lna:1 predicted:1 soc:2 p2:1 inhomogeneous:1 closely:1 correct:1 thick:1 bin:6 argued:1 larval:1 exploring:1 mm:1 pine:1 largest:1 onj:2 mit:1 gaussian:1 modified:1 varying:1 conjunction:1 salamander:4 steeply:1 contrast:2 demarcation:1 am:2 accumulated:1 typically:2 lj:1 entire:1 jq:1 among:1 smoothing:1 equal:3 distilled:1 biology:1 identical:1 lit:1 nearly:1 thin:1 stimulus:10 few:1 retina:4 randomly:1 national:2 individual:1 fire:1 light:4 tj:4 shorter:1 iv:1 circle:2 plotted:1 isolated:1 dev:1 contiguous:1 deviation:5 uniform:1 johnson:2 peri:2 peak:1 probabilistic:4 michael:1 continuously:1 squared:1 recorded:1 containing:1 account:1 potential:4 exclude:1 de:3 retinal:5 sec:1 summarized:1 depends:2 performed:2 extracellularly:1 start:1 contribution:1 ni:1 variance:6 miller:2 spaced:1 ofthe:1 raw:1 comp:1 fo:1 reach:1 definition:1 raster:1 acquisition:1 pp:5 recovers:1 auditory:2 intrinsically:1 conversation:1 oftimes:1 nerve:2 higher:1 dt:2 response:14 ooo:1 refractoriness:17 furthermore:1 aj:1 usa:1 effect:2 assigned:1 spatially:1 white:1 during:2 m:20 complete:1 l1:1 instantaneous:2 stimulation:3 spiking:8 refractory:29 exponentially:1 occurred:1 numerically:1 measurement:1 cambridge:2 boxcar:1 similarly:1 fano:3 had:2 reliability:1 add:1 recent:1 inequality:1 continue:1 seen:5 minimum:1 greater:2 recognized:1 period:32 dashed:1 ii:1 smoother:1 signal:2 pnas:1 match:5 characterized:1 calculation:1 determination:1 divided:2 molecular:1 prevented:1 vision:1 poisson:6 histogram:2 cell:17 addition:1 remarkably:1 addressed:1 interval:8 median:1 ot:1 exhibited:1 recording:2 hz:4 regularly:1 integer:1 near:1 counting:2 enough:2 fit:2 identified:1 silent:1 otj:2 cause:1 jj:1 action:4 dramatically:1 generally:1 useful:2 covered:1 simplest:1 generate:3 flicker:5 millisecond:1 notice:2 dotted:2 estimated:2 discrete:2 affected:1 monitor:1 deweese:1 fraction:1 sum:1 ringer:1 jitter:5 uncertainty:1 extends:1 almost:1 comparable:1 layer:1 hi:1 bound:1 ct:3 distinguish:1 identifiable:1 yielded:1 markus:2 wc:1 department:1 jr:1 across:4 describes:1 unity:1 restricted:1 equation:1 count:5 end:1 meister:7 include:1 warland:2 parsed:1 mjb:1 question:1 quantity:2 spike:57 occurs:1 added:1 bialek:2 unable:1 thank:1 cellular:1 code:1 relationship:2 illustration:1 baylor:1 daylight:1 rise:2 diamond:3 neuron:16 finite:1 extended:1 variability:3 excluding:1 sharp:1 intensity:3 extensive:1 beyond:2 bar:1 pattern:2 oj:2 including:1 event:21 difficulty:1 eye:2 acknowledges:2 koberle:1 lewen:1 berry:13 determining:1 relative:8 ina:1 generation:1 versus:1 generator:4 foundation:1 consistent:1 pi:1 summary:1 repeat:2 last:3 free:13 silence:2 institute:1 taking:1 absolute:8 sparse:1 van:2 boundary:2 calculated:5 curve:2 transition:1 made:1 projected:1 far:1 obtains:1 reproduces:1 table:2 ruyter:3 investigated:1 necessarily:1 constructing:1 linearly:1 neurosci:2 noise:1 repeated:7 neuronal:1 fig:10 precision:28 exponential:1 ib:1 saturate:1 removing:1 er:1 exists:1 drew:1 illustrates:1 interneuron:1 ganglion:15 visual:1 expressed:1 contained:1 pip2:1 ma:2 viewed:1 narrower:1 absence:1 tiger:1 change:1 included:1 determined:1 total:1 photoreceptor:1 tby:1 support:2 assessed:1 heaviside:1 ex:1 |
505 | 1,462 | Characterizing Neurons in the Primary
Auditory Cortex of the Awake Primate
U sing Reverse Correlation
R. Christopher deC harms
decharms@phy.ucsf.edu
Michael M . Merzenich
merz@phy.ucsf.edu
w. M. Keck Center for Integrative Neuroscience
University of California, San Francisco CA 94143
Abstract
While the understanding of the functional role of different classes
of neurons in the awake primary visual cortex has been extensively
studied since the time of Hubel and Wiesel (Hubel and Wiesel, 1962),
our understanding of the feature selectivity and functional role of
neurons in the primary auditory cortex is much farther from complete. Moving bars have long been recognized as an optimal stimulus
for many visual cortical neurons, and this finding has recently been
confirmed and extended in detail using reverse correlation methods
(Jones and Palmer, 1987; Reid and Alonso, 1995; Reid et al., 1991;
llingach et al., 1997). In this study, we recorded from neurons in the
primary auditory cortex of the awake primate, and used a novel reverse correlation technique to compute receptive fields (or preferred
stimuli), encompassing both multiple frequency components and ongoing time. These spectrotemporal receptive fields make clear that
neurons in the primary auditory cortex, as in the primary visual cortex, typically show considerable structure in their feature processing
properties, often including multiple excitatory and inhibitory regions
in their receptive fields. These neurons can be sensitive to stimulus
edges in frequency composition or in time, and sensitive to stimulus
transitions such as changes in frequency. These neurons also show
strong responses and selectivity to continuous frequency modulated
stimuli analogous to visual drifting gratings.
1
Introduction
It is known that auditory neurons are tuned for a number of independent feature
parameters of simple stimuli including frequency (Merzenich et al., 1973), intensity
(Sutter and Schreiner, 1995), amplitude modulation (Schreiner and Urbas, 1988), and
Characterizing Auditory Cortical Neurons Using Reverse Correlation
125
others. In addition, auditory cortical responses to multiple stimuli can enhance or suppress one another in a time dependent fashion (Brosch and Schreiner, 1997; Phillips
and Cynader, 1985; Shamma and Symmes, 1985), and auditory cortical neurons can
be highly selective for species-specific vocalizations (Wang et al., 1995; Wollberg and
Newman, 1972), suggesting complex acoustic processing by these cells. It is not yet
known if these many independent selectivities of auditory cortical neurons reflect a
discernible underlying pattern of feature decomposition, as has often been suggested
(Merzenich et al., 1985; Schreiner and Mendelson, 1990; Wang et al., 1995). Further,
since sustained firing rate responses in the auditory cortex to tonal stimuli are typically much lower than visual responses to drifting bars (deCharms and Merzenich,
1996b), it has been suggested that the preferred type of auditory stimulus may still
not be known (Nelken et al., 1994). We sought to develop an unbiased method for
determining the full feature selectivity of auditory cortical neurons, whatever it might
be, in frequency and time based upon reverse correlation.
2
Methods
Recordings were made from a chronic array of up to 49 individually placed ultrafine extracellular Iridium microelectrodes, placed in the primary auditory cortex of
the adult owl monkey. The electrodes had tip lengths of 10-25microns, which yield
impedance values of .5-SMOhm and good isolation of signals from individual neurons
or clusters of nearby neurons. We electrochemically activated these tips to add an
ultramicroscopic coating of Iridium Oxide, which leaves the tip geometry unchanged,
but decreases the tip impedance by more than an order of magnitude, resulting in
substantially improved recording signals. These signals are filtered from .3-8kHz,
sampled at 20kHz, digitized, and sorted. The stimuli used were a variant of random
VlsuII Cortn: Reveree Correlltlon
U.lng 2?D VI.nl Pltternl In Time
Auditory Cortex: Rever.e Correlltlon
U.lng 1?D Auditory Pltternl (Chordl) In Tim.
t -Om.ec
t- 20m.ec
t-40msec
t - 40m.ec
x
x
II
!
Spltlotemporal Receptive Field
II
SplkeT .. ln.
Spectrotempoul Receptive Field
Figure 1: Schematic of stimuli used for reverse correlation.
white noise which was designed to allow us to characterize the responses of neurons
in time and in frequency. As shown in figure 1, these stimuli are directly analogous
to stimuli that have been used previously to characterize the response properties of
neurons in the primary visual cortex (Jones and Palmer, 1987; Reid and Alonso,
1995; Reid et al., 1991). In the visual case, stimuli consist of spatial checkerboards
that span some portion of the two-dimensional visual field and change pattern with
a short sampling interval. In the auditory case, which we have studied here, the
stimuli chosen were randomly selected chords, which approximately evenly span a
R C. deChanns and M M. Merzenich
126
portion of the one-dimensional receptor surface of the cochlea. These stimuli consist
of combinations of pure tones, all with identical phase and all with 5 msec cosineshaped ramps in amplitude when they individually turn on or off. Each chord was
created by randomly selecting frequency values from 84 possible values which span
7 octaves from 110Hz to 14080Hz in even semitone steps. The density of tones in
each stimulus was 1 tone per octave on average, or 7 tones per chord, but the stimuli
were selected stochastically so a given chord could be composed of a variable number
of tones of randomly selected frequencies. We have used sampling rates of 10-100
chords/second, and the data here are from stimuli with 50 chords/second. Stimuli
with random, asynchronous onset times of each tone produce similar results. These
stimuli were presented in the open sound field within an acoustical isolation chamber at 44. 1kHz sampling rate directly from audio compact disk, while the animal sat
passively in the sound field or actively performed an auditory discrimination task,
receiving occasional juice rewards. The complete characterization set lasted for ten
minutes, thereby including 30,000 individual chords.
Spike trains were collected from mUltiple sites in the cortex simultaneously during the
presentation of our characterization stimulus set, and individually reverse correlated
with the times of onset of each of the tonal stimuli. The reverse correlation method
computes the number of spikes from a neuron that were detected, on average, during
a given time preceding, during, or following a particular tonal stimulus component
from our set of chords. These values are presented in spikes/s for all of the tones
in the stimulus set, and for some range of time shifts. This method is somewhat
analogous in intention to a method developed earlier for deriving spectrotemporal
receptive fields for auditory midbrain neurons (Eggermont et al., 1983), but previous
methods have not been effective in the auditory cortex.
3
Results
Figure 2 shows the spectrotemporal responses of neurons from four locations in the
primary auditory cortex. In each panel, the time in milliseconds between the onset of
a particular stimulus component and a neuronal spike is shown along the horizontal
axis. Progressively greater negative time shifts indicate progressively longer latencies
from the onset of a stimulus component until the neuronal spikes. The frequency
of the stimulus component is shown along the vertical axis, in octave spacing from
a 110Hz standard, with twelve steps per octave. The brightness corresponds to the
average rate of the neuron, in spk/s, driven by a particular stimulus component.
The reverse-correlogram is thus presented as a stimulus triggered spike rate average,
analogous to a standard peristimulus time histogram but reversed in time, and is
identical to the spectrogram of the estimated optimal stimulus for the cell (a spike
triggered stimulus average which would be in units of mean stimulus denSity).
A minority of neurons in the primary auditory cortex have spectrotemporal receptive fields that show only a single region of increased rate, which corresponds to the
traditional characteristic frequency of the neuron, and no inhibitory region. We have
found that cells of this type (less than 10%, not shown) are less common than cells
with multimodal receptive field structure. More commonly, neurons have regions of
both increased and decreased firing rate relative to their mean rate within their receptive fields. For terminological convemence, these will be referred to as excitatory
and inhibitory regions, though these changes in rate are not diagnostic of an underlying mechanism. Neurons with receptive fields of this type can serve as detectors
of stimulus edges in both frequency space, and in time. The neuron shown in figure
2a has a receptive field structure indicative of lateral inhibition in frequency space.
This cell prefers a very narrow range of frequencies, and decreases its firing rate for
nearby frequencies, giving the characteristic of a sharply-tuned bandpass filter. This
Characterizing Auditory Cortical Neurons Using Reverse Correlation
a)
...
>
127
b)
3
3.5
2.5
3
40
30
!: 2.5
2
g 1.5
!!!
3
2
8
20
1.5
0.5
-100
10
-50
0
msec
d)
c)
N
3
J:
0
.,>
g
1.5
..
0
15
3.5
2.5
-SO
msec
)!
2
3
10
10;:::
2.5
5
.8
'">
2
0
5 .,
1.5
<
~
U
-5
-10
CD
-100
-50
msec
0
0
~O
-40
-20
msec
Figure 2: Spectrotemporal receptive fields of neurons in the primary auditory cortex
of the awake primate. These receptive fields are computed as described in methods.
Receptive field structures read from left to right correspond to a preferred stimulus for
the neuron, with light shading indicating more probable stimulus components to evoke
a spike, and dark shading indicating less probable components. Receptive fields read
from right to left indicate the response of the neuron in time to a particular stimulus
component. The colorbars correspond to the average firing rates of the neurons in
Hz at a given time preceding, during, or following a particular stimulus component.
type of response is the auditory analog of a visual or tactile edge detector with lateral
inhibition. Simple cells in the primary visual cortex typically show similar patterns
of center excitation along a short linear segment, surrounded by inhibition (Jones
and Palmer, 1987;?Reid and Alonso, 1995; Reid et al., 1991). The neuron shown in
figure 2b shows a decrease in firing rate caused by a stimulus frequency which at a
later time causes an increase in rate. This receptive field structure is ideally suited
to detect stimulus transients; and can be thought of as a detector of temporal edges.
Neurons in the auditory cortex typically prefer this type of stimulus, which is initially
soft or silent and later loud. This corresponds to a neuronal response which shows
an increase followed by a decrease in firing rate. This is again analogous to neuronal
responses in the primary visual cortex, which also typically show a firing rate pattern to an optimal stimulus of excitation followed by inhibition, and preference for
stimulus transients such as when a stimulus is first off and then comes on.
The neuron shown in figures 2c shows an example which has complex receptive field
structure, with multiple regions. Cells of this type would be indicative of selectivity for feature conjunctions or quite complex stimuli, perhaps related to sounds in
the animal's learned environment. Cells with complex receptive field structures are
common in the awake auditory cortex, and we are in the process of quantifying the
percentages of cells that fit within these different categories.
Neurons were observed which respond with increased rate to one frequency range at
one time, and a different frequency range at a later time, indicative of selectivity for
frequency modulations(Suga, 1965). Regions of decreased firing rate can show similar
patterns. The neuron shown in figure 2d is an example of this type. This pattern
is strongly analogous to motion energy detectors in the visual system (Adelson and
Bergen, 1985), which detect stimuli moving in space, and these cells are selective for
changes in frequency.
R. C. deCharms and M M. Merzenich
128
2 oct/sec
6 oct/sec
10 oct/sec 14 oct/sec 30 oct/sec 100 oct/sec
?2 oct/sec ?6 oct/sec ? 10 oct/sec ?14 oct/sec ?30 oct/sec ?100 oct/sec
Figure 3: Parametric stimulus set used to explore neuronal responses to continuously
changing stimulus frequency. Images axe spectrograms of stimuli from left to right in
time, and spanning seven octaves of frequency from bottom to top. Each stimulus is
one second. Numbers indicate the sweep rate of the stimuli in octaves per second.
Based on the responses shown, we wondered whether we could find a more optimal
class of stimuli for these neuron, analogous to the use of drifting bars or gratings in the
primary visual cortex. We have created auditory stimuli which correspond exactly to
the preferred stimulus computed for a paxticulax cell from the cell's spectrotemporal
receptive field (manuscript in prepaxation), and we have also designed a paxametric
class of stimuli which are designed to be particularly effective for neurons selective
for stimuli of changing amplitude or frequency, which are presented here. The stimuli
shown in figure 3 are auditory analogous of visual drifting grating stimuli. The
stimuli axe shown as spectrograms, where time is along the horizontal axis, frequency
content on an octave scale is along the vertical axis, and brightness corresponds to the
intensity of the signal. These stimuli contain frequencies that change in time along an
octave frequency scale so that they repeatedly pass approximately linearly through a
neurons receptive field, just as a drifting grating would pass repeatedly through the
receptive field of a visual neuron. These stimuli axe somewhat analogous to drifting
ripple stimuli which have recently been used by Kowalski, et.al. to characterize the
linearity of responses of neurons in the anesthetized ferret auditory cortex (Kowalski
et al., 1996a; Kowalski et al., 1996b).
Neurons in the auditory cortex typically respond to tonal stimuli with a brisk onset
response at the stimulus transient, but show sustained rates that axe far smaller than
found in the visual or somatosensory systems (deCharms and Merzenich, 1996a).
We have found neurons in the awake animal that respond with high firing rates and
significant selectivity to the class of moving stimuli shown in figure 3. An outstanding
example of this is shown in figure 4. The neuron in this example showed a very high
sustained firing rate to the optimal drifting stimulus, as high as 60 Hz?for one second.
The neuron shown in this example also showed considerable selectivity for stimulus
velocity, as well as some selectivity for stimulus direction.
4
Conclusions
These stimuli enable us to efficiently quantify the response characteristics of neurons in the awake primaxy auditory cortex, as well as producing optimal stimuli for
particular neurons. The data that we have gathered thus far extend our knowledge
about the complex receptive field structure of cells in the primary auditory cortex,
Cha racterizing Auditory Cortical Neurons Using Reverse Correlation
129
2 oct/sec
6 oct/sec
10 octIsec
14 oct/sec
30 oct/sec
100 oct/sec
-2 oct/sec
-6 oct/sec
-10 oct/sec
-14 oct/sec
-30 oct/sec
-100 oct/sec
Figure 4: Responses of a neuron in the primary auditory cortex of the awake primate to example stimuli take form our characterization set, as shown in figure 3. In
each panel, the average response rate histogram in spikes per second is shown below
rastergrams showing the individual action potentials elicited on,each of twenty trials.
and show some considerable analogy with neurons in the primary visual cortex. In
addition, they indicate that it is possible to drive auditory cortical cells to high rates
of sustained firing, as in the visual cortex. This method will allow a number of future
questions to be addressed. Since we have recorded many neurons simultaneously, we
are interested in the interactions among large populations of neurons and how these
relate to stimuli. We are also recording responses to these stimuli while monkeys are
performing cognitive tasks involving attention and learning, and we hope that this
will give us insight into the effects on cell selectivity of the context provided by other
stimuli, the animal's behavioral state or awareness of the stimuli, and the animal's
prior learning of stimulus sets.
5
References
Adelson EH, Bergen JR (1985) Spatiotemporal energy models for the perception of
motion. J. Opt. Soc. Am. A, 2, 284-299.
Brosch M, Schreiner CE (1997) Time course of forward masking tuning curves in cat
primary auditory cortex. J Neurophysiol, 77, 923-43.
deCharms Re, Merzenich MM (1996a) Primary cortical representation of sounds by
the coordination of action-potential timing. Nature, 381, 610-3.
deCharms RC, Merzenich MM (1996b) Primary cortical representation of sounds by
the coordination of action-potential timing. Nature, 381, 610-613.
EggeI1I).ont JJ, Aertsen AM, Johannesma PI (1983) Quantitative characterisation
procedure for auditory neurons based on the spectro-temporal receptive field. Hear
Res, 10, 167-90.
Hubel DH, Wiesel TN (1962) Receptive fields, binocular interaction and functional
archtecture in the cat's visual cortex. J. Physiol., 160, 106-154.
Jones JP, Palmer LA (1987) The two-dimensional spatial structure of simple receptive
130
R. C. deCharms and M M. Merzenich
fields in cat striate cortex. J Neurophysiol, 58, 1187-211.
Kowalski N, Depireux DA, Shamma SA (1996a) Analysis of dynamic spectra in ferret
primary auditory cortex. I. Characteristics of single-unit responses to moving ripple
spectra. J Neurophysiol, 76, 3503-23.
Kowalski N, Depireux DA, Shamma SA (1996b) Analysis of dynamic spectra in ferret primary auditory cortex. II. Prediction of unit responses to arbitrary dynamic
spectra. J Neurophysiol, 76, 3524-34.
Merzenich MM, Jenkins WM, Middlebrooks JC (1985) Observations and hypotheses
on special organizational features of the central auditory nervous system. In: Dynamic Aspects of Neocortical Function. Edited by E. G. a. W. M. C. G. Edelman.
New York: Wiley, pp. 397-423.
Merzenich MM, Knight PL, Roth GL (1973) Cochleotopic organization of primary
auditory cortex in the cat. Brain Res, 63, 343-6.
Nelken I, Prut Y, Vaadia E, Abeles M (1994) In search of the best stimulus: an
optimization procedure for finding efficient stimuli in the? cat auditory cortex. Hear
Res, 72, 237-53.
Phillips DP, Cynader MS (1985) Some neural mechanisms in the cat's auditory cortex
underlying sensitivity to combined tone and wide-spectrum noise stimuli. Hear Res,
18, 87-102.
Reid RC, Alonso JM (1995) Specificity of monosynaptic connections from thalamus
to visual cortex. Nature, 378,281-4.
Reid RC, Soodak RE, Shapley RM (1991) Directional selectivity and spatiotemporal
structure of receptive fields of simple cells in cat striate cortex. J Neurophysiol, 66,
505-29.
Ringach DL, Hawken MJ, Shapley R (1997) Dynamics of orientation tuning in
macaque primary visual cortex. Nature, 387, 281-4.
Schreiner CE, Mendelson JR (1990) Functional topography of cat primary auditory
cortex: distribution of integrated excitation. J Neurophysiol, 64, 1442-59.
Schreiner CE, Urbas JV (1988) Representation of amplitude in the auditory cortex
of the cat. II. Comparison between cortical fields. Hear. Res., 32, 49-64.
Shamma SA, Symmes D (1985) Patterns of inhibition in auditory cortical cells in
awake squirrel monkeys. Hear Res, 19, 1-13.
Suga N (1965) Responses of cortical auditory neurones to frequency modulated sounds
in echo-locating bats. Nature, 206, 890-l.
Sutter ML, Schreiner CE (1995) Topography of intensity tuning in cat primary auditory cortex: single-neuron versus multiple-neuron recordings. J Neurophysiol, 73,
190-204.
Wang X, Merzenich MM, Beitel R, Schreiner CE (1995) Representation of a speciesspecific vocalization in the primary auditory cortex of the common marmoset: temporal and spectral characteristics. J Neurophysiol, 74, 2685-706.
Wollberg Z, Newman JD (1972) Auditory cortex of squirrel monkey: response patterns of single cells to species-specific vocalizations. Science, 175, 212-214.
| 1462 |@word trial:1 wiesel:3 disk:1 open:1 cha:1 integrative:1 decomposition:1 brightness:2 thereby:1 shading:2 phy:2 selecting:1 tuned:2 yet:1 physiol:1 discernible:1 designed:3 progressively:2 discrimination:1 leaf:1 selected:3 nervous:1 tone:8 indicative:3 sutter:2 short:2 farther:1 filtered:1 characterization:3 location:1 preference:1 rc:3 along:6 edelman:1 sustained:4 shapley:2 behavioral:1 brain:1 ont:1 jm:1 provided:1 monosynaptic:1 underlying:3 linearity:1 panel:2 substantially:1 monkey:4 developed:1 finding:2 temporal:3 quantitative:1 exactly:1 rm:1 middlebrooks:1 whatever:1 unit:3 producing:1 reid:8 timing:2 receptor:1 firing:11 modulation:2 approximately:2 might:1 studied:2 shamma:4 palmer:4 range:4 suga:2 bat:1 procedure:2 johannesma:1 thought:1 intention:1 specificity:1 context:1 center:2 chronic:1 roth:1 attention:1 wondered:1 pure:1 schreiner:9 insight:1 array:1 deriving:1 population:1 analogous:9 hypothesis:1 velocity:1 particularly:1 observed:1 role:2 bottom:1 wang:3 region:7 decrease:4 chord:8 knight:1 edited:1 environment:1 reward:1 ideally:1 dynamic:5 segment:1 serve:1 cynader:2 upon:1 neurophysiol:8 spk:1 multimodal:1 cat:10 train:1 effective:2 detected:1 newman:2 quite:1 ramp:1 echo:1 vocalization:3 triggered:2 vaadia:1 interaction:2 electrode:1 keck:1 cluster:1 ripple:2 produce:1 tim:1 develop:1 sa:3 strong:1 soc:1 grating:4 indicate:4 come:1 somatosensory:1 quantify:1 direction:1 coating:1 filter:1 transient:3 enable:1 owl:1 opt:1 probable:2 pl:1 mm:5 squirrel:2 sought:1 spectrotemporal:6 coordination:2 sensitive:2 individually:3 hope:1 depireux:2 conjunction:1 lasted:1 detect:2 am:2 dependent:1 bergen:2 typically:6 integrated:1 initially:1 selective:3 interested:1 among:1 orientation:1 animal:5 spatial:2 special:1 field:30 sampling:3 identical:2 jones:4 adelson:2 future:1 others:1 stimulus:79 randomly:3 microelectrodes:1 composed:1 simultaneously:2 individual:3 geometry:1 phase:1 organization:1 highly:1 nl:1 light:1 activated:1 edge:4 re:8 increased:3 earlier:1 soft:1 organizational:1 characterize:3 spatiotemporal:2 abele:1 combined:1 density:2 twelve:1 sensitivity:1 off:2 receiving:1 michael:1 enhance:1 tip:4 continuously:1 again:1 reflect:1 recorded:2 central:1 stochastically:1 oxide:1 cognitive:1 checkerboard:1 actively:1 suggesting:1 potential:3 sec:23 jc:1 caused:1 vi:1 onset:5 performed:1 later:3 portion:2 wm:1 elicited:1 masking:1 om:1 prut:1 characteristic:5 efficiently:1 yield:1 correspond:3 kowalski:5 gathered:1 directional:1 confirmed:1 drive:1 detector:4 energy:2 frequency:27 pp:1 sampled:1 auditory:50 knowledge:1 amplitude:4 manuscript:1 response:23 improved:1 though:1 strongly:1 just:1 binocular:1 correlation:9 until:1 horizontal:2 christopher:1 perhaps:1 effect:1 contain:1 unbiased:1 merzenich:13 read:2 ringach:1 white:1 semitone:1 during:4 excitation:3 m:1 octave:8 complete:2 neocortical:1 tn:1 motion:2 image:1 novel:1 recently:2 common:3 juice:1 functional:4 khz:3 jp:1 axe:4 analog:1 extend:1 significant:1 composition:1 phillips:2 tuning:3 had:1 moving:4 cortex:43 surface:1 longer:1 inhibition:5 add:1 showed:2 driven:1 reverse:11 selectivity:11 greater:1 somewhat:2 preceding:2 spectrogram:3 recognized:1 signal:4 ii:4 multiple:6 full:1 sound:6 thalamus:1 long:1 schematic:1 prediction:1 variant:1 involving:1 histogram:2 cochlea:1 dec:1 cell:18 addition:2 spacing:1 addressed:1 interval:1 decreased:2 ferret:3 recording:4 hz:5 isolation:2 fit:1 silent:1 shift:2 whether:1 tonal:4 tactile:1 locating:1 york:1 cause:1 jj:1 prefers:1 repeatedly:2 action:3 neurones:1 latency:1 clear:1 dark:1 extensively:1 ten:1 category:1 percentage:1 inhibitory:3 millisecond:1 neuroscience:1 estimated:1 per:5 diagnostic:1 urbas:2 four:1 characterisation:1 changing:2 jv:1 ce:5 micron:1 respond:3 lng:2 hawken:1 prefer:1 followed:2 sharply:1 awake:9 nearby:2 aspect:1 span:3 passively:1 performing:1 extracellular:1 combination:1 jr:2 smaller:1 primate:4 marmoset:1 midbrain:1 ln:1 brosch:2 previously:1 turn:1 mechanism:2 jenkins:1 occasional:1 spectral:1 chamber:1 drifting:7 jd:1 top:1 giving:1 eggermont:1 unchanged:1 sweep:1 question:1 spike:9 receptive:26 primary:27 parametric:1 striate:2 traditional:1 aertsen:1 dp:1 reversed:1 lateral:2 alonso:4 evenly:1 acoustical:1 seven:1 collected:1 spanning:1 minority:1 length:1 relate:1 decharms:7 negative:1 suppress:1 twenty:1 vertical:2 neuron:55 observation:1 sing:1 extended:1 digitized:1 arbitrary:1 intensity:3 connection:1 california:1 acoustic:1 learned:1 narrow:1 macaque:1 adult:1 bar:3 suggested:2 below:1 pattern:8 perception:1 hear:5 including:3 eh:1 axis:4 created:2 prior:1 understanding:2 determining:1 relative:1 encompassing:1 topography:2 analogy:1 versus:1 awareness:1 surrounded:1 cd:1 pi:1 excitatory:2 course:1 placed:2 gl:1 asynchronous:1 allow:2 wide:1 characterizing:3 anesthetized:1 curve:1 cortical:14 transition:1 computes:1 nelken:2 forward:1 made:1 commonly:1 san:1 ec:3 far:2 compact:1 spectro:1 preferred:4 evoke:1 ml:1 hubel:3 sat:1 harm:1 francisco:1 spectrum:5 continuous:1 search:1 impedance:2 nature:5 mj:1 ca:1 brisk:1 complex:5 da:2 linearly:1 noise:2 neuronal:5 site:1 referred:1 fashion:1 wiley:1 msec:6 bandpass:1 minute:1 specific:2 showing:1 peristimulus:1 iridium:2 dl:1 mendelson:2 consist:2 magnitude:1 suited:1 explore:1 visual:21 correlogram:1 corresponds:4 loud:1 dh:1 oct:23 sorted:1 presentation:1 quantifying:1 considerable:3 change:5 content:1 specie:2 pas:2 merz:1 la:1 indicating:2 soodak:1 modulated:2 ucsf:2 outstanding:1 ongoing:1 audio:1 correlated:1 |
506 | 1,463 | Statistical Models of Conditioning
Peter Dayan*
Brain & Cognitive Sciences
E25-2IDMIT
Cambridge, MA 02139
Theresa Long
123 Hunting Cove
Williamsburg, VA 23185
Abstract
Conditioning experiments probe the ways that animals make predictions about rewards and punishments and use those predictions to control their behavior. One standard model of conditioning paradigms which involve many conditioned stimuli suggests
that individual predictions should be added together. Various key
results show that this model fails in some circumstances, and motivate an alternative model, in which there is attentional selection
between different available stimuli. The new model is a form of
mixture of experts, has a close relationship with some other existing psychological suggestions, and is statistically well-founded.
1
Introduction
Classical and instrumental conditioning experiments study the way that animals
learn about the causal texture of the world (Dickinson, 1980) and use this information to their advantage. Although it reached a high level of behavioral sophistication, conditioning has long since gone out of fashion as a paradigm for studying
learning in animals, partly because of the philosophical stance of many practitioners, that the neurobiological implementation of learning is essentially irrelevant.
However, more recently it has become possible to study how conditioning phenomena are affected by particular lesions or pharmacological treatments to the
brain (eg Gallagher & Holland, 1994), and how particular systems, during simple
learning tasks, report information that is consistent with models of conditioning
(Gluck & Thompson, 1987; Gabriel & Moore, 1989).
In particular, we have studied the involvement of the dopamine (DA) system in
the ventral tegmental area of vertebrates in reward based learning (Montague et
aI, 1996; Schultz et aI, 1997). The activity of these cells is consistent with a model
in which they report a temporal difference (TO) based prediction error for reward
"This work was funded by the Surdna Foundation.
118
P. Dayan and T. Long
(Sutton & Barto, 1981; 1989). This prediction error signal can be used to learn correct
predictions and also to learn appropriate actions (Barto, Sutton & Anderson, 1983).
The DA system is important since it is crucially involved in normal reward learning,
and also in the effects of drugs of addiction, self stimulation, and various neural
diseases.
The TO model is consistent with a whole body of experiments, and has even correctly anticipated new experimental findings. However, like the Rescorla-Wagner
(RW; 1972) or delta rule, it embodies a particular additive model for the net prediction made when there are multiple stimuli. Various sophisticated conditioning
experiments have challenged this model and found it wanting. The results support
competitive rather than additive models. Although ad hoc suggestions have been
made to repair the model, none has a sound basis in appropriate prediction. There
is a well established statistical theory for competitive models, and it is this that we
adopt.
In this paper we review existing evidence and theories, show what constraints
a new theory must satisfy, and suggest and demonstrate a credible candidate.
Although it is based on behavioral data, it also has direct implications for our
neural theory.
2 Data and Existing Models
Table 1 describes some of the key paradigms in conditioning (Dickinson, 1980;
Mackintosh, 1983). Although the collection of experiments may seem rather arcane
(the standard notation is even more so), in fact it shows exactly the basis behind
the key capacity of animals in the world to predict events of consequence. We will
extract further biological constraints implied by these and other experiments in the
discussion.
In the table,l (light) and s (tone) are potential predictors (called conditioned stimuli or
CSs), of a consequence, r, such a~ the delivery of a reward (called an unconditioned
stimulus or US). Even though we use TO rules in practice, we discuss some of the
abstract learning rules without much reference to the detailed time course of trials.
The same considerations apply to TD.
In Pavlovian conditioning, the light acquires a positive association with the reward
in a way that can be reasonably well modeled by:
,6.Wl(t) = al(t)(r(t) - wl(t?l(t),
(1)
where let) E {O, I} represents the presence of the light in trial t (s(t) will similarly
represent the presence of a tone), Wl(t) (we will often drop the index t) represents
the strength of the expectation about the delivery of reward ret) in trial t if the
light is also delivered, and al(t) is the learning rate. This is just the delta rule. It
also captures well the probabilistic contingent nature of conditioning - for binary
ret) E {O, I}, animals seem to assess il = P[r(t)ll(t) = 1J - P[r(t)ll(t) = OJ, and then
only expect reward following the light (in the model, have WI > 0) if il > O.
Pavlovian conditioning is easy to explain under a whole wealth of rules. The trouble
comes in extending equation 1 to the case of multiple predictors (in this paper we
consider just two). The other paradigms in table 1 probe different aspects of this.
The one that is most puzzling is (perversely) called downwards unblocking (Holland,
1988). In a first set of trials, an association is established between the light and two
presentations of reward separated by a few (u) seconds. In a second set, a tone is
included with the light, but the second reward is dropped. The animal amasses
less reward in conjunction with the tone. However, when presented with the tone
Statistical Models of Conditioning
119
1
Name
Pavlovian
2
Overshadowing
l+s-tr
3
Inhibitory
l-tr }
{ Z+s-t?
4
Blocking
Upwards unblocking
Downwards unblockinK
5
.6
Set 1
I
l-tr
l-tr
I -t rflur
Set 2
-t r
l+s-tr
1+ s -t rflur
l+s-tr
Test
r
l~
{I
~ r~
s~
1
r'i
}
s~f
s~?
s~r
s~ ?~
Table 1: Paradigms. Sets 1 and 2 are separate sets of learning trials, which are continued
until convergence. Symbols land s indicate presentation of lights and tones as potential
predictors. The 't+ in the test set indicates that the associations of the predictors are tested,
prodUCing the listed results. In overshadowing, association with the reward can be divided
between the light and the sound, indicated by r!. In some cases overshadowing favours
one stimulus at the complete expense of the other; and at the end of very prolonged training,
all effects of overshadowing can disappear. In blocking, the tone makes no prediction of r.
In set 2 of inhibitory conditioning, the two types of trials are interleaved and the outcome
is that the tone predicts the absence of reward. In upwards and downwards unblocking,
the 6" indicates that the delivery of two rewards is separated by time u. For downwards
unblocking, if u is small, then s is associated with the absence of r; if u is large, then s is
associated with the presence of r.
alone, the animal expects the presence rather than the absence of reward. On the face
of it, this seems an insurmountable challenge to prediction-based theories. First
we describe the existing theories, then we formalise some potential replacements.
One theory (called a US-processing theory) is due to Rescorla & Wagner (RW; 1972),
and, as pointed out by Sutton & Barto (1981), is just the delta rule. For RW, the
animal constructs a net prediction:
V(t) = wi(t)l(t)
+ ws{t)s{t)
(2)
for r(t), and then changes flWi(t) = Cti(t)(r(t) - V(t?l(t) (and similarly for ws(t?
using the prediction error r(t) - V(t). Its foundation in the delta rule makes it
computationally appropriate (Marr, 1982) as a method of making predictions. TD
uses the same additive model in equation 2, but uses r(t) + V(t + 1) - V(t) as the
prediction error.
RW explains overshadowing, inhibitory conditioning, blocking, and upwards unblocking, but not downwards unblocking. In overshadowing, the terminal association between I and r is weaker if I and s are simultaneously trained - this is
expected under RW since learning stops when V(t) = r{t), and W, and Ws will
share the prediction. In inhibitory conditioning, the sound comes to predict the absence of r. The explanation of inhibitory conditioning is actually quite complicated
(Konorski, 1967; Mackintosh, 1983); however RW provides the simple account that
WI = r for the I -t r trials, forcing Ws = -r for the 1+ s -t . trials. In blocking,
the prior association between I and r means that Wi = r in the second set of trials,
leading to no learning for the tone (since V(t) - r(t) = 0). In upwards unblocking,
Wi = r at the start of set 2. Therefore, r(t) - WI = r > 0, allowing Ws to share in the
prediction.
As described above, downwards unblocking is the key thorn in the side of RW.
Since the TD rule combines the predictions from different stimuli in a similar way,
P. Dayan and T. Long
120
it also fails to account properly for downwards unblocking. This is one reason why
it is incorrect as a model of reward learning.
The class of theories (called CS-processing theories) that is alternative to RW does
not construct a net prediction V(t), but instead uses equation 1 for all the stimuli,
only changing the learning rates O!l(t) and O!s(t) as a function of the conditioning
history of the stimuli (eg Mackintosh, 1975; Pearce & Hall, 1980; Grossberg, 1982).
A standard notion is that there is a competition between different stimuli for a
limited capacity learning processor (Broadbent, 1958; Mackintosh, 1975; Pearce &
Hall, 1980), translating into competition between the learning rates. In blocking,
nothing unexpected happens in the second set of trials and equally, the tone does
not predict anything novel. In either case as is set to '" 0 and so no learning
happens. In these models, downwards unblocking now makes qualitative sense:
the surprising consequences in set 2 can be enough to set as ?0, but then learning
according to equation 1 can make Ws > O. Whereas Mackintosh's (1975) and Pearce
and Hall's (1980) models only consider competition between the stimuli for learning,
Grossberg's (1982) model incorporates competition during representation, so the net
prediction on a trial is affected by competitive interactions between the stimuli. In
essence, our model provides a statistical formalisation of this insight.
3
New Models
From the previous section, it would seem that we have to abandon the computational basis of the RW and TD models in terms of making collective predictions
about the reward. The CS-processing models do not construct a net prediction of
the reward, or say anything about how possibly conflicting information based on
different stimuli should be integrated. This is a key flaw - doing anything other
than well-founded prediction is likely to be maladaptive. Even quite successful
pre-synaptic models, such as Grossberg (1982), do not justify their predictions.
We now show that we can take a different, but still statistically-minded approach to combination in which we specify a parameterised probability distribution P[r(t)ls(t), l(t)] and perform a form of maximum likelihood (ML) inference,
updating the parameters to maximise this probability over the samples. Consider
three natural models of P[r(t)/s(t), l(t)]:
Pa[r(t)ls(t),l(t)]
P M[r(t)/s(t), l(t)]
P J[r(t)/s(t), l(t)]
N[w1l(t)
+ wss(t), (72]
7l"1 (t)N[Wl' (72]
N[WI7l"I(t)l(t)
+ 7l"s(t)N[ws, (72] + 1i'(t).,v[w, r2]
+ wsnAt)s(t), (72]
(3)
(4)
(5)
where N[J.L, (72] is a normal distribution, with mean J.L and variance (72. In the latter
two cases, 0 ::; 7l"1 (t) + 7l" s (t) ::; I, implementing a form of competition between
the stimuli, and 7l".(t) = 0 if stimulus * is not presented. In equation 4, N[w, r2]
captures the background expectation if neither the light nor the tone wins, and
1i'(t) = 1 - 7l"1(t) - n"s{t). We will show that the data argue against the first two and
support the third of these models.
Rescorla-Wagner: Pa[r(t)/s(t), l(t)]
The RW rule is derived as ML inference based on equation 3. The only difference
is the presence of the variance, (72. This is useful for capturing the partial reinforcement effect (see Mackintosh, 1983), in which if r(t) is corrupted by substantial
noise (ie (72 ?0), then learning to r is demonstrably slower. As we discussed above,
121
Statistical Models of Conditioning
downwards unblocking suggests that animals are not using P G [r( t) Is(t), I (t)] as the
basis for their predictions.
Competitive mixture of experts: P M[r(t)ls(t), l(t)]
PM[r(t)ls(t),l(t)] is recognisable as the generative distribution in a mixture
of Gaussians model (Nowlan, 1991; Jacobs et ai, 1991b). Key in this model
are the mixing proportions 7r1(t) and 7r s (t). Online variants of the E phase of
the EM algorithm (Dempster et ai, 1977) compute posterior responsibilities as
ql(t) + qs(t) + q(t) = 1, where ql(t) <X 7r1(t)e-(r(t)-w1I(t)),2/2(T2 (and similarly for the
others), and then perform a partial M phase as
L\wl(t) <X (r(t) - WI (t?ql(t)
L\ws(t) <X (r(t) - ws(t?qs(t)
(6)
which has just the same character as the presynaptic rules (depending on how 7r1 (t)
is calculated). As in the mixture of experts model, each expert (each stimulus here)
that seeks to predict r(t) (ie each stimulus * for which q. (t) f; 0) has to predict
the whole of r(t) by itself. This means that the model can capture downwards
unblocking in the following way. The absence of the second r in the second set of
trials forces 7r s (t) > 0, and, through equation 6, this in turn means that the tone will
come to predict the presence of the first r. The time u between the rewards can be
important because of temporal discounting. This means that there are sufficiently
large values of u for which the inhibitory effect of the absence of the second reward
will be dominated. Note also that the expected reward based on l(t) and s(t) is the
sum
(7)
Although the net prediction given in equation 7 is indeed based on all the stimuli,
it does not directly affect the course of learning. This means that the model has
difficulty with inhibitory conditioning. The trouble with inhibitory conditioning is
that the model cannot use Ws < 0 to counterbalance WI > 0 - it can at best set Ws =0,
which is experimentally inaccurate. Note, however, this form of competition bears
some interesting similarities with comparator models of conditioning (see Miller &
Matzel, 1989). It also has some problems in explaining overshadowing, for similar
reasons.
Cooperative mixture of experts: P J[r(t)ls(t), l(t)]
The final model P J[r(t)ls(t), l(t)] is just like the mixture model that Jacobs et al
(1991a) suggested (see also Bordley, 1982). One statistical formulation of this model
considers that, independently,
where Pl(t) and Ps(t) are inverse variances. This makes
(72
= (Pl(t)
+ Ps(t?-l 7r1(t)
= PI(t)(72
7r s (t) = Ps(t)(72.
Normative learning rules should emerge from a statistical model of uncertainty in
the world. Short of such a model, we used:
7r1 (t)
L\WI = o:w-(-) 6(t)
PI t
where 6(t) = r(t) - 7r1(t)Wl (t) - 7rs (t)ws (t) is the prediction error; the 1/ Pl(t) term
in changing WI makes learning slower if WI is more certainly related to r (ie if PI (t) is
greater); the 0.1 substitutes for background noise; if 62 (t) is too large, then PI + Ps
P. Dayan and T. Long
122
t'reolctrVe vanances; I:SloCklng
I:SIOCKlng ana unDloCklng
......... ...
r-u"gj;\'--'--'
8
,
t'reolctove vanances: UnDlocKlng
i
6
;:_1:1
/~~
.... -"-
4
r . ""
~~L-~1~
0 --~2~
0 --~M
~~
~
Tim. to 2nd I8warn
00
200
400
600
Trial
800
1000
00
200
400
600
800
1000
Trial
Figure 1: Blocking and downwards unblocking with 5 steps to the first reward; and a
variable number to the second. Here, the discount factor "y = 0.9, and O:w = 0.5, O:p = 0.02,
f.L = 0.75. For blocking, the second reward remains; for unblocking it is removed after 500
trials. a) The terminal weight for the sound after learning - for blocking it is always small
and positive; for downwards unblocking, it changes from negative at small ~u to positive
at large ~ u. b,c) Predictive variances Pl(t) and P.. (t). In blocking, although there is a small
change when the sound is introduced because of additivity of the variances, learning to the
sound is substantially prevented. In downwards unblocking, the surprise omission of the
second reward makes the sound associable and unblocks learning to it.
pr
is shared out in proportion of
to capture the insight that there can be dramatic
changes to variabilities; and the variabilities are bottom-limited.
Figure 1 shows the end point and course of learning in blocking and downwards
unblocking. Figure 1a confirms that the model captures downwards unblocking,
making the terminal value of Ws negative for short separations between the rewards and positive for long separations. By comparison, in the blocking condition,
for which both rewards are always presented, W s is always small and positive.
Figures 1b,c show the basis behind this behaviour in terms of Pl(t) and Ps{t). In
particular, the heightened associability of the sound in unblocking following the
prediction error when the second reward is removed accounts for the behavior.
As for the mixture of experts model (and also for comparator models), the presence
of 11'j(t) and nAt) makes the explanation of inhibitory conditioning and overshadowing a little complicated. For instance, if the sound is associable (Ps(t) ? 0), then
it can seem to act as a conditioned inhibitor even if Ws = O. Nevertheless, unlike
the mixture of experts model, the fact that learning is based on the joint prediction
makes true inhibitory conditioning possible.
4
Discussion
Downwards unblocking may seem like an extremely abstruse paradigm with which
to refute an otherwise successful and computationally sound model. However, it
is just the tip of a conditioning iceberg that would otherwise sink TD. Even in
other reinforcement learning applications of TO, there is no a priori reason why
predictions should be made according to equation 2 - the other statistical models
in equations 4 and 5 could also be used. Indeed, it is easy to generate circumstances
in which these more competitive models will perform better. For the neurobiology,
experiments on the behavior of the DA system in these conditioning tasks will help
specify the models further.
The model is incomplete in various important ways. First, it makes no distinction
between preparatory and consumatory conditioning (Konorski, 1967). There is
evidence that the predictions a CS makes about the affective value of USs fall in
a different class from the predictions it makes about the actual USs that appear.
Statistical Models of Conditioning
123
For instance, an inhibitory stimulus reporting the absence of expected delivery of
food can block learning to the delivery of shock, implying that aversive events
form a single class. The affective value forms the preparatory aspect, is likely
what is reported by the DA cells, and perhaps controls orienting behavior, the
characteristic reaction of animals to the conditioned stimuli that may provide an
experimental handle on the attention they are paid. Second, the model does not
use opponency (Konorski, 1967; Solomon & Corbit, 1974; Grossberg, 1982) to
handle inhibitory conditioning. This is particularly important, since the dynamics
of the interaction between the opponent systems may well be responsible for the
importance of the delay u in downwards unblocking. Serotonin is an obvious
candidate as an opponent system to DA (Montague et a11996). We also have not
specified a substrate for the associabilities or the attentional competition - the DA
system itself may well be involved. Finally, we have not specified an overall model
of how the animal might expect the contingency of the world to change over time
- which is key to the statistical justification of appropriate learning rules.
References
[1] Barto, AG, Sutton, RS & Anderson, CW (1983). IEEE Transactions on Systems, Man, and
Cybernetics, 13, pp 834-846.
[2] Bordley, RF (1982). Journal of the Operational Research Society, 33, 171-174.
[3] Broadbent, DE (1958). Perception and Communication. London: Pergamon.
[4] Buhusi, CV & Schmajuk, NA. Hippocampus, 6, 621-642.
[5] Dempster, AP, Laird, NM & Rubin, DB (1977). Proceedings of the Royal Statistical Society,
B-39,1-38.
[6] Dickinson, A (1980). Contemporary Animal Learning Theory. Cambridge, England: Cambridge University Press.
[7] Gabriel, M & Moore, J, editors (1989). Learning and Computational Neuroscience. Cambridge, MA: MIT Press.
[8] Gallagher, M & Holland, PC (1994). PNAS, 91, 11771-6.
[9] Gluck, MA & Thompson, RF (1987). Psychological Reviews, 94, 176-191.
[10] Grossberg, S (1982). Psychological Review, 89,529-572.
[11] Holland, PC (1988). Journal of Experimental Psychology: Animal Behavior Processes, 14,
261-279.
[12] Jacobs, RA, Jordan, MI & Barto, AG (1991). Cognitive Science, 15, 219-250.
[13] Jacobs, RA, Jordan, MI, Nowlan, SJ & Hinton, GE (1991). Neural Computation, 3, 79-87.
[14] Konorski, J (1967). Integrative Activity of the Brain. Chicago, 11: Chicago University Press.
[15] Mackintosh, NJ (1975). Psychological Review, 82,276-298.
[16] Mackintosh, NJ (1983). Conditioning and Associative Learning. Oxford, UK: Oxford University Press.
[17] Marr, 0 (1982). Vision. New York, NY: Freeman.
[18] Miller, RR & Matzel, LD (1989). In SB Klein & RR Mowrer, editors, Contemporary
Learning Theories: Pavlovian Conditioning and the Status of Traditional Theory. Hillsdale,
NJ: Lawrence Erlbaum.
[19] Montague, PR, Dayan, P & Sejnowski, TK (1996). Journal of Neuroscience, 16, 1936-1947.
[20] Nowlan, SJ (1991). Soft Competitive Adaptation: Neural Network Learning Algorithms Based
on Fitting Statistical Mixtures. PhD Thesis, Department of Computer Science, CamegieMellon University.
[21] Pearce, JM & Hall, G (1980). Psychological Review, 87,532-552.
[22] Rescorla, RA & Wagner, AR (1972). In AH Black & WF Prokasy, editors, Classical Conditioning II: Current Research and Theory, pp 64-69. New York, NY: Appleton-CenturyCrofts.
[23] Schultz, W, Dayan, P & Montague, PR (1997). Science, 275, 1593-1599.
[24] Solomon, RL & Corbit, JD (1974). Psychological Review, 81, 119-145.
[25] Sutton, RS & Barto, AG (1981). Psychological Review, 882, pp 135-170.
[26] Sutton, RS & Barto, AG (1989). In Gabriel & Moore (1989).
| 1463 |@word trial:15 proportion:2 instrumental:1 nd:1 seems:1 hippocampus:1 integrative:1 confirms:1 seek:1 crucially:1 r:4 jacob:4 paid:1 dramatic:1 tr:6 ld:1 hunting:1 existing:4 reaction:1 current:1 surprising:1 nowlan:3 must:1 additive:3 chicago:2 drop:1 alone:1 generative:1 implying:1 tone:12 short:2 provides:2 direct:1 become:1 addiction:1 incorrect:1 qualitative:1 combine:1 fitting:1 behavioral:2 affective:2 indeed:2 ra:3 preparatory:2 expected:3 nor:1 behavior:5 brain:3 terminal:3 freeman:1 td:5 prolonged:1 little:1 actual:1 food:1 jm:1 vertebrate:1 notation:1 what:2 substantially:1 ret:2 finding:1 ag:4 nj:3 temporal:2 act:1 exactly:1 uk:1 control:2 appear:1 producing:1 positive:5 maximise:1 dropped:1 consequence:3 sutton:6 oxford:2 ap:1 might:1 black:1 studied:1 suggests:2 limited:2 gone:1 statistically:2 grossberg:5 responsible:1 practice:1 block:1 area:1 drug:1 pre:1 suggest:1 cannot:1 close:1 selection:1 attention:1 l:6 thompson:2 independently:1 rule:12 continued:1 insight:2 q:2 marr:2 handle:2 notion:1 justification:1 cs:1 heightened:1 dickinson:3 substrate:1 us:5 pa:2 mowrer:1 particularly:1 updating:1 predicts:1 maladaptive:1 blocking:11 cooperative:1 bottom:1 capture:5 removed:2 contemporary:2 disease:1 substantial:1 dempster:2 reward:27 dynamic:1 motivate:1 trained:1 predictive:1 basis:5 sink:1 joint:1 montague:4 various:4 additivity:1 separated:2 describe:1 london:1 sejnowski:1 outcome:1 quite:2 say:1 otherwise:2 serotonin:1 itself:2 laird:1 abandon:1 unconditioned:1 delivered:1 online:1 hoc:1 advantage:1 final:1 associative:1 net:6 rr:2 rescorla:4 interaction:2 adaptation:1 mixing:1 competition:7 convergence:1 p:6 extending:1 r1:6 tk:1 tim:1 depending:1 help:1 insurmountable:1 c:3 come:3 indicate:1 correct:1 translating:1 ana:1 implementing:1 hillsdale:1 explains:1 behaviour:1 biological:1 refute:1 pl:5 sufficiently:1 hall:4 normal:2 lawrence:1 predict:6 ventral:1 adopt:1 wl:6 minded:1 mit:1 tegmental:1 always:3 inhibitor:1 rather:3 barto:7 conjunction:1 derived:1 properly:1 indicates:2 likelihood:1 sense:1 wf:1 inference:2 flaw:1 dayan:6 sb:1 integrated:1 inaccurate:1 w:15 overall:1 priori:1 animal:13 construct:3 represents:2 anticipated:1 report:2 stimulus:20 t2:1 others:1 few:1 simultaneously:1 individual:1 phase:2 replacement:1 certainly:1 mixture:9 light:10 behind:2 pc:2 implication:1 partial:2 incomplete:1 opponency:1 causal:1 formalise:1 psychological:7 instance:2 soft:1 ar:1 overshadowing:8 w1i:1 challenged:1 expects:1 predictor:4 delay:1 successful:2 erlbaum:1 too:1 reported:1 corrupted:1 punishment:1 ie:3 probabilistic:1 tip:1 together:1 e25:1 na:1 thesis:1 nm:1 solomon:2 possibly:1 cognitive:2 expert:7 leading:1 account:3 potential:3 de:1 satisfy:1 ad:1 responsibility:1 doing:1 reached:1 competitive:6 start:1 complicated:2 ass:1 il:2 matzel:2 variance:5 characteristic:1 miller:2 none:1 cybernetics:1 processor:1 history:1 ah:1 unblocking:21 explain:1 synaptic:1 against:1 pp:3 involved:2 obvious:1 associated:2 mi:2 stop:1 treatment:1 credible:1 sophisticated:1 actually:1 specify:2 formulation:1 though:1 anderson:2 just:6 parameterised:1 until:1 broadbent:2 perhaps:1 indicated:1 orienting:1 name:1 effect:4 true:1 discounting:1 stance:1 moore:3 eg:2 pharmacological:1 ll:2 self:1 during:2 acquires:1 essence:1 anything:3 complete:1 demonstrate:1 upwards:4 consideration:1 novel:1 recently:1 iceberg:1 stimulation:1 rl:1 conditioning:32 association:6 discussed:1 cambridge:4 appleton:1 ai:4 cv:1 pm:1 similarly:3 pointed:1 funded:1 similarity:1 gj:1 cove:1 posterior:1 involvement:1 irrelevant:1 forcing:1 binary:1 contingent:1 greater:1 paradigm:6 signal:1 ii:1 multiple:2 sound:10 pnas:1 england:1 long:6 divided:1 equally:1 prevented:1 va:1 prediction:31 variant:1 circumstance:2 expectation:2 essentially:1 dopamine:1 vision:1 represent:1 cell:2 whereas:1 background:2 wealth:1 unlike:1 db:1 incorporates:1 seem:5 jordan:2 practitioner:1 presence:7 easy:2 enough:1 affect:1 psychology:1 favour:1 peter:1 york:2 action:1 gabriel:3 useful:1 detailed:1 involve:1 listed:1 discount:1 demonstrably:1 rw:10 generate:1 inhibitory:12 delta:4 neuroscience:2 correctly:1 klein:1 affected:2 key:7 nevertheless:1 changing:2 neither:1 shock:1 sum:1 inverse:1 uncertainty:1 thorn:1 reporting:1 separation:2 delivery:5 interleaved:1 capturing:1 konorski:4 activity:2 strength:1 constraint:2 dominated:1 aspect:2 extremely:1 pavlovian:4 department:1 according:2 combination:1 describes:1 em:1 character:1 wi:11 making:3 happens:2 repair:1 pr:3 computationally:2 equation:10 remains:1 discus:1 turn:1 ge:1 end:2 studying:1 available:1 gaussians:1 opponent:2 probe:2 apply:1 appropriate:4 alternative:2 slower:2 jd:1 substitute:1 trouble:2 embodies:1 disappear:1 classical:2 society:2 implied:1 pergamon:1 added:1 traditional:1 win:1 cw:1 attentional:2 separate:1 capacity:2 argue:1 presynaptic:1 considers:1 reason:3 modeled:1 relationship:1 index:1 ql:3 expense:1 negative:2 aversive:1 implementation:1 collective:1 perform:3 allowing:1 pearce:4 neurobiology:1 variability:2 communication:1 hinton:1 omission:1 introduced:1 specified:2 philosophical:1 distinction:1 conflicting:1 established:2 suggested:1 corbit:2 perception:1 challenge:1 rf:2 oj:1 royal:1 explanation:2 event:2 natural:1 force:1 difficulty:1 wanting:1 counterbalance:1 extract:1 review:7 prior:1 expect:2 bear:1 a11996:1 suggestion:2 interesting:1 foundation:2 contingency:1 consistent:3 rubin:1 editor:3 share:2 land:1 pi:4 course:3 side:1 weaker:1 explaining:1 fall:1 face:1 wagner:4 emerge:1 schmajuk:1 calculated:1 world:4 made:3 collection:1 reinforcement:2 schultz:2 founded:2 transaction:1 prokasy:1 sj:2 neurobiological:1 status:1 ml:2 why:2 table:4 learn:3 reasonably:1 nature:1 operational:1 da:6 whole:3 noise:2 nothing:1 lesion:1 body:1 fashion:1 downwards:17 bordley:2 ny:2 formalisation:1 fails:2 candidate:2 third:1 normative:1 symbol:1 r2:2 evidence:2 importance:1 texture:1 gallagher:2 nat:1 phd:1 conditioned:4 gluck:2 surprise:1 sophistication:1 likely:2 unexpected:1 holland:4 ma:3 cti:1 comparator:2 presentation:2 shared:1 absence:7 man:1 change:5 experimentally:1 included:1 justify:1 called:5 partly:1 experimental:3 puzzling:1 support:2 latter:1 theresa:1 tested:1 phenomenon:1 mackintosh:8 |
507 | 1,464 | Regularisation in Sequential Learning
Algorithms
J oao FG de Freitas
Cambridge University
Engineering Department
Cambridge CB2 IPZ England
jfgf@eng.cam.ac.uk
[Corresponding author]
Mahesan Niranjan
Cambridge University
Engineering Department
Cambridge CB2 IPZ England
niranjan@eng.cam.ac.uk
Andrew H Gee
Cambridge University
Engineering Department
Cambridge CB2 IPZ England
ahg@eng.cam.ac.uk
Abstract
In this paper, we discuss regularisation in online/sequential learning algorithms. In environments where data arrives sequentially,
techniques such as cross-validation to achieve regularisation or
model selection are not possible. Further, bootstrapping to determine a confidence level is not practical. To surmount these
problems, a minimum variance estimation approach that makes use
of the extended Kalman algorithm for training multi-layer perceptrons is employed. The novel contribution of this paper is to show
the theoretical links between extended Kalman filtering, Sutton's
variable learning rate algorithms and Mackay's Bayesian estimation framework. In doing so, we propose algorithms to overcome
the need for heuristic choices of the initial conditions and noise
covariance matrices in the Kalman approach.
1
INTRODUCTION
Model estimation involves building mathematical representations of physical processes using measured data. This problem is often referred to as system identification, time-series modelling or machine learning. In many occasions, the system
being modelled varies with time. Under this circumstance, the estimator needs to be
Regularisation in Sequential Learning Algorithms
459
updated sequentially. Online or sequential learning has many applications in tracking and surveillance, control systems, fault detection, communications, econometric
systems, operations research, navigation and other areas where data sequences are
often non-stationary and difficult to obtain before the actual estimation process.
To achieve acceptable generalisation, the complexity of the estimator needs to be
judiciously controlled. Although there are various reliable schemes for controlling
model complexity when training en bloc (batch processing), the same cannot be
said about sequential learning. Conventional regularisation techniques cannot be
applied simply because there is no data to cross-validate. Consequently, there is
ample scope for the design of sequential methods of controlling model complexity.
2
NONLINEAR ESTIMATION
A dynamical system may be described by the following discrete, stochastic state
space representation:
(1)
Wk +dk
(2)
g(Wk, tk) + Vk
where it has been assumed that the model parameters (Wk E R<J) constitute the
states of the system, which in our case represent the weights of a multi-layer perceptron (MLP). g is a nonlinear vector function that may change at each estimation
step k, tk denotes the time at the k-th estimation step and dk and Vk represent
zero mean white noise with covariances given by Qk and Rk respectively. The noise
terms are often called the process noise (dk) and the measurement noise (Vk). The
system measurements are encoded in the output vector Yk E Rm.
The estimation problem may be reformulated as having to compute an estimate
Wk of the states Wk using the set of measurements Yk = {Yl, Y2, "', Yk}. The
estimate Wk can be used to predict future values of the output y. We want Wk to be
an unbiased, minimum variance and consistent estimate (Gelb 1984). A minimum
variance (unbiased) estimate is one that has its variance less than or equal to that of
any other unbiased estimator. Since the variance of the output Y depends directly
on the variance of the parameter estimates (Astrom 1970), the minimum variance
framework constitutes a regularisation scheme for sequential learning.
The conditional probability density function of Wk given Yk (p(wkIYk)) constitutes
the complete solution of the estimation problem (Bar-Shalom and Li 1993, Ho and
Lee 1964, Jazwinski 1970). This is simply because p(wkIYk) embodies all the statistical information about Wk given the measurements Yk and the initial condition Woo
This is essentially the Bayesian approach to estimation, where instead of describing
a model by a single set of parameters, it is expressed in terms of the conditional
probability p(wkIYk) (Jaynes 1986, Jazwinski 1970). The estimate Wk can be computed from p(wklY k) according to several criteria, namely MAP estimation (peak
of the posterior), minimum variance estimation (centroid of the posterior) and minimax estimation (median of the posterior).
The Bayesian solution to the optimal estimation problem is (Ho and Lee 1964):
P(Wk+1,Yk+I IYk)
p(Yk+1I Yk)
J p(Yk+1IYk, Wk+l )p(wk+1lwk)P(Wk IYk)dwk
J J p(Yk+lIYk, Wk+1 )p(Wk+llwk)p(Wk IYk)dwk+l dWk (3)
where the integrals run over the parameter space. This functional integral difference
equation governing the evolution of the posterior density function is not suitable
1. R G. d. Freitas, M. Niranjan andA. H. Gee
460
for practical implementation (Bar-Shalom and Li 1993, Jazwinski 1970). It involves
propagating a quantity (the posterior density function) that cannot be described
by a finite number of parameters. The situation in the linear case is vastly simpler.
There the mean and covariance are sufficient statistics for describing the Gaussian
posterior density function.
In view of the above statements, it would be desirable to have a framework for nonlinear estimation similar to the one for linear-Gaussian estimation. The extended
Kalman filter (EKF) constitutes an attempt in this direction (Bar-Shalom and Li
1993, Gelb 1984). The EKF is a minimum variance estimator based on a Taylor
series expansion of the nonlinear function g(w) around the previous estimate. The
EKF equations for a linear expansion are given by:
+ Qk)Gk+l [Rk + Gk+1 (Pk + Qk)Gk+1]-1
Wk + Kk+l(Yk+l - Gk+l Wk)
Pk + Qk - Kk+lGk+l (Pk + Qk)
(Pk
(4)
(5)
(6)
where Pk denotes the covariance of the weights. In the general multiple input,
multiple output (MIMO) case, g E ~m is a vector function and G represents the
Jacobian of the network outputs with respect to the weights.
The EKF provides a minimum variance Gaussian approximation to the posterior
probability density function. In many cases, p(wkIYk) is a multi-modal (several
peaks) function. In this scenario, it is possible to use a committee of Kalman
filters, where each individual filter approximates a particular mode, to produce
a more accurate approximation (Bar-Shalom and Li 1993, Kadirkamanathan and
Kadirkamanathan 1995). The parameter covariances of the individual estimators
may be used to determine the contribution of each estimator to the committee. In
addition, the parameter covariances serve the purpose of placing confidence intervals
on the output prediction.
3
TRAINING MLPs WITH THE EKF
One of the earliest implementations of EKF trained MLPs is due to Singhal and
Wu (Singhal and Wu 1988). In their method, the network weights are grouped
into a single vector w that is updated in accordance with the EKF equations. The
entries of the Jacobian matrix are calculated by back-propagating the m output
values through the network.
The algorithm proposed by Singhal and Wu requires a considerable computational
effort. The complexity is of the order mq2 multiplications per estimation step. Shah,
Palmieri and Datum (1992) and Puskorius and Feldkamp (1991) have proposed
strategies for decoupling the global EKF estimation algorithm into local EKF estimation sub-problems, thereby reducing the computational time. The EKF is an improvement over conventional MLP estimation techniques, such as back-propagation,
in that it makes use of second order statistics (covariances). These statistics are
essential for placing error bars on the predictions and for combining separate networks into committees of networks. Further, it has been proven elsewhere that the
back-propagation algorithm is simply a degenerate of the EKF algorithm (Ruck,
Rogers, Kabrisky, Maybeck and Oxley 1992).
However, the EKF algorithm for training MLPs suffers from serious difficulties,
namely choosing the initial conditions (wo, Po) and the noise covariance matrices
Rand Q. In this work, we propose the use of maximum likelihood techniques,
such as back-propagation computed over a small set of initial data, to initialise the
Regularisation in Sequential Learning Algorithm~
461
EKF-MLP estimator. The following two subsections? describe ways of overcoming
the difficulty of choosing R and Q.
3.1
ELIMINATING Q BY UPDATING P WITH
BACK-PROPAGATION
.
To circumvent the problem of choosing the process noise covariance Q, while at the
same time increasing computational efficiency, it is possible to extend an algorithm
proposed by Sutton (Sutton 1992) to the nonlinear case. In doing so, the weights covariance is approximated by a diagonal matrix with entries given by pqq = exp(,8q),
where,8 is updated by error back-propagation (de Freitas, Niranjan and Gee 1997).
The Kalman gain K k and the weights estimate Wk are updated using a variation of
the Kalman equations, where the Kalman gain and weights update equations are
independent of Q (Gelb 1984), while the weights covariance P is updated by backpropagation. This algorithm lessens the burden of choosing the matrix Q by only
having to choose the learning rate scalar 1]. The performance of the EKF algorithm
with P updated by back-propagation will be analysed in Section 4.
3.2
KALMAN FILTERING AND BAYESIAN TECHNIQUES
A further improvement on the EKF algorithm for training MLPs would be to update
Rand Q automatically each estimation step. This can be done by borrowing some
ideas from the Bayesian estimation field. In particular, we shall attempt to link
Mackay's work (Mackay 1992, Mackay 1994) on Bayesian estimation for neural
networks with the EKF estimation framework. This theoretical link should serve
to enhance both methods.
Mackay expresses the prior, likelihood and posterior density functions in terms of
the following Gaussian approximations:
1
(a
2
(7)
p(w) = (27r)q/2 a -q/2 exp - "2llwll )
1
,8~
2
p(Yklw) = (27r)n/2,8-n/2 exp ( - "2 L.,..(Yk - fn,q(w, <Pk)) )
A
(8)
k=l
p(wIYk )
1
1
T
= (27r)q/2IAI- 1 / 2 exp ( - 2(w - WMP) A(w - WMP))
(9)
where in,q(w, <Pk) represents the estimator and the hyper-parameters a and ,8 control the variance of the prior distribution of weights and the variance of the measurement noise. a also plays the role of the regularisation coefficient. The posterior
is obtained by approximating it with a Gaussian function, whose mean wMP is given
by a minimum of the following regularised error function:
a
S(w) = "2llwl12
,8~
+ "2 L.,..(Yk -
A
fn,q(w, <Pk))
2
(10)
k==l
The posterior covariance A is the Hessian of the above error function.
In Mackay's estimation framework, also known as the evidence framework, the
parameters ware obtained by minimising equation (10), while the hyper-parameters
a and ,8 are obtained by maximising the evidence p(Yk la,,8) after approximating
the posterior density function by a Gaussian function. In doing so, the following
recursive formulas for a and ,8 are obtained:
'1
n-'1
ak+1 = L: q
2
and
,8k+1 =
n
2
i=l Wi
L::k=l (Yk - in,q(Wk, <Pk))
A
J. F. G. d. Freitas, M. Niranjan andA. H. Gee
462
The quantity 'Y represents the effective number of parameters 'Y = 2J~=1 >.:~a' where
the Ai correspond to the eigenvalues of the Hessian of the error function without
the regularisation term .
Instead of adopting Mackay's evidence framework, it is possible to maximise the
posterior density function by performing integrations over the hyper-parameters
analytically (Buntine and Weigend 1991, Mackay 1994). The latter approach is
known as the MAP framework for 0 and {3. The hyper-parameters computed by
the MAP framework differ from the ones computed by the evidence framework in
that the former makes use of the total number of parameters and not only the
effective number of parameters. That is, 0 and {3 are updated according to:
q
n
Ok+l = ",q
2
and
{3k+1 =
n
2
L.,..i=l Wi
l:k=l (Yk - /n,q(Wk , <Pk))
A
By comparing the equations for the prior, likelihood and posterior density functions
in the Kalman filtering framework (Ho and Lee 1964) with equations (7), (8) and
(9) we can establish the following relations:
P=A- 1
Q=o-IIq_A- 1
,
and
R={3-1Im
where Iq and 1m represent identity matrices of sizes q and m respectively. Therefore,
it is possible to update Q and R sequentially by expressing them in terms of the
sequential updates of 0 and {3.
-60~--:'::
10---::2'=""0-
::'::30-
--=4'=""0-::'::5o----:6O
':--7=o----:8o~-90
:'::---,-'
100
I
-~-8~10~82~0-~
~M~0--~
~860~-8~
70--880
~-8~90-~900
I
Figure 1: Prediction using the conventional EKF algorithm for a network with 20
hidden neurons. Actual output [. . .J and estimated output [-J.
4
RESULTS
To compare the performance of the conventional EKF algorithm, the EKF algorithm with P updated by back-propagation, and the EKF algorithm with Rand
Q updated sequentially according to the Bayesian MAP framework, noisy data was
generated from the following nonlinear, non-stationary, multivariate process:
Y
(t) - { Xl (t) + X2 (t) + v(t)
4sin(xdt)) + X2(t) sin(0.03(t - 200)) + v(t}
200
1 ::; t ::; 200
< t ::; 1000
Regularisation in Sequential Learning Algorithms
463
where the inputs Xi are uniformly distributed random sequences with variance equal .
to 1 and v(t) corresponds to uniformly distributed noise with variance equal to 0.1.
Figure 1 shows the prediction obtained using the conventional EKF algorithm. To
\
"
," ,
,,
3.S
~
I
,
I
,
I
,
I
,
I
,
3
.?
ri
s
I
2
~ 1.5
a:
05
10
Tnal
12
14
16
18
20
Figure 2: Output error for the conventional EKF algorithm [. .. ], the EKF algorithm
with P updated by back-propagation [- . -], the EKF algorithm with Rand Q
updated sequentially according to the Bayesian MAP framework [-], and the EKF
algorithm with the Bayesian evidence framework [- - -] .
compare the four estimation frameworks, an MLP with 20 neurons in the hidden
layer was selected. The initial conditions were obtained by using back-propagation
on the first 100 samples and assigning to P a diagonal matrix with diagonal elements
equal to 10. The matrices R and Q in the conventional EKF algorithm were chosen,
by trial and error, to be identity matrices. In the EKF algorithm with P updated
by back-propagation, R was chosen to be equal to the identity matrix, while the
learning rate was set to 0.01. Finally, in the EKF algorithm with Rand Q updated
sequentially, the initial Rand Q matrices were chosen to be identity matrices. The
prediction errors obtained for each method with random input data are shown in
Figure 2.
It is difficult to make a fair comparison between the four nonlinear estimation
methods because their parameters were optimised independently. However, the
results suggest that the prediction obtained with the conventional EKF training
outperforms the predictions of the other methods. This may be attributed to the
facts that, firstly, in this simple problem it is possible to guess the optimal values for
Rand Q and, secondly, the algorithms to update the noise covariances may affect
the regularisation performance of the EKF algorithm. This issue, and possible
solutions, is explored in depth by the authors in (de Freitas et al. 1997).
5
Conclusions
In this paper, we point out the links between Kalman filtering, gradient descent
algorithms with variable learning rates and Bayesian estimation. This results in two
algorithms for eliminating the problem of choosing the initial conditions and the
noise covariance matrices in the training of MLPs with the EKF. These algorithms
are illustrated on a toy problem here, but more extensive experiments have been
reported in (de Freitas et al. 1997).
Improved estimates may be readily obtained by combining the estimators into com-
464
J. R G. d. Freitas, M. Niranjan and A. H. Gee
mit tees or extending the training methods to recurrent networks. Finally, the computational time may be reduced by decoupling the network weights.
Acknowledgements
Joao FG de Freitas is financially supported by two University of the Witwatersrand
Merit Scholarships, a Foundation for Research Development Scholarship (South
Africa) and a Trinity College External Studentship (Cambridge).
References
Astrom, K. J. (1970). Introduction to Stochastic Control Theory, Academic Press.
Bar-Shalom, Y. and Li, X. R. (1993). Estimation and Tracking: Principles, Techniques and Software, Artech House, Boston.
Buntine, W. L. and Weigend, A. S. (1991). Bayesian back-propagation, Complex
Systems 5: 603-643.
de Freitas, J., Niranjan, M. and Gee, A. (1997).
Hierarchichal BayesianKalman models for regularisation and ARD in sequential learning, Technical Report CUED/F-INFENG/TR 307, Cambridge University, http:j jsvrwww .eng.cam.ac.ukj-jfgf.
Gelb, A. (ed.) (1984). Applied Optimal Estimation, MIT Press.
Ho, Y. C. and Lee, R. C. K. (1964). A Bayesian approach to problems in stochastic
estimation and control, IEEE Transactions on Automatic Control AC-9: 333339.
Jaynes, E. T. (1986). Bayesian methods: General background, in J. H. Justice (ed.),
Maximum Entropy and Bayesian Methods in Applied Statistics, Cambridge
University Press, pp. 1-25.
Jazwinski, A. H. (1970). Stochastic Processes and Filtering Theory, Academic Press.
Kadirkamanathan, V. and Kadirkamanathan, M. (1995). Recursive estimation
of dynamic modular RBF networks, in D. S. Touretzky, M. C. Mozer and
M. E. Hasselmo (eds), Advances in Neural Information Processing Systems 8,
pp. 239-245.
Mackay, D. J. C. (1992). Bayesian interpolation, Neural Computation 4(3): 415-447.
Mackay, D. J. C. (1994). Hyperparameters: Optimise or integrate out?, in G. Heidbreder (ed.), Maximum Entropy and Bayesian Methods.
Puskorius, G. V. and Feldkamp, 1. A. (1991). Decoupled extended Kalman filter
training of feedforward layered networks, International Joint Conference on
Neural Networks, Seattle, pp. 307-312.
Ruck, D. W., Rogers, S. K., Kabrisky, M., Maybeck, P. S. and Oxley, M. E. (1992).
Comparative analysis of backpropagation and the extended Kalman filter for
training multilayer perceptrons, IEEE Transactions on Pattern Analysis and
Machine Intelligence 14(6): 686-690.
Shah, S., Palmieri, F. and Datum, M. (1992). Optimal filtering algorithms for fast
learning in feedforward neural networks, Neural Networks 5: 779-787.
Singhal, S. and Wu, 1. (1988). Training multilayer perceptrons with the extended
Kalman algorithm, in D. S. Touretzky (ed.), Advances in Neural Information
Processing Systems, Vol. 1, San Mateo, CA, pp. 133-140.
Sutton, R. S. (1992). Gain adaptation beats least squares?, Proceedings of the
Seventh Yale Workshop on Adaptive Learning Systems, pp. 161-166.
| 1464 |@word trial:1 eliminating:2 justice:1 lwk:1 eng:4 covariance:14 thereby:1 tr:1 initial:7 series:2 outperforms:1 freitas:9 africa:1 comparing:1 jaynes:2 com:1 analysed:1 assigning:1 readily:1 fn:2 update:5 stationary:2 intelligence:1 selected:1 guess:1 provides:1 firstly:1 simpler:1 mathematical:1 multi:3 feldkamp:2 automatically:1 actual:2 increasing:1 heidbreder:1 joao:1 bootstrapping:1 rm:1 uk:3 control:5 before:1 maximise:1 engineering:3 accordance:1 local:1 sutton:4 ak:1 ware:1 optimised:1 interpolation:1 mateo:1 practical:2 recursive:2 witwatersrand:1 backpropagation:2 cb2:3 area:1 confidence:2 suggest:1 cannot:3 selection:1 layered:1 conventional:8 map:5 independently:1 estimator:9 initialise:1 variation:1 updated:13 controlling:2 play:1 regularised:1 element:1 approximated:1 updating:1 role:1 yk:16 mozer:1 environment:1 complexity:4 cam:4 dynamic:1 trained:1 serve:2 efficiency:1 po:1 joint:1 various:1 fast:1 describe:1 effective:2 hyper:4 choosing:5 whose:1 heuristic:1 encoded:1 modular:1 statistic:4 noisy:1 online:2 sequence:2 eigenvalue:1 propose:2 adaptation:1 combining:2 degenerate:1 achieve:2 validate:1 seattle:1 extending:1 produce:1 comparative:1 tk:2 cued:1 iq:1 andrew:1 ac:5 propagating:2 recurrent:1 measured:1 ard:1 involves:2 differ:1 direction:1 filter:5 stochastic:4 rogers:2 trinity:1 secondly:1 im:1 around:1 exp:4 scope:1 predict:1 purpose:1 estimation:32 lessens:1 oxley:2 grouped:1 hasselmo:1 mit:2 gaussian:6 ekf:31 surveillance:1 earliest:1 vk:3 improvement:2 modelling:1 likelihood:3 centroid:1 jazwinski:4 borrowing:1 relation:1 hidden:2 issue:1 development:1 integration:1 mackay:10 equal:5 field:1 having:2 represents:3 ipz:3 placing:2 constitutes:3 future:1 report:1 serious:1 individual:2 attempt:2 detection:1 mlp:4 navigation:1 arrives:1 puskorius:2 accurate:1 integral:2 decoupled:1 taylor:1 theoretical:2 mahesan:1 singhal:4 entry:2 mimo:1 seventh:1 lgk:1 buntine:2 reported:1 varies:1 density:9 peak:2 international:1 lee:4 yl:1 bloc:1 enhance:1 vastly:1 choose:1 external:1 li:5 toy:1 de:6 wk:22 coefficient:1 depends:1 ahg:1 view:1 doing:3 contribution:2 mlps:5 square:1 variance:14 qk:5 correspond:1 modelled:1 bayesian:16 identification:1 suffers:1 touretzky:2 ed:5 pp:5 attributed:1 gain:3 palmieri:2 subsection:1 back:12 ok:1 modal:1 improved:1 rand:7 iai:1 done:1 governing:1 wmp:3 nonlinear:7 propagation:11 mode:1 building:1 y2:1 unbiased:3 evolution:1 analytically:1 former:1 illustrated:1 white:1 sin:2 criterion:1 occasion:1 complete:1 oao:1 novel:1 functional:1 physical:1 extend:1 approximates:1 measurement:5 expressing:1 cambridge:9 ai:1 automatic:1 posterior:13 multivariate:1 shalom:5 scenario:1 tee:1 fault:1 minimum:8 employed:1 determine:2 artech:1 multiple:2 desirable:1 technical:1 england:3 jfgf:2 cross:2 minimising:1 academic:2 niranjan:7 controlled:1 prediction:7 infeng:1 multilayer:2 essentially:1 circumstance:1 represent:3 adopting:1 addition:1 want:1 background:1 interval:1 median:1 south:1 ample:1 feedforward:2 affect:1 idea:1 judiciously:1 effort:1 wo:1 reformulated:1 hessian:2 constitute:1 maybeck:2 reduced:1 http:1 estimated:1 per:1 discrete:1 shall:1 vol:1 express:1 four:2 wkiyk:4 econometric:1 surmount:1 weigend:2 run:1 wu:4 acceptable:1 layer:3 datum:2 yale:1 x2:2 ri:1 software:1 performing:1 department:3 according:4 llwll:1 wi:2 equation:8 discus:1 describing:2 committee:3 merit:1 operation:1 batch:1 ho:4 shah:2 denotes:2 xdt:1 embodies:1 scholarship:2 establish:1 approximating:2 quantity:2 kadirkamanathan:4 strategy:1 diagonal:3 said:1 financially:1 gradient:1 link:4 separate:1 maximising:1 kalman:14 kk:2 difficult:2 statement:1 gk:4 design:1 implementation:2 neuron:2 finite:1 descent:1 gelb:4 beat:1 situation:1 extended:6 communication:1 overcoming:1 namely:2 extensive:1 bar:6 dynamical:1 pattern:1 reliable:1 optimise:1 suitable:1 difficulty:2 circumvent:1 minimax:1 scheme:2 iyk:4 woo:1 prior:3 acknowledgement:1 multiplication:1 regularisation:12 filtering:6 proven:1 validation:1 foundation:1 integrate:1 sufficient:1 consistent:1 principle:1 elsewhere:1 supported:1 gee:6 perceptron:1 fg:2 distributed:2 overcome:1 calculated:1 depth:1 studentship:1 author:2 adaptive:1 san:1 transaction:2 global:1 sequentially:6 anda:2 assumed:1 xi:1 ca:1 decoupling:2 expansion:2 complex:1 ruck:2 pk:10 noise:11 hyperparameters:1 fair:1 astrom:2 referred:1 en:1 sub:1 xl:1 house:1 jacobian:2 rk:2 formula:1 explored:1 dk:3 evidence:5 essential:1 burden:1 workshop:1 sequential:11 boston:1 entropy:2 simply:3 expressed:1 llwl12:1 tracking:2 scalar:1 corresponds:1 ukj:1 conditional:2 identity:4 consequently:1 rbf:1 considerable:1 change:1 generalisation:1 reducing:1 uniformly:2 dwk:3 called:1 total:1 la:1 perceptrons:3 college:1 latter:1 |
508 | 1,465 | Adaptive choice of grid and time
reinforcement learning
?
In
Stephan Pareigis
stp@numerik.uni-kiel.de
Lehrstuhl Praktische Mathematik
Christian-Albrechts-Uni versitiit Kiel
Kiel, Germany
Abstract
We propose local error estimates together with algorithms for adaptive a-posteriori grid and time refinement in reinforcement learning. We consider a deterministic system with continuous state and
time with infinite horizon discounted cost functional. For grid refinement we follow the procedure of numerical methods for the
Bellman-equation. For time refinement we propose a new criterion,
based on consistency estimates of discrete solutions of the Bellmanequation. We demonstrate, that an optimal ratio of time to space
discretization is crucial for optimal learning rates and accuracy of
the approximate optimal value function.
1
Introduction
Reinforcement learning can be performed for fully continuous problems by discretizing state space and time, and then performing a discrete algorithm like Q-Iearning
or RTDP (e.g. [5]). Consistency problems arise if the discretization needs to be
refined, e.g. for more accuracy, application of multi-grid iteration or better starting
values for the iteration of the approximate optimal value function. In [7] it was
shown, that for diffusion dominated problems, a state to time discretization ratio
k/ h of Ch'r, I > 0 has to hold, to achieve consistency (i.e. k = o(h)). It can be
shown, that for deterministic problems, this ratio must only be k / h = C, C a constant, to get consistent approximations of the optimal value function. The choice
of the constant C is crucial for fast learning rates, optimal use of computer memory
resources and accuracy of the approximation.
We suggest a procedure involving local a-posteriori error estimation for grid refinement, similar to the one used in numerical schemes for the Bellman-equation (see
[4]). For the adaptive time discretization we use a combination from step size con-
Adaptive Choice of Grid and Time in Reinforcement Learning
1037
trol for ordinary differential equations and calculations for the rates of convergence
of fully discrete solutions of the Bellman-equation (see [3]). We explain how both
methods can be combined and applied to Q-Iearning. A simple numerical example
shows the effects of suboptimal state space to time discretization ratio, and provides
an insight in the problems of coupling both schemes.
2
Error estimation for adaptive choice of grid
We want to approximate the optimal value function V :
of the following problem: Minimize
n C IRd
J(x, u(.)) :=
1
00
n -+
IR in a state space
u(.): IR+ -+ A measurable,
e- pr g(Yx,u( .)(r), u(r))dr,
(1)
where 9 : n X A -+ IR+ is the cost function, and Yx,u( .)(.) is the solution of the
differential equation
(2)
y(t) f(y(t), u(t)), y(O) x.
=
=
As a trial space for the approximation of the optimal value function (or Q-function)
we use locally linear elements on simplizes Si, i = 1, ... , N s which form a triangulation of the state space, N s the number of simplizes. The vertices shall be called
Xi, i
1, . .. , N, N the dimension of the trial space 1 . This approach has been used
in numerical schemes for the Bellman-equation ([2], [4]). We will first assume, that
the grid is fixed and has a discretization parameter
=
k
= maxdiam{Si}.
i
Other than in the numerical case, where the updates are performed in the vertices of
the triangulation, in reinforcement learning only observed information is available.
We will assume, that in one time step of size h > 0, we obtain the following
information:
? the current state Yn E
n,
? an action an E A,
? the subsequent state Yn+1 := YYn,a n(h)
? the local cost rn = r(Yn, an) =
Joh e-PTg(YYn,an(r),an(r))dr.
The state Yn, in which an update is to be made, may be any state in
finite, and an locally constant .
n. A
shall be
The new value of the fully discrete Q-function Qi (Yn, an) should be set to
shall be
where
rn
(
)
+ e -phTTk
v h Yn+l ,
V; (Yn+d = minaQi(Yn+l,a). We call the right side the update function
(3)
We will update Qi in the vertices {Xd~l of the triangulation in one ofthe following
two ways:
1
When an adaptive grid is used, then N s and N depend on the refinement.
Kaczmarz-update. Let
nates, such that
>.7
(AI, .. . , AN) be the vector of barycentric coordi-
N
Yn =
2: Aixi,
O:SAi:Sl, foralli=I, ... ,N.
i=1
Then update
(4)
Kronecker-update. Let 53 Yn and x be the vertex of 5, closest to Yn (if there
is a draw, then the update can be performed in all winners). Then update Q~ only
in x according to
(5)
Each method has some assets and drawbacks . In our computer simulations the
Kaczmarz-update seemed to be more stable over the Kronecker-update (see [6]) .
However , examples may be constructed where a (Holder-) continuous bounded optimal value function V is to be approximated, and the Kaczmarz-update produces
an approximation with arbitrarily high "."sup-norm (place a vertex x of the triangulation in a point where
V is infinity, and use as update states the vertex x in
turn with an arbitrarily close state x) .
d:
Kronecker-update will provide a bounded approximation if V is bounded. Let
be the fully-discrete optimal value function
Vhk (xd = min{r(xi'
a) + e-PhVhk (Yxi,a(h)),
a
Vhk
i = 1, . . . , N .
Then it can be shown, that an approximationyerformed by Kronecker-update will
(with respect to the "."sup-norm),
eventually be caught in an c-neighborhood of
if the data points Yo, Yl, Y2, . . . are sufficiently dense. Under regularity conditions
on V, c may be bounded by2
VI:
(6)
As a criterion for grid refinement we choose a form of a local a posteriori error estimate as defined in [4] . Let
(x) = mina Q~ (x, a) be the current iterate of the optimal value function. Let ax E U be the minimizing control ax = argmina Q~ (x, a).
Then we define
vI:
(7)
vI:,
If Vhk is in the c-neighborhood of
then it can be shown, that (for every x E
and simplex Sx with x E Sx, ax as above)
n
O:S e(x) :S sup P(z , az , Vhk) - inf P( z , az , Vhk).
z ESz:
z ES:t
If Vhk is Lipschitz-continuous, then an estimate using only Gronwall's inequality
bounds the right side and therefore e(x) by C
where C depends on the Lipschitzconstants of
and the cost g .
vl'
p\'
2With respect to the results in [3] we assume, that also
E:
~ C(h
+ 7;:)
can be shown.
Adaptive Choice of Grid and Time in Reinforcement Learning
1039
The value ej := maXxesj eh(x) defines a function, which is locally constant on every
simplex. We use ej, j = 1, ... , N as an indicator function for grid refinement. The
(global) tolerance value tolk for ej shall be set to
Ns
tolk
= C * (L; edlNs,
i=l
where we have chosen 1 :::; C :::; 2. We approximate the function e on the simplizes
in the following way, starting in some Yn E Sj:
1.
2.
3.
4.
apply a control a E U constantly on [T, T + h]
receive value rn and subsequent state Yn+l
calculate the update value Ph(x, a, Vf)
if (IPh(x,a, vt) - Vt(x)l ~ ej) then ej := IPh(x,a, Vhk) - Vt(x)1
It is advisable to make grid refinements in one sweep. We also store (different to
the described algorithm) several past values of ej in every simplex, to be able to
distinguish between large ej due to few visits in that simplex and the large ej due to
space discretization error. For grid refinement we use a method described in ([1]).
3
A local criterion for time refinement
Why not take the smallest possible sampling rate? There are two arguments for
adaptive time discretization. First, a bigger time step h naturally improves (decreases) the contraction rate of the iteration, which is e- ph . The new information
is conveyed from a point further away (in the future) for big h, without the need
to store intermediate states along the trajectory. It is therefore reasonable to start
with a big h and refine where needed.
The second argument is, that the grid and time discretization k and h stand in a
certain relation. In [3] the estimate
lV(x) - vt(x)1 :::; C(h +
k
..Jh)'
for all x
En,
C a constant
is proven (or similar estimates, depending on the regularity of V). For obvious
reasons, it is desirable to start with a coarse grid (storage, speed), i.e. k large.
Having a too small h in this case will make the approximation error large. Also
here, it is reasonable to start with a big h and refine where needed.
What can serve as a refinement criterion for the time step h? In numerical schemes
for ordinary differential equations, adaptive step size control is performed by estimating the local truncation error of the Taylor series by inserting intermediate
points. In reinforcement learning, however, suppose the system has a large truncation error (i.e. it is difficult to control) in a certain region using large h and locally
constant control functions. If the optimal value function is nearly constant in this
region, we will not have to refine h. The criterion must be, that at an intermediate
point, e.g. at time h12, the optimal value function assumes a value considerably
smaller (better) than at time h . However, if this better value is due to error in the
state discretization, then do not refine the time step.
?
We define a function H on the simplices of the triangulation. H(S) > holds the
time-step which will be used when in simplex S. Starting at a state Yn E n, Yn E Sn
at time T > 0, with the current iterate of the Q-function Q~ (Vhk respectively) the
following is performed:
s. Pareigis
1040
1. apply a control a E U constantly on [T, T + h]
2. take a sample at the intermediate state z = YYn,a(h/2)
3. if (H(Sn) < C*vdiam{Sn}) then end.
else:
4. compute Vl(z) = millb Q~(z, b)
5. compute Ph/2(Yn, a, Vt) = rh/2(Yn, a) + e- ph / 2Vt(z)
6. compute Ph(Yn, a, Vt) = rh(Yn, a) +e-phVl(Yn+d
7. if (Ph/ 2(Yn, a, Vhk) S Ph(Yn , a, Vhk)-tol) update H(Sn)
= H(Sn)/2
The value C is currently set to
C
2
= C(Yn, a) = -lrh/2(Yn,
a) p
rh(Yn, a)/,
whereby a local value of MI~gh2 is approximated, MJ (x)
approximation of l\7g(x, a)1 (if 9 is sufficiently regular).
= maxa If(x, a)l,
Lg an
tol depends on the local value of Vhk and is set to
tOl(x) = 0.1 * vt(x).
How can a Q-function Q:~:~(x, a), with state dependent time and space discretisation be approximated and stored? We have stored the time discretisation function
H locally constant on every simplex. This implies (if H is not constant on 0), that
there will be vertices Xj, such that adjacent triangles hold different values of H .
The Q-function, which is stored in the vertices, then has different choices of H(xj).
We solved this problem, by updating a function Q'H(Xj, a) with Kaczmarz-update
and the update value PH(Yn) (Yn , a, Vt), Yn in an to Xj adjacent simplex, regardless
of the different H-values in Xj. Q'H(Xj, a) therefore has an ambiguous semantic:
it is the value if a is applied for 'some time ', and optimal from there on. 'some
time'depends here on the value of H in the current simplex. It can be shown, that
IQ~(Xj)/2(xj,a) - Q'H(Xj)(xj,a)1 is less than the space discretization error.
4
A simple numerical example
We demonstrate the effects of suboptimal values for space and time discretisation
with the following problem. Let the system equation be
iJ = f(y, u) :=
(~1 ~) (y -
v),
v
=(
.375 )
.375
'
yEO = [0,1] x [0,1] (8)
The stationary point of the uncontrolled system is v. The eigenvalues of the system
are {u + i, U - i}, u E [-c, cJ. The system is reflected at the boundary.
The goal of the optimal control shall be steer the solution along a given trajectory in
state space (see figure 1), minimizing the integral over the distance from the current
state to the given trajectory. The reinforcement or cost function is therefore chosen
to be
g(y) dist(L, y)t,
(9)
=
where L denotes the set of points in the given trajectory. The cost functional takes
the form
( 10)
Adaptive Choice of Grid and Time in Reinforcement Learning
1041
IL
0.5
~
o~
o
______
~
______
~
0.5
Figure 1: The left picture depicts the L-form of the given trajectory. The stationary
point of the system is at (.375, .375) (depicted as a big dot). The optimal value function
computed by numerical schemes on a fine fixed grid is depicted with too large time discretization (middle) and small time discretization (right) (rotated by about 100 degrees
for better viewing). The waves in the middle picture show the effect of too large time steps
in regions where 9 varies considerably.
In the learning problem, the adaptive grid mechanism tries to resolve the waves
(figure 1, middle picture) which come from the large time discretization. This is
depicted in figure 2. We used only three different time step sizes (h
0.1, 0.05 and
0.025) and started globally with the coarsest step size 0.1.
=
Figure 2: The adaptive grid mechanism refines correctly. However, in the left picture,
unnecessary effort is spended in resolving regions, in which the time step should be refined
urgently. The right picture shows the result, if adaptive time is also used. Regions outside
the L-form are refined in the early stages of learning while h was still large. An additional
coarsening should be considered in future work. We used a high rate of random jumps in
the process and locally a certainty equivalence controller to produce these pictures.
1042
5
S. Pareigis
Discussion of the methods and conclusions
We described a time and space adaptive method for reinforcement learning with
discounted cost functional. The ultimate goal would be, to find a self tuning algorithm which locally adjusted the time and space discretization automatically to the
optimal ratio. The methods worked fine in the problems we investigated, e.g. nonlinearities in the system showed no problems. Nevertheless, the results depended
on the choice of the tolerance values C, tol and tolk' We used only three time discretization steps to prevent adjacent triangles holding time discretization values too
far apart. The smallest state space resolution in the example is therefore too fine
for the finest time resolution. A solution can be, to eventually use controls that are
of higher order (in terms of approximation of control functions) than constant (e.g.
linear, polynomial, or locally constant on subintervals of the finest time interval).
This corresponds to locally open loop controls.
The optimality of the discretization ratio time/space could not be proven. Some
discontinuous value functions 9 gave problems, and we had problems handling stiff
systems, too.
The learning period was considerably shorter (about factor 100 depending on the
requested accuracy and initial data) in the adaptive cases as opposed to fixed grid
and time with the same accuracy.
From our experience, it is difficult in numerical analysis to combine adaptive time
and space discretization methods. To our knowledge this concept has not yet been
applied to the Bellman-equation. Theoretical work is still to be done. We are aware,
that triangulation of the state space yields difficulties in implementation in high
dimensions. In future work we will be using rectangular grids. We will also make
some comparisons with other algorithms like Parti-game ([5]). To us, a challenge is
seen in handling discontinuous systems and cost functions as they appear in models
with dry friction for example, as well as algebro-differential systems as they appear
in robotics.
References
[1] E. Bansch. Local mesh refinement in 2 and 3 dimensions. IMPACT Comput.
Sci. Engrg. 3, Vol. 3:181-191, 1991.
[2] M. Falcone. A numerical approach to the infinite horizon problem of deterministic control theory. Appl Math Optim 15:1-13, 1987.
[3] R. Gonzalez and M. Tidball. On the rates of convergence of fully discrete
solutions of Hamilton-Jacobi equations. INRIA, Rapports de Recherche, No
1376, Programme 5, 1991.
[4] L. Griine. An adaptive grid scheme for the discrete Hamilton-Jacobi-Bellman
equation. Numerische Mathematik, Vol. 75, No. 3:319-337, 1997.
[5] A. W. Moore and C. G. Atkeson. The parti-game algorithm for variable resolution reinforcement learning in multidimensional state-spaces. Machine Learning,
Volume 21, 1995.
[6] S. Pareigis. Lernen der Losung der Bellman-Gleichung durch Beobachtung von
kontinuierlichen Prozepen. PhD thesis, Universitat Kiel, 1996.
[7] S. Pareigis. Multi-grid methods for reinforcement learning in controlled diffusion
processes. In D. S. Touretzky, M. C . Mozer, and M. E. Hasselmo, editors,
Advances in Neural Information Processing Systems, volume 9. The MIT Press,
Cambridge, 1997.
| 1465 |@word trial:2 middle:3 joh:1 polynomial:1 norm:2 open:1 simulation:1 contraction:1 initial:1 series:1 past:1 current:5 discretization:19 optim:1 urgently:1 si:2 yet:1 must:2 finest:2 refines:1 subsequent:2 numerical:10 mesh:1 christian:1 update:19 stationary:2 recherche:1 provides:1 coarse:1 math:1 kiel:4 albrechts:1 along:2 constructed:1 differential:4 combine:1 dist:1 multi:2 bellman:7 discounted:2 globally:1 automatically:1 resolve:1 estimating:1 bounded:4 what:1 rtdp:1 maxa:1 certainty:1 sai:1 every:4 multidimensional:1 xd:2 iearning:2 control:11 yn:28 appear:2 hamilton:2 local:9 depended:1 inria:1 equivalence:1 appl:1 kaczmarz:4 procedure:2 foralli:1 regular:1 suggest:1 get:1 close:1 storage:1 measurable:1 deterministic:3 regardless:1 starting:3 caught:1 rectangular:1 resolution:3 numerische:1 insight:1 parti:2 suppose:1 aixi:1 element:1 approximated:3 praktische:1 updating:1 observed:1 solved:1 calculate:1 region:5 decrease:1 mozer:1 trol:1 depend:1 serve:1 triangle:2 fast:1 neighborhood:2 refined:3 outside:1 eigenvalue:1 propose:2 inserting:1 loop:1 achieve:1 az:2 convergence:2 regularity:2 produce:2 yxi:1 rotated:1 coupling:1 advisable:1 depending:2 iq:1 ij:1 implies:1 come:1 drawback:1 discontinuous:2 viewing:1 adjusted:1 hold:3 sufficiently:2 considered:1 early:1 smallest:2 estimation:2 currently:1 by2:1 esz:1 hasselmo:1 mit:1 ej:8 ax:3 yo:1 posteriori:3 dependent:1 vl:2 relation:1 germany:1 stp:1 aware:1 having:1 sampling:1 nearly:1 future:3 simplex:8 few:1 integral:1 experience:1 shorter:1 discretisation:3 taylor:1 theoretical:1 steer:1 ordinary:2 cost:8 vertex:8 too:6 universitat:1 stored:3 varies:1 considerably:3 combined:1 yl:1 together:1 von:1 thesis:1 opposed:1 choose:1 dr:2 yeo:1 nonlinearities:1 de:2 rapport:1 vi:3 depends:3 performed:5 try:1 sup:3 start:3 wave:2 minimize:1 il:1 ir:3 accuracy:5 holder:1 yield:1 ofthe:1 dry:1 trajectory:5 asset:1 explain:1 touretzky:1 obvious:1 naturally:1 mi:1 jacobi:2 con:1 pareigis:5 knowledge:1 improves:1 cj:1 higher:1 follow:1 reflected:1 lehrstuhl:1 done:1 stage:1 defines:1 effect:3 concept:1 y2:1 moore:1 semantic:1 adjacent:3 game:2 self:1 ambiguous:1 whereby:1 criterion:5 mina:1 demonstrate:2 functional:3 winner:1 volume:2 cambridge:1 ai:1 tuning:1 grid:24 consistency:3 engrg:1 had:1 dot:1 stable:1 argmina:1 closest:1 showed:1 triangulation:6 stiff:1 inf:1 apart:1 store:2 certain:2 inequality:1 discretizing:1 arbitrarily:2 durch:1 vt:9 der:2 seen:1 additional:1 period:1 resolving:1 desirable:1 calculation:1 visit:1 bigger:1 controlled:1 qi:2 impact:1 involving:1 controller:1 iteration:3 robotics:1 receive:1 want:1 fine:3 interval:1 else:1 crucial:2 coarsening:1 call:1 intermediate:4 stephan:1 iterate:2 xj:10 gave:1 suboptimal:2 ultimate:1 effort:1 ird:1 action:1 tol:4 locally:9 ph:8 sl:1 correctly:1 discrete:7 shall:5 vol:2 nevertheless:1 prevent:1 diffusion:2 place:1 reasonable:2 h12:1 draw:1 gonzalez:1 vf:1 bound:1 uncontrolled:1 distinguish:1 refine:4 kronecker:4 infinity:1 worked:1 dominated:1 speed:1 argument:2 min:1 optimality:1 friction:1 performing:1 coarsest:1 according:1 coordi:1 combination:1 smaller:1 pr:1 equation:11 resource:1 mathematik:2 turn:1 eventually:2 mechanism:2 needed:2 end:1 available:1 apply:2 away:1 ptg:1 assumes:1 denotes:1 yx:2 sweep:1 distance:1 sci:1 reason:1 ratio:6 minimizing:2 difficult:2 lg:1 holding:1 implementation:1 finite:1 rn:3 barycentric:1 able:1 challenge:1 memory:1 difficulty:1 eh:1 indicator:1 scheme:6 picture:6 started:1 sn:5 fully:5 proven:2 lv:1 degree:1 conveyed:1 consistent:1 editor:1 truncation:2 side:2 jh:1 lrh:1 tolerance:2 boundary:1 dimension:3 stand:1 seemed:1 made:1 adaptive:17 reinforcement:12 refinement:12 jump:1 programme:1 far:1 atkeson:1 sj:1 approximate:4 uni:2 global:1 unnecessary:1 xi:2 continuous:4 why:1 mj:1 subintervals:1 requested:1 investigated:1 dense:1 rh:3 big:4 arise:1 en:1 depicts:1 simplices:1 n:1 comput:1 nates:1 phd:1 sx:2 horizon:2 depicted:3 ch:1 iph:2 corresponds:1 constantly:2 goal:2 lipschitz:1 infinite:2 called:1 e:1 handling:2 |
509 | 1,466 | Independent Component Analysis for
identification of artifacts in
Magnetoencephalographic recordings
Ricardo Vigario 1 ; Veikko J ousmiiki2 ,
Matti Hiimiiliiinen2, Riitta Hari2, and Erkki Oja 1
1 Lab.
of Computer & Info. Science
Helsinki University of Technology
P.O. Box 2200, FIN-02015 HUT, Finland
{Ricardo.Vigario, Erkki.Oja}@hut.fi
2 Brain
Research Unit, Low Temperature Lab.
Helsinki University of Technology
P.O. Box 2200, FIN-02015 HUT, Finland
{veikko, msh, hari}@neuro.hut.fi
Abstract
We have studied the application of an independent component analysis
(ICA) approach to the identification and possible removal of artifacts
from a magnetoencephalographic (MEG) recording. This statistical technique separates components according to the kurtosis of their amplitude
distributions over time, thus distinguishing between strictly periodical
signals, and regularly and irregularly occurring signals. Many artifacts
belong to the last category. In order to assess the effectiveness of the
method, controlled artifacts were produced, which included saccadic eye
movements and blinks, increased muscular tension due to biting and the
presence of a digital watch inside the magnetically shielded room. The
results demonstrate the capability of the method to identify and clearly
isolate the produced artifacts.
1 Introduction
When using a magnetoencephalographic (MEG) record, as a research or clinical tool, the
investigator may face a problem of extracting the essential features of the neuromagnetic
? Corresponding author
R. Vigario,
230
v. Jousmiiki, M. Hiimiiliiinen, R. Hari and E. Oja
signals in the presence of artifacts. The amplitude of the disturbance may be higher than
that of the brain signals, and the artifacts may resemble pathological signals in shape. For
example, the heart's electrical activity, captured by the lowest sensors of a whole-scalp
magnetometer array, may resemble epileptic spikes and slow waves (Jousmili and Hari
1996).
The identification and eventual removal of artifacts is a common problem in electroencephalography (EEG), but has been very infrequently discussed in context to MEG (Hari
1993; Berg and Scherg 1994).
The simplest and eventually most commonly used artifact correction method is rejection,
based on discarding portions of MEG that coincide with those artifacts. Other methods
tend to restrict the subject from producing the artifacts (e.g. by asking the subject to fix the
eyes on a target to avoid eye-related artifacts, or to relax to avoid muscular artifacts). The
effectiveness of those methods can be questionable in studies of neurological patients, or
other non-co-operative subjects. In eye artifact canceling, other methods are available and
have recently been reviewed by Vigario (I 997b) whose method is close to the one presented
here, and in Jung et aI. (1998).
This paper introduces a new method to separate brain activity from artifacts, based on the
assumption that the brain activity and the artifacts are anatomically and physiologically
separate processes, and that their independence is reflected in the statistical relation between the magnetic signals generated by those processes.
The remaining of the paper will include an introduction to the independent component
analysis, with a presentation of the algorithm employed and some justification of this approach. Experimental data are used to illustrate the feasibility of the technique, followed
by a discussion on the results.
2
Independent Component Analysis
Independent component analysis is a useful extension of the principal component analysis
(PC A). It has been developed some years ago in context with blind source separation applications (Jutten and Herault 1991; Comon 1994). In PCA. the eigenvectors of the signal
covariance matrix C = E{xx T } give the directions oflargest variance on the input data
x. The principal components found by projecting x onto those perpendicular basis vectors
are uncorrelated, and their directions orthogonal.
However, standard PCA is not suited for dealing with non-Gaussian data. Several authors, from the signal processing to the artificial neural network communities, have shown
that information obtained from a second-order method such as PCA is not enough and
higher-order statistics are needed when dealing with the more demanding restriction of
independence (Jutten and Herault 1991; Comon 1994). A good tutorial on neural ICA implementations is available by Karhunen et al. (1997). The particular algorithm used in this
study was presented and derived by Hyvarinen and Oja (1997a. 1997b).
2.1
The model
In blind source separation, the original independent sources are assumed to be unknown,
and we only have access to their weighted sum. In this model, the signals recorded in an
MEG study are noted as xk(i) (i ranging from 1 to L, the number of sensors used, and
k denoting discrete time); see Fig. 1. Each xk(i) is expressed as the weighted sum of M
ICAfor Identification of Artifacts in MEG Recordings
231
independent signals Sk(j), following the vector expression:
M
Xk = La(j)sdj) = ASk,
(1)
j=l
where Xk = [xk(1), ... , xk(L)]T is an L-dimensional data vector, made up of the L mixtures at discrete time k. The sk(1), ... , sk(M) are the M zero mean independent source
signals, and A = [a(1), . .. , a(M)] is a mixing matrix independent of time whose elements
ail are th.e unknown coefficients of the mixtures. In order to perform ICA, it is necessary
to have at least as many mixtures as there are independent sources (L ~ M). When this
relation is not fully guaranteed, and the dimensionality of the problem is high enough,
we should expect the first independent components to present clearly the most strongly
independent signals, while the last components still consist of mixtures of the remaining
signals. In our study, we did expect that the artifacts, being clearly independent from the
brain activity, should come out in the first independent components. The remaining of the
brain activity (e.g. a and J-L rhythms) may need some further processing.
The mixing matrix A is a function of the geometry of the sources and the electrical conductivities of the brain, cerebrospinal fluid, skull and scalp. Although this matrix is unknown.
we assume it to be constant, or slowly changing (to preserve some local constancy).
The problem is now to estimate the independent signals Sk (j) from their mixtures, or the
equivalent problem of finding the separating matrix B that satisfies (see Eq. 1)
(2)
In our algorithm, the solution uses the statistical definition of fourth-order cumulant or
kurtosis that, for the ith source signal, is defined as
kurt(s(i)) = E{s(i)4} - 3[E{s(i)2}]2,
where E( s) denotes the mathematical expectation of s.
2.2 The algorithm
The initial step in source separation, using the method described in this article, is whitening, or sphering. This projection of the data is used to achieve the uncorrelation between
the solutions found, which is a prerequisite of statistical independence (Hyvarinen and Oja
1997a). The whitening can as well be seen to ease the separation of the independent signals (Karhunen et al. 1997). It may be accomplished by PCA projection: v = V x, with
E{ vv T } = I. The whitening matrix V is given by
-=T ,
V -- A- 1 / 2 .....
where A = diag[-\(1), ... , -\(M)] is a diagonal matrix with the eigenvalues of the data
covariance matrix E{xxT}, and 8 a matrix with the corresponding eigenvectors as its
columns.
Consider a linear combination y = w T v of a sphered data vector v, with Ilwll = 1. Then
E{y2} = .1 andkurt(y) = E{y4}-3, whose gradientwithrespecttow is 4E{v(wTv)3} .
Based on this, Hyvarinen and Oja (1997a) introduced a simple and efficient fixed-point
algorithm for computing ICA, calculated over sphered zero-mean vectors v, that is able to
find one of the rows of the separating matrix B (noted w) and so identify one independent
source at a time - the corresponding independent source can then be found using Eq. 2.
This algorithm, a gradient descent over the kurtosis, is defined for a particular k as
1. Take a random initial vector Wo of unit norm. Let l = 1.
232
R. Vigario,
v. Jousmiiki, M. Hiimiiliiinen, R. Hari and E. Oja
2. Let Wi = E{V(W[.1 v)3} - 3Wl-I. The expectation can be estimated using a
large sample OfVk vectors (say, 1,000 vectors).
3. Divide Wi by its norm (e.g. the Euclidean norm
4.
Ilwll = JLi wI J.
lflwT wi-II is not close enough to 1, let I = 1+1 andgo back to step 2.
Otherwise,
output the vector Wi.
In order to estimate more than one solution, and up to a maximum of lvI, the algorithm
may be run as many times as required. It is, nevertheless, necessary to remove the infonnation contained in the solutions already found, to estimate each time a different independent
component. This can be achieved, after the fourth step of the algorithm, by simply subtracting the estimated solution s = w T v from the unsphered data Xk . As the solution is
defined up to a multiplying constant, the subtracted vector must be multiplied by a vector
containing the regression coefficients over each vector component of Xk.
3
Methods
The MEG signals were recorded in a magnetically shielded room with a 122-channel
whole-scalp Neuromag-122 neuromagnetometer. This device collects data at 61 locations
over the scalp, using orthogonal double-loop pick-up coils that couple strongly to a local
source just underneath, thus making the measurement "near-sighted" (HamaHi.inen et al.
1993).
One of the authors served as the subject and was seated under the magnetometer. He kept
his head immobile during the measurement. He was asked to blink and make horizontal
saccades, in order to produce typical ocular artifacts. Moreover, to produce myographic
artifacts, the subject was asked to bite his teeth for as long as 20 seconds. Yet another
artifact was created by placing a digital watch one meter away from the helmet into the
shieded room. Finally, to produce breathing artifacts, a piece of metal was placed next
to the navel. Vertical and horizontal electro-oculograms (VEOG and HEOG) and electrocardiogram (ECG) between both wrists were recorded simultaneously with the MEG, in
order to guide and ease the identification of the independent components. The bandpassfiltered MEG (0.03-90 Hz), VEOG, HEOG, and ECG (0.1-100 Hz) signals were digitized
at 297 Hz, and further digitally low-pass filtered, with a cutoff frequency of 45 Hz and
downsampled by a factor of 2. The total length of the recording was 2 minutes. A second
set of recordings was perfonned, to assess the reproducibility of the results.
Figure 1 presents a subset of 12 spontaneous MEG signals from the frontal, temporal and
occipital areas. Due to the dimension of the data (122 magnetic signals were recorded), it
is impractical to plot all MEG signals (the complete set is available on the internet - see
reference list for the adress (Vigario 1997a?. Also both EOG channels and the electrocardiogram are presented.
4
Results
Figure 2 shows sections of9 independent components (IC's) found from the recorded data,
corresponding to a I min period, starting 1 min after the beginning of the measurements.
The first two IC's, with a broad band spectrum, are clearly due to the musclular activity
originated from the biting. Their separation into two components seems to correspond, on
the basis of the field patterns, to two different sets of muscles that were activated during
the process. IC3 and IC5 are, respectively showing the horizontal eye movements and the
eye blinks, respectively. IC4 represents cardiac artifact that is very clearly extracted. In
agreement with Jousmaki and Hari (1996), the magnetic field pattern of IC4 shows some
predominance on the left.
ICA/or Identification 0/ Artifacts in MEG Recordings
233
MEG [ 1000 fTlcm
I--
---l
saccades
I--
---l
blinking
EOG [
500 IlV
ECG [
500 IlV
I--
biting
---l MEG
~=::::::::::::::=::
~?'104~
rJ. .........
M
,J.\.......1iIIiM~..
::
t...
:;::::::;:::~=
~::::::::;::=
~ ?? ",~Jrt
..,.
t
....
~,.~ . ? .J..
.
.../\""$""~I
2 t
:;
:;
4
~
5 t
., ... ...., ,'fIJ'\,
..........-.
,..,d
,LIlt ... .,
I?
.,............. ................. "
....,..,."........ .
.... Dei ..... "
.'''IIb'''*. rt
-1I\JY. ? ---
I p", . . . . , . . . . . . . . . . . . at ...'....
I; rp ..
,P....
,
.,...............' tMn':M.U
... ,
, ..... '
U\..,.--II..------'-__
ooII..Jl,,-
".'tIItS
5 ~
6 t
VEOG
It ... 11.1. HEOG
~UijuJJJ.LU Wl Uij.lJU.LllU.UUUllUUij,UU~ijJJJ
ECG
10 s
Figure 1: Samples of MEG signals, showing artifacts produced by blinking, saccades,
biting and cardiac cycle. For each of the 6 positions shown, the two orthogonal directions
of the sensors are plotted.
The breathing artifact was visible in several independent components, e.g. IC6 and IC7. It
is possible that, in each breathing the relative position and orientation of the metallic piece
with respect to the magnetometer has changed. Therefore, the breathing artifact would be
associated with more than one column of the mixing matrix A, or to a time varying mixing
vector.
To make the analysis less sensible to the breathing artifact, and to find the remaining artifacts, the data were high-pass filtered, with cutoff frequency at 1 Hz. Next, the independent
component IC8 was found. It shows clearly the artifact originated at the digital watch,
located to the right side of the magnetometer.
The last independent component shown, relating to the first minute of the measurement,
shows an independent component that is related to a sensor presenting higher RMS (root
mean squared) noise than the others.
5
Discussion
The present paper introduces a new approach to artifact identification from MEG recordings, based on the statistical technique of Independent Component Analysis. Using this
method, we were able to isolate both eye movement and eye blinking artifacts, as well as
R. Vigario,
234
v. Jousmiiki, M HtJmlJliiinen, R. Hari and E. Oja
cardiac, myographic, and respiratory artifacts.
The basic asswnption made upon the data used in the study is that of independence between brain and artifact waveforms. In most cases this independence can be verified by the
known differences in physiological origins of those signals. Nevertheless, in some eventrelated potential (ERP) studies (e.g. when using infrequent or painful stimuli), both the
cerebral and ocular signals can be similarly time-locked to the stimulus. This local time
dependence could in principle affect these particular ICA studies. However, as the independence between two signals is a measure of the similarity between their joint amplitude
distribution and the product of each signal's distribution (calculated throughout the entire
signal, and not only close to the stimulus applied), it can be expected that the very local
relation between those two signals, during stimulation, will not affect their global statistical
relation.
6
Acknowledgment
Supported by a grant from Junta Nacional de Investiga~ao Cientifica e Tecnologica, under
its 'Programa PRAXIS XXI' (R.Y.) and the Academy of Finland (R.H.).
References
Berg, P. and M. Scherg (1994). A multiple source approach to the correction of eye
artifacts. Electroenceph. clin. Neurophysiol. 90, 229-241.
Comon, P. (1994). Independent component analysis - a new concept? Signal Processing 36,287-314.
Hamalainen, M., R. Hari, R. Ilmoniemi, 1. Knuutila, and O. Y. Lounasmaa (1993, April).
Magnetoencephalography-theory, instrumentation, and applications to noninvasive
studies of the working human brain. Reviews o/Modern Physics 65(2), 413-497.
Hari, R. (1993). Magnetoencephalography as a tool of clinical neurophysiology. In
E. Niedermeyer and F. L. da Silva (Eds.), Electroencephalography. Basic principles, clinical applications, and relatedjields, pp. 1035-1061 . Baltimore: Williams
& Wilkins.
Hyvarinen, A. and E. Oja (l997a). A fast fixed-point algorithm for independent component analysis. Neural Computation (9), 1483-1492.
Hyvarinen, A. and E. Oja (1997b). One-unit learning rules for independent component
analysis. In Neural Information Processing Systems 9 (Proc. NIPS '96). MIT Press.
Jousmiiki, Y. and R. Hari (1996). Cardiac artifacts in magnetoencephalogram. Journal
o/Clinical Neurophysiology 13(2), 172-176.
Jung, T.-P., C. Hwnphries, T.-W. Lee, S. Makeig, M. J. McKeown, Y. lragui, and
T. Sejnowski (1998). Extended ica removes artifacts from electroencephalographic
recordings. In Neural Information Processing Systems 10 (Proc. NIPS '97). MIT
Press.
Jutten, C. and 1. Herault (1991). Blind separation of sources, part i: an adaptive algorithm based on neuromimetic architecture. Signal Processing 24, 1-10.
Karhunen, J., E. Oja, L. Wang, R. Vigmo, and J. Joutsensalo (1997). A class of neural
networks for independent component analysis. IEEE Trans. Neural Networks 8(3),
1-19.
Vigmo, R. (1997a). WWW adress for the MEG data:
http://nuc1eus.hut.firrvigarioINIPS97_data.html.
Vigmo, R. (1997b). Extraction of ocular artifacts from eeg using independent component analysis. To appear in Electroenceph. c/in. Neurophysiol.
ICAfor Identification ofArtifacts in MEG Recordings
235
~~~
IC1
------,--y~-------------------------.-.------~~.. ,.. ~
U
...
IC2
IC3
.",
''' ... '' .. '
<> .
).\~
.\ C:> ?
\
~~~}a
~~-"
____I4-_. _
. . . ._.---_._. . . .-.__
. """"""?t;_-"'' '....
~
. . . . .-......,.....
~_1
IC4
IC5
IC6
~
...W"
....
"1011
...~"_f....
..".,.""'_
/tJ'IfII/'h
I' ......
d1b ..
~*W,.'tJ ......
r' .. ns...
IC7
ICB
ICg
~._-~.,.
. . . . .t . .
Wt:n:ePWt.~..,.~I'NJ'~~
I
10 s
I
Figure 2: Nine independent components found from the MEG data. For each component the
left, back and right views of the field patterns generated by these components are shown full line stands for magnetic flux coming out from the head, and dotted line the flux inwards.
| 1466 |@word neurophysiology:2 norm:3 seems:1 riitta:1 covariance:2 pick:1 initial:2 denoting:1 kurt:1 yet:1 must:1 visible:1 shape:1 remove:2 plot:1 device:1 xk:8 beginning:1 ith:1 record:1 junta:1 filtered:2 location:1 mathematical:1 magnetoencephalographic:3 inside:1 expected:1 ica:7 brain:9 electroencephalography:2 xx:1 moreover:1 lowest:1 ail:1 developed:1 finding:1 impractical:1 nj:1 temporal:1 questionable:1 makeig:1 unit:3 conductivity:1 grant:1 appear:1 producing:1 local:4 studied:1 ecg:4 scherg:2 collect:1 co:1 ease:2 perpendicular:1 locked:1 acknowledgment:1 wrist:1 area:1 projection:2 downsampled:1 tmn:1 shielded:2 onto:1 close:3 periodical:1 context:2 restriction:1 equivalent:1 www:1 williams:1 occipital:1 starting:1 painful:1 rule:1 array:1 his:2 jli:1 justification:1 target:1 spontaneous:1 infrequent:1 distinguishing:1 us:1 origin:1 agreement:1 element:1 infrequently:1 iib:1 located:1 constancy:1 electrical:2 wang:1 cycle:1 movement:3 digitally:1 asked:2 neuromagnetic:1 upon:1 basis:2 neurophysiol:2 joint:1 xxt:1 fast:1 sejnowski:1 artificial:1 whose:3 say:1 relax:1 otherwise:1 statistic:1 eigenvalue:1 kurtosis:3 subtracting:1 product:1 coming:1 biting:4 loop:1 mixing:4 reproducibility:1 achieve:1 academy:1 double:1 produce:3 mckeown:1 unsphered:1 illustrate:1 eq:2 ic1:1 resemble:2 come:1 uu:1 direction:3 waveform:1 fij:1 human:1 fix:1 ao:1 of9:1 strictly:1 extension:1 correction:2 hut:5 ic:2 finland:3 ilv:2 proc:2 infonnation:1 helmet:1 predominance:1 wl:2 tool:2 weighted:2 neuromag:1 mit:2 clearly:6 sensor:4 gaussian:1 avoid:2 varying:1 eventrelated:1 derived:1 electroencephalographic:1 underneath:1 entire:1 relation:4 uij:1 orientation:1 html:1 herault:3 icb:1 magnetically:2 field:3 ilmoniemi:1 extraction:1 placing:1 broad:1 represents:1 others:1 stimulus:3 modern:1 pathological:1 oja:11 praxis:1 preserve:1 simultaneously:1 geometry:1 jrt:1 introduces:2 mixture:5 pc:1 activated:1 tj:2 necessary:2 orthogonal:3 divide:1 euclidean:1 plotted:1 increased:1 column:2 asking:1 subset:1 nacional:1 lee:1 physic:1 squared:1 recorded:5 containing:1 slowly:1 ricardo:2 potential:1 de:1 electrocardiogram:2 coefficient:2 blind:3 piece:2 root:1 view:1 lab:2 portion:1 wave:1 capability:1 ass:2 variance:1 correspond:1 identify:2 blink:3 identification:8 produced:3 lu:1 multiplying:1 served:1 ago:1 icg:1 canceling:1 ed:1 definition:1 frequency:2 ocular:3 pp:1 hwnphries:1 associated:1 couple:1 lounasmaa:1 ask:1 dimensionality:1 blinking:3 amplitude:3 back:2 higher:3 tension:1 reflected:1 april:1 box:2 strongly:2 just:1 working:1 horizontal:3 jutten:3 artifact:38 concept:1 y2:1 during:3 noted:2 rhythm:1 presenting:1 complete:1 demonstrate:1 temperature:1 silva:1 ranging:1 epwt:1 fi:2 recently:1 common:1 stimulation:1 immobile:1 cerebral:1 jl:1 belong:1 discussed:1 he:2 relating:1 measurement:4 ai:1 bandpassfiltered:1 similarly:1 access:1 similarity:1 whitening:3 l997a:1 lju:1 instrumentation:1 accomplished:1 muscle:1 captured:1 seen:1 employed:1 period:1 signal:30 ii:2 multiple:1 full:1 rj:1 clinical:4 long:1 magnetometer:4 sdj:1 jy:1 controlled:1 feasibility:1 neuro:1 regression:1 basic:2 patient:1 expectation:2 vigario:7 achieved:1 xxi:1 operative:1 baltimore:1 source:13 recording:9 isolate:2 tend:1 subject:5 electro:1 hz:5 regularly:1 effectiveness:2 extracting:1 near:1 presence:2 enough:3 independence:6 affect:2 architecture:1 restrict:1 expression:1 pca:4 epileptic:1 rms:1 wo:1 nine:1 useful:1 eigenvectors:2 band:1 category:1 simplest:1 http:1 tutorial:1 dotted:1 estimated:2 discrete:2 tecnologica:1 nevertheless:2 changing:1 erp:1 cutoff:2 verified:1 kept:1 year:1 sum:2 run:1 fourth:2 throughout:1 separation:6 lilt:1 internet:1 followed:1 guaranteed:1 activity:6 scalp:4 i4:1 helsinki:2 erkki:2 min:2 sphering:1 according:1 combination:1 cardiac:4 wi:5 skull:1 making:1 cerebrospinal:1 comon:3 anatomically:1 projecting:1 heart:1 eventually:1 needed:1 irregularly:1 neuromimetic:1 available:3 prerequisite:1 multiplied:1 away:1 magnetic:4 subtracted:1 rp:1 original:1 denotes:1 remaining:4 include:1 clin:1 sighted:1 already:1 spike:1 saccadic:1 rt:1 dependence:1 diagonal:1 hamalainen:1 gradient:1 separate:3 separating:2 sensible:1 meg:19 length:1 y4:1 msh:1 inwards:1 info:1 fluid:1 implementation:1 ic6:2 unknown:3 perform:1 vertical:1 fin:2 descent:1 extended:1 head:2 digitized:1 jousmaki:1 community:1 introduced:1 required:1 nip:2 ic4:3 trans:1 able:2 pattern:3 breathing:5 bite:1 perfonned:1 demanding:1 disturbance:1 technology:2 dei:1 eye:9 created:1 eog:2 review:1 removal:2 meter:1 relative:1 fully:1 expect:2 digital:3 teeth:1 metal:1 article:1 principle:2 uncorrelated:1 seated:1 row:1 changed:1 jung:2 placed:1 last:3 supported:1 guide:1 side:1 vv:1 face:1 calculated:2 dimension:1 noninvasive:1 stand:1 electroenceph:2 author:3 commonly:1 made:2 coincide:1 adaptive:1 hyvarinen:5 flux:2 dealing:2 global:1 hari:10 assumed:1 inen:1 navel:1 spectrum:1 physiologically:1 sk:4 reviewed:1 channel:2 matti:1 eeg:2 metallic:1 joutsensalo:1 ic2:1 diag:1 lvi:1 did:1 da:1 whole:2 noise:1 adress:2 respiratory:1 fig:1 slow:1 n:1 position:2 originated:2 minute:2 discarding:1 showing:2 list:1 physiological:1 essential:1 consist:1 ilwll:2 knuutila:1 occurring:1 karhunen:3 rejection:1 suited:1 simply:1 expressed:1 contained:1 neurological:1 watch:3 saccade:3 veikko:2 satisfies:1 extracted:1 coil:1 presentation:1 magnetoencephalography:2 eventual:1 room:3 included:1 muscular:2 typical:1 wt:1 principal:2 total:1 pas:2 experimental:1 la:1 wilkins:1 berg:2 sphered:2 cumulant:1 frontal:1 investigator:1 |
510 | 1,467 | A Revolution: Belief Propagation
Graphs With Cycles
?
In
Brendan J. Frey?
http://wvw.cs.utoronto.ca/-frey
Department of Computer Science
University of Toronto
David J. C. MacKay
http://vol.ra.phy.cam.ac.uk/mackay
Department of Physics, Cavendish Laboratory
Cambridge University
Abstract
Until recently, artificial intelligence researchers have frowned upon
the application of probability propagation in Bayesian belief networks that have cycles. The probability propagation algorithm is
only exact in networks that are cycle-free. However, it has recently
been discovered that the two best error-correcting decoding algorithms are actually performing probability propagation in belief
networks with cycles.
1
Communicating over a noisy channel
Our increasingly wired world demands efficient methods for communicating bits of
information over physical channels that introduce errors. Examples of real-world
channels include twisted-pair telephone wires, shielded cable-TV wire, fiber-optic
cable, deep-space radio, terrestrial radio, and indoor radio. Engineers attempt
to correct the errors introduced by the noise in these channels through the use
of channel coding which adds protection to the information source, so that some
channel errors can be corrected. A popular model of a physical channel is shown
in Fig. 1. A vector of K information bits u = (Ut, ... ,UK), Uk E {O, I} is encoded,
and a vector of N codeword bits x = (Xl! ... ,XN) is transmitted into the channel.
Independent Gaussian noise with variance (12 is then added to each codeword bit,
.. Brendan Frey is currently a Beckman Fellow at the Beckman Institute for Advanced
Science and Technology, University of Illinois at Urbana-Champaign.
B. J Frey and D. J. C. MacKay
480
Gaussian noise
with variance (J2
--U--'~~I~__
E
__
nc_o_d_e_r__
~1 ~ .I~___
x
y
D_e_c_od_e_r__
~--U~.~
Figure 1: A communication system with a channel that adds Gaussian noise to the
transmitted discrete-time sequence.
producing the real-valued channel output vector y = (Y!, ... ,YN). The decoder
must then use this received vector to make a guess U at the original information
vector. The probability P" (e) of bit error is minimized by choosing the Uk that
maximizes P(ukly) for k = 1, ... , K. The rate K/N of a code is the number of
information bits communicated per codeword bit. We will consider rate ~ 1/2
systems in this paper, where N == 2K.
The simplest rate 1/2 encoder duplicates each information hit: X2k-l = X2k = Uk,
k = 1, ... , K. The optimal decoder for this repetition code simply averages together
pairs of noisy channel outputs and then applies a threshold:
Uk
= 1
if
(Y2k-l
+ Y2k)/2 > 0.5,
0 otherwise.
(1)
Clearly, this procedure has the effect of reducing the noise variance by a factor of
1/2. The resulting probability p,,(e) that an information bit will be erroneously
decoded is given by the area under the tail of the noise Gaussian:
p,,(e)
-0.5)
= 4> ( (J2/2
'
(2)
where 4>0 is the cumulative standard normal distribution. A plot of p,,(e) versus (J
for this repetition code is shown in Fig. 2, along with a thumbnail picture that shows
the distribution of noisy received signals at the noise level where the repetition code
gives p,,(e) == 10- 5 ?
More sophisticated channel encoders and decoders can be used to increase the tolerable noise level without increasing the probability of a bit error. This approach can
in principle improve performance up to a bound determined by Shannon (1948). For
a given probability of bit error P,,(e), this limit gives the maximum noise level that
can be tolerated, no matter what channel code is used. Shannon's proof was nonconstructive, meaning that he showed that there exist channel codes that achieve his
limit, but did not present practical encoders and decoders. The curve for Shannon's
limit is also shown in Fig. 2.
The two curves described above define the region of interest for practical channel
coding systems. For a given P,,(e), if a system requires a lower noise level than the
repetition code, then it is not very interesting. At the other extreme, it is impossible
for a system to tolerate a higher noise level than Shannon's limit.
2
Decoding Hamming codes by probability propagation
One way to detect errors in a string of bits is to add a parity-check bit that is
chosen so that the sum modulo 2 of all the bits is O. If the channel flips one bit,
the receiver will find that the sum modulo 2 is 1, and can detect than an error
occurred. In a simple Hamming code, the codeword x consists of the original vector
A Revolution: Belief Propagation in Graphs with Cycles
481
le-l
Shannon
limit
Concatenated
Code
le-5
'-----I~_----'c..........L..L_
0.2
_ _ _ _....L.__ _ _..L____L_......u._~LI__...J.........J
0.4
0.5
0.6
Standard deviation of Gaussian noise, U
0.3
0.7
0.8
Figure 2: Probability of bit error Ph (e) versus noise level u for several codes with
rates near 1/2, using 0/1 signalling. It is impossible to obtain a Ph(e) below Shannon's limit (shown on the far right for rate 1/2). "H-PP" = Hamming code (rate
4/7) decoded by probability propagation (5 iterations); "H-Exact" = Hamming
code decoded exactly; "LDPCC-PP" = low-density parity-check coded decoded by
probability propagation; "TC-PP" = turbo code decoded by probability propagation. The thumbnail pictures show the distribution of noisy received signals at the
noise levels where the repetition code and the Shannon limit give Ph (e) = 10- 5 .
u in addition to several parity-check bits, each of which depends on a different
subset of the information bits. In this way, the Hamming code can not only detect
errors but also correct them.
The code can be cast in the form of the conditional probabilities that specify a
Bayesian network. The Bayesian network for a K = 4, N = 7 rate 4/7 Hamming
code is shown in Fig. 3a. Assuming the information bits are uniformly random,
we have P(Uk) = 0.5, Uk E {0,1}, k = 1,2,3,4. Codeword bits 1 to 4 are direct copies of the information bits: P(xkluk) = 6(Xk,Uk), k = 1,2,3,4, where
6(a, b) = 1 if a = b and 0 otherwise. Codeword bits 5 to 7 are parity-check
bits: P(XSIUI,U2,U3) = 6(X5,Ul EB U2 EB U3), P(XaIU.,U2,U4) = 6(Xa,Ul EB U2 EB U4),
P(x7Iu2,U3,U4) = 6(X7,u2EBU3EBu4), where EB indicates addition modulo 2 (XOR).
Finally, the conditional channel probability densities are
(3)
for n = 1, ... , 7.
The probabilities P(ukly) can be computed exactly in this belief network, using
Lauritzen and Spiegelhalter's algorithm (1988) or just brute force computation.
However, for the more powerful codes discussed below, exact computations are
intractable. Instead, one way the decoder can approximate the probabilities P( Uk Iy)
is by applying the probability propagation algorithm (Pearl 1988) to the Bayesian
network. Probability propagation is only approximate in this case because the
B. 1. Frey and D. J. C. MacKay
482
(a)
(b)
(d)
(c)
Figure 3: (a) The Bayesian network for a K = 4, N = 7 Hamming code. (b) The
Bayesian network for a K = 4, N = 8 low-density parity-check code. (c) A block
diagram for the turbocode linear feedback shift register. (d) The Bayesian network
for a K = 6, N = 12 turbocode.
network contains cycles (ignoring edge directions), e.g., UI-XS-U2-X6-UI. Once a
channel output vector y is observed, propagation begins by sending a message from
Yn to Xn for n = 1, ... ,7. Then, a message is sent from Xk to Uk for k = 1,2,3,4.
An iteration now begins by sending messages from the information variables 'Ill, U2,
U3, U4 to the parity-check variables Xs, X6, X7 in parallel. The iteration finishes by
sending messages from the parity-check variables back to the information variables
in parallel. Each time an iteration is completed, new estimates of P( Uk Iy) for
k = 1,2,3,4 are obtained.
The Pb (e) - (j curve for optimal decoding and the curve for the probability propagation decoder (5 iterations) are shown in Fig. 2. Quite surprisingly, the performance
of the iterative decoder is quite close to that of the optimal decoder. Our expectation was that short cycles would confound the probability propagation decoder.
However, it seems that good performance can be obtained even when there are short
cycles in the code network.
For this simple Hamming code, the complexities of the probability propagation decoder and the exact decoder are comparable. However, the similarity in performance
between these two decoders prompts the question: "Can probability propagation
decoders give performances comparable to exact decoding in cases where exact decoding is computationally intractable?"
A Revolution: Belief Propagation in Graphs with Cycles
3
483
A leap towards the limit: Low-density parity-check codes
Recently, there has been an explosion of interest in the channel coding community in
two new coding systems that have brought us a leap closer to Shannon's limit. Both
of these codes can be described by Bayesian networks with cycles, and it turns out
that the corresponding iterative decoders are performing probability propagation
in these networks.
Fig. 3b shows the Bayesian network for a simple low-density parity-check code (Gallager 1963). In this network, the information bits are not represented explicitly.
Instead, the network defines a set of allowed configurations for the codewords. Each
parity-check vertex qi requires that the codeword bits {Xn}nEQ; to which qi is connected have even parity:
P(qil{xn}nEQ;)
= 8(qi'
EB xn),
(4)
nEQi
where q is clamped to 0 to ensure even parity. Here, Qi is the set of indices of
the codeword bits to which parity-check vertex qi is connected. The conditional
probability densities for the channel outputs are the same as in Eq. 3.
One way to view the above code is as N binary codeword variables along with a
set of linear (modulo 2) equations. If in the end we want there to be K degrees
of freedom, then the number of linearly independent parity-check equations should
be N - K. In the above example, there are N = 8 codeword bits and 4 paritychecks, leaving K = 8 - 4 = 4 degrees of freedom. It is these degrees of freedom
that we use to represent the information vector u. Because the code is linear, a
K -dimensional vector u can be mapped to a valid x simply by multiplying by an
N x K matrix (using modulo 2 addition). This is how an encoder can produce a
low-density parity-check codeword for an input vector.
Once a channel output vector y is observed, the iterative probability propagation
decoder begins by sending messages from y to x. An iteration now begins by sending
messages from the codeword variables x to the parity-check constraint variables q.
The iteration finishes by sending messages from the parity-check constraint variables
back to the codeword variables. Each time an iteration is completed, new estimates
of P(xnly) for n = 1, . . . , N are obtained. After a valid (but not necessarily correct)
codeword has been found, or a prespecified limit on the number of iterations has
been reached, decoding stops. The estimate of the codeword is then mapped back
to an estimate ii of the information vector.
Fig. 2 shows the performance of a K = 32,621, N = 65,389 low-density paritycheck code when decoded as described above. (See MacKay and Neal (1996) for
details.) It is impressively close to Shannon's limit - significantly closer than the
"Concatenated Code" (described in Lin and Costello (1983? which was considered
the best practical code until recently.
4
Another leap: Turbocodes
The codeword for a turbocode (Berrou et al. 1996) consists of the original information vector, plus two sets of bits used to protect the information. Each of these two
sets is produced by feeding the information bits into a linear feedback shift register
(LFSR), which is a type of finite state machine. The two sets differ in that one
set is produced by a permuted set of information bits; i.e., the order of the bits is
scrambled in a fixed way before the bits are fed into the LFSR. Fig. 3c shows a block
diagram (not a Bayesian network) for the LFSR that was used in our experiments.
B. 1. Frey and D. J. C. MacKay
484
Each box represents a delay (memory) element, and each circle performs addition
modulo 2. When the kth information bit arrives, the machine has a state Sk which
can be written as a binary string of state bits b4b3b2blbo as shown in the figure. bo
of the state Sk is determined by the current input Uk and the previous state Sk-l'
Bits b1 to b4 are just shifted versions of the bits in the previous state.
Fig. 3d shows the Bayesian network for a simple turbocode. Notice that each
state variable in the two constituent chains depends on the previous state and an
information bit. In each chain, every second LFSR output is not transmitted. In
this way, the overall rate of the code is 1/2, since there are K = 6 information bits
and N = 6 + 3 + 3 = 12 codeword bits. The conditional probabilities for the states
of the non permuted chain are
P(sllsl-I' Uk)
=1
if state
sl follows Sk-l for input Uk,
0 otherwise.
(5)
The conditional probabilities for the states in the other chain are similar, except that
the inputs are permuted. The probabilities for the information bits are uniform,
and the conditional probability densities for the channel outputs are the same as in
Eq.3.
Decoding proceeds with messages being passed from the channel output variables
to the constituent chains and the information bits. Next, messages are passed
from the information variables to the first constituent chain, SI. Messages are
passed forward and then backward along this chain, in the manner of the forwardbackward algorithm (Smyth et al. 1997). After messages are passed from the
first chain to the second chain s2, the second chain is processed using the forwardbackward algorithm. To complete the iteration, messages are passed from S2 to the
information bits.
Fig. 2 shows the performance of a K = 65,536, N = 131,072 turbocode when
decoded as described above, using a fixed number (18) of iterations. (See Frey
(1998) for details.) Its performance is significantly closer to Shannon's limit than the
performances of both the low-density parity-check code and the textbook standard
"Concatenated Code" .
5
Open questions
We are certainly not claiming that the NP-hard problem (Cooper 1990) of probabilistic inference in general Bayesian networks can be solved in polynomial time by
probability propagation. However, the results presented in this paper do show that
there are practical problems which can be solved using approximate inference in
graphs with cycles. Iterative decoding algorithms are using probability propagation
in graphs with cycles, and it is still not well understood why these decoders work
so well. Compared to other approximate inference techniques such as variational
methods, probability propagation in graphs with cycles is unprincipled. How well
do more principled decoders work? In (MacKay and Neal 1995), a variational decoder that maximized a lower bound on n~=1 P(ukly) was presented for low-density
parity-check codes. However, it was found that the performance of the variational
decoder was not as good as the performance of the probability propagation decoder.
It is not difficult to design small Bayesian networks with cycles for which probability
propagation is unstable. Is there a way to easily distinguish between those graphs for
which propagation will work and those graphs for which propagation is unstable? A
belief that is not uncommon in the graphical models community is that short cycles
are particularly apt to lead probability propagation astray. Although it is possible to
design networks where this is so, there seems to be a variety of interesting networks
A Revolution: Belief Propagation in Graphs with Cycles
485
(such as the Hamming code network described above) for which propagation works
well, despite short cycles.
The probability distributions that we deal with in decoding are very special distributions: the true posterior probability mass is actually concentrated in one microstate
in a space of size 2M where M is large (e.g., 10,000). The decoding problem is to
find this most probable microstate, and it may be that iterative probability propagation decoders work because the true probability distribution is concentrated in
this microstate.
We believe that there are many interesting and contentious issues in this area that
remain to be resolved.
Acknowledgements
We thank Frank Kschischang, Bob McEliece, and Radford Neal for discussions related to this work, and Zoubin Ghahramani for comments on a draft of this paper.
This research was supported in part by grants from the Gatsby foundation, the Information Technology Research Council, and the Natural Sciences and Engineering
Research Council.
References
C. Berrou and A. Glavieux 1996. Near optimum error correcting coding and decoding: Turbo-codes. IEEE 7hmsactions on Communications 44, 1261-1271.
G. F. Cooper 1990. The computational complexity of probabilistic inference using
Bayesian belief networks. Artificial Intelligence 42, 393-405.
B. J. Frey 1998. Graphical Models for Machine Learning and Digital Communication, MIT Press, Cambridge, MA. See http://vwv . cs. utoronto. cal -frey.
R. G. Gallager 1963. Low-Density Parity-Check Codes, MIT Press, Cambridge,
MA.
S. Lin and D. J. Costello, Jr. 1983. Error Control Coding: Fundamentals and
Applications, Prentice-Hall Inc., Englewood Cliffs, NJ.
S. L. Lauritzen and D. J. Spiegelhalter 1988. Local computations with probabilities
on graphical structures and their application to expert systems. Journal of the
Royal Statistical Society B 50, 157-224.
D. J. C. MacKay and R. M. Neal 1995. Good codes based on very sparse matrices.
In Cryptography and Coding. 5th IMA Conference, number 1025 in Lecture Notes
in Computer Science, 100-111, Springer, Berlin Germany.
D. J. C. MacKay and R. M. Neal 1996. Near Shannon limit performance of low
density parity check codes. Electronics Letters 32, 1645-1646. Due to editing
errors, reprinted in Electronics Letters 33, 457-458.
J. Pearl 1988. Probabilistic Reasoning in Intelligent Systems, Morgan Kaufmann,
San Mateo, CA.
C. E. Shannon 1948. A mathematical theory of communication. Bell System Technical Journal 27, 379-423, 623-656.
P. Smyth, D. Heckerman, and M. I. Jordan 1997. Probabilistic independence networks for hidden Markov probability models. Neural Computation 9, 227-270.
| 1467 |@word version:1 polynomial:1 seems:2 open:1 electronics:2 configuration:1 contains:1 phy:1 current:1 protection:1 si:1 must:1 written:1 neq:2 plot:1 intelligence:2 guess:1 signalling:1 xk:2 short:4 prespecified:1 draft:1 toronto:1 mathematical:1 along:3 direct:1 consists:2 vwv:1 manner:1 introduce:1 ra:1 increasing:1 begin:4 twisted:1 maximizes:1 mass:1 what:1 string:2 textbook:1 lfsr:4 nj:1 fellow:1 every:1 exactly:2 hit:1 uk:15 brute:1 control:1 grant:1 yn:2 producing:1 before:1 understood:1 frey:9 engineering:1 local:1 limit:13 despite:1 cliff:1 plus:1 eb:6 mateo:1 practical:4 block:2 communicated:1 procedure:1 area:2 bell:1 significantly:2 zoubin:1 shielded:1 close:2 cal:1 prentice:1 impossible:2 applying:1 correcting:2 communicating:2 his:1 cavendish:1 modulo:6 smyth:2 exact:6 element:1 particularly:1 u4:4 qil:1 observed:2 solved:2 region:1 cycle:17 connected:2 forwardbackward:2 principled:1 ui:2 complexity:2 turbocodes:1 cam:1 upon:1 easily:1 resolved:1 represented:1 fiber:1 artificial:2 choosing:1 quite:2 encoded:1 valued:1 otherwise:3 encoder:2 noisy:4 sequence:1 j2:2 achieve:1 wvw:1 constituent:3 optimum:1 wired:1 produce:1 ac:1 lauritzen:2 received:3 eq:2 c:2 differ:1 direction:1 correct:3 feeding:1 probable:1 considered:1 hall:1 normal:1 u3:4 beckman:2 radio:3 currently:1 leap:3 council:2 repetition:5 brought:1 clearly:1 mit:2 gaussian:5 check:19 indicates:1 brendan:2 detect:3 inference:4 hidden:1 germany:1 overall:1 issue:1 ill:1 special:1 mackay:9 once:2 represents:1 minimized:1 np:1 intelligent:1 duplicate:1 nonconstructive:1 ima:1 attempt:1 freedom:3 interest:2 message:12 englewood:1 certainly:1 uncommon:1 extreme:1 arrives:1 chain:10 edge:1 closer:3 explosion:1 circle:1 y2k:2 deviation:1 subset:1 vertex:2 uniform:1 delay:1 encoders:2 tolerated:1 density:13 fundamental:1 probabilistic:4 physic:1 decoding:11 together:1 iy:2 expert:1 coding:7 matter:1 inc:1 register:2 explicitly:1 depends:2 view:1 reached:1 parallel:2 xor:1 variance:3 kaufmann:1 maximized:1 bayesian:14 produced:2 multiplying:1 researcher:1 bob:1 pp:3 proof:1 hamming:9 stop:1 popular:1 ut:1 sophisticated:1 actually:2 back:3 tolerate:1 higher:1 x6:2 specify:1 editing:1 box:1 xa:1 just:2 until:2 mceliece:1 propagation:30 defines:1 believe:1 effect:1 true:2 laboratory:1 neal:5 deal:1 x5:1 complete:1 performs:1 reasoning:1 meaning:1 variational:3 recently:4 permuted:3 physical:2 b4:1 tail:1 he:1 occurred:1 discussed:1 cambridge:3 illinois:1 similarity:1 add:3 posterior:1 showed:1 codeword:17 binary:2 transmitted:3 morgan:1 berrou:2 signal:2 ii:1 champaign:1 technical:1 lin:2 coded:1 qi:5 expectation:1 iteration:11 represent:1 addition:4 want:1 diagram:2 source:1 leaving:1 comment:1 sent:1 jordan:1 near:3 variety:1 independence:1 finish:2 reprinted:1 shift:2 passed:5 ul:2 deep:1 ph:3 concentrated:2 processed:1 simplest:1 http:3 sl:1 exist:1 shifted:1 notice:1 thumbnail:2 per:1 discrete:1 vol:1 threshold:1 pb:1 contentious:1 backward:1 graph:9 sum:2 letter:2 powerful:1 comparable:2 bit:41 x2k:2 bound:2 distinguish:1 turbo:2 optic:1 constraint:2 erroneously:1 x7:2 performing:2 department:2 tv:1 jr:1 remain:1 heckerman:1 increasingly:1 cable:2 costello:2 confound:1 computationally:1 equation:2 turn:1 flip:1 fed:1 end:1 sending:6 tolerable:1 apt:1 original:3 include:1 ensure:1 completed:2 graphical:3 concatenated:3 ghahramani:1 society:1 added:1 question:2 codewords:1 kth:1 thank:1 mapped:2 berlin:1 decoder:21 astray:1 terrestrial:1 unstable:2 assuming:1 code:40 index:1 difficult:1 frank:1 claiming:1 design:2 wire:2 markov:1 urbana:1 finite:1 communication:4 discovered:1 community:2 prompt:1 david:1 introduced:1 pair:2 cast:1 protect:1 pearl:2 proceeds:1 below:2 indoor:1 royal:1 memory:1 belief:9 natural:1 force:1 advanced:1 improve:1 technology:2 spiegelhalter:2 picture:2 acknowledgement:1 lecture:1 interesting:3 impressively:1 versus:2 digital:1 foundation:1 degree:3 principle:1 surprisingly:1 parity:21 free:1 copy:1 supported:1 l_:1 institute:1 sparse:1 curve:4 feedback:2 xn:5 world:2 cumulative:1 valid:2 forward:1 san:1 far:1 unprincipled:1 approximate:4 receiver:1 b1:1 scrambled:1 iterative:5 sk:4 why:1 channel:23 ca:2 kschischang:1 ignoring:1 necessarily:1 did:1 linearly:1 s2:2 noise:14 allowed:1 cryptography:1 fig:10 gatsby:1 cooper:2 decoded:7 xl:1 clamped:1 utoronto:2 revolution:4 x:2 intractable:2 demand:1 tc:1 simply:2 gallager:2 bo:1 u2:6 applies:1 radford:1 springer:1 ma:2 conditional:6 glavieux:1 towards:1 hard:1 telephone:1 determined:2 corrected:1 reducing:1 uniformly:1 except:1 engineer:1 shannon:12 |
511 | 1,468 | Learning Human-like Knowledge by Singular
Value Decomposition: A Progress Report
Thomas K. Landauer
Darrell Laham
Department of Psychology & Institute of Cognitive Science
University of Colorado at Boulder Boulder, CO 80309-0345
{landauer, dlaham}@psych.colorado.edu
Peter Foltz
Department of Psychology
New Mexico State University Las Cruces, NM 88003-8001
pfoltz@crl.nmsu.edu
Abstract
Singular value decomposition (SVD) can be viewed as a method for
unsupervised training of a network that associates two classes of events
reciprocally by linear connections through a single hidden layer. SVD
was used to learn and represent relations among very large numbers of
words (20k-60k) and very large numbers of natural text passages (lk70k) in which they occurred. The result was 100-350 dimensional
"semantic spaces" in which any trained or newly aibl word or passage
could be represented as a vector, and similarities were measured by the
cosine of the contained angle between vectors. Good accmacy in
simulating human judgments and behaviors has been demonstrated by
performance on multiple-choice vocabulary and domain knowledge
tests, emulation of expert essay evaluations, and in several other ways.
Examples are also given of how the kind of knowledge extracted by this
method can be applied.
1 INTRODUCTION
Traditionally, imbuing machines with human-like knowledge has relied primarily on
explicit coding of symbolic facts into computer data structures and algorithms. A serious
limitation of this approach is people's inability to access and express the vast reaches of
unconscious knowledge on which they rely, knowledge based on masses of implicit
inference and irreversibly melded data. A more important deficiency of this state of affairs
is that by coding the knowledge ourselves, (as we also do when we assign subjectively
hypothesized rather than objectively identified features to input or output nodes in a neural
net) we beg important questions of how humans acquire and represent the cOOed
knowledge in the fIrSt place.
T. K. Landauer; D. Laham and P. Foltz
46
Thus. from both engineering and scientific perspectives. there are reasons to try to design
learning machines that can ocquire human-like quantities of human-like knowledge from
the same sources as humans. The success of such techniques would not prove that the
same mechanisms are used by humans. but because we presently do not know how the
problem can be solved in principle, successful simulation may offer theoretical insights
as well as practical applications. In the work reported here we have found a way to induce
significant amounts of knowledge about the meanings of passages and of their constituent
vocabularies of words by training on large bodies of natural text. In general terms, the
method simultaneously extracts the similarity between words (the likelihood of being
used in passages that convey similar ideas) and the similarity of passages (the likelihood
of containing words of similar meaning). The conjoint estimation of similarity is
accomplished by a fundamentally simple representational technique that exploits mutual
constraints implicit in the occurrences of very many words in very many contexts. We
view the resultant system both as a means for automatically learning much of the
semantic content of words and passages. and as a potential computational model for the
process that underlies the corresponding human ability.
While the method starts with data about the natural contextual co-occurrences of words. it
uses them in a different manner than has previously been applied. A long -standing
objection to co-occurrence statistics as a source of linguistic knowledge (Chomsky's
1957) is that many grammatically acceptable expressions. for example sentences with
potentially unlimited embedding structures. cannot be produced by a finite Markov
process whose elements are transition probabilities from word to word. If word-word
probabilities are insufficient to generate language. then. it is argued, acquiring estimates
of such probabilities cannot be a way that language can be learned. However, our
approach to statistical knowledge learning differs from those considered in the past in two
ways. First. the basic associational data from which knowledge is induced are not
transition frequencies between successive individual words or phrases. but rather the
frequencies with which particular words appear as components of relatively large natural
passages, utterances of the kind that humans use to convey complete ideas. The result of
this move is that the statistical regularities reflected are relations among unitary
expressions of meaning. rather than syntactic constraints on word order that may serve
additional purposes such as output and input processing efficiencies. error protection. or
esthetic style. Second, the mutual constraints inherent in a multitude of such local ~
occurrence relations are jointly satisfied by being forced into a global representation of
lower dimensionality. This constraint satisfaction. a form of induction. was accomplished
by singular value decomposition. a linear factorization technique that produces a
representational structure equivalent to a three layer neural network model with linear
activation functions.
2 THE TEXT ANALYSIS MODEL AND METHOD
The text analysis process that we have explored is called Latent Semantic Analysis (LSA)
(Deerwester et al .? 1990; Landauer and Dumais. 1997). It comprises four steps:
(1) A large body of text is represented as a matrix [ij], in which rows stand for individual
word types. columns for meaning-bearing passages such as sentences or paragraphs. mel
cells contain the frequency with which a word occurs in a passage.
(2) Cell entries (freqi) are transformed to:
log(freqi, + I)
L(( fref/,] {freq. ]'l
-1-, ~fr:~v *10 ~fre~v )
a measure of the first order association of a word and its context.
Learning Human Knowledge by Singular Value Decomposition: A Progress Report
47
(3) The matrix is then subjected to singular value decomposition (Berry, 1992):
[ij] = [ik] [kk] Uk]'
in which [ik] and Uk] have orthonormal columns, [kk] is a diagonal matrix of singular
values, and k <= max (ij).
(4) Finally, all but the d largest singular values are set to zero. Pre-multiplication of the
right-hand matrices produces a least-squares best approximation to the original matrix
given the number of dimensions, d, (hidden units in a corresponding neural net model
representation) that are retained. The SVD with dimension reduction constitutes a
constraint-satisfaction induction process in that it predicts the original observations on the
basis of linear relations among the abstracted representations of the data that it has
retained. By hypothesis, the analysis induces human-like relationships among passages
and words because humans also make inferences about semantic relationships from
abstracted representations based on limited data, and do so by an analogous process.
In the result, each word and passage is represented as a vector of length d. Performance
depends strongly on the choice of number of dimensions. The optimal number is
typically around 300. The similarity of any two words, any two text passages, or any
word and any text passage, are computed by measures on their vectors. We have most
often used the cosine (of the contained angle between the vectors in semantic d-space) which we interpret as the degree of qualitative similarity of meaning. The length of
vectors is also useful and interpretable.
3 TESTS OF LSA'S PERFORMANCE
LSA's ability to simulate human knowledge and meaning relations has been tested in a
variety of ways. Here we describe two relatively direct sources of evidence and briefly list
several others.
3.1
VOCABULARY & DOMAIN KNOWLEDGE TESTS
In all cases, LSA was ftrst trained on a large text corpus intended to be representative of
the text from which humans gain most of the semantic knowledge to be simulated. In a
previously reported test (Landauer and Dumais, 1997), LSA was trained on approximately
five million words of text sampled from a high-school level encyclopedia, then tested on
multiple choice items from the Educational Testing Service Test of English as a Foreign
Language (TOEFL). These test questions present a target word or short phrase and ask the
student to choose the one of four alternative words or phrases that is most similar in
meaning. LSA's answer was determined by computing the cosine between the derived
vector for the target word or phrase and each of the alternatives and choosing the largest.
LSA was correct on 64% of the 80 items, identical to the average of a large sample of
students from non-English speaking countries who had applied for admission to U. S.
colleges. When in error, LSA made choices positively correlated (product-moment r = .44)
with those preferred by students. We have recently replicated this result with training on a
similar sized sample from the Associated Press newswire
In a new set of tests, LSA was trained on a popular introductory psychology textbook
(Myers, 1995) and tested with the same four-alternative multiple choice tests used for
students in two large classes. In these experiments, LSA's score was about 6O%-lower
than the class averages but above passing level, and far above guessing probability. Its
errors again resembled those of students; it got right about half as many of questions rated
difftcult by the test constructors as ones rated easy (Landauer, Foltz and Laham, 1997).
3.2
ESSA Y TESTS
48
T. K. Landauer, D. Laham and P. Foltz
Word-wad meaning similarities are a good test of knowledge-indeed, vocabulary tests
are the best single measure of human intelligence. However, they are not sufficient to
assess the correspondence of LSA and human knowledge because people usually express
knowledge via larger verbal strings, such as sentences, paragraphs and articles. Thus, just
as multiple choice tests of student knowledge are often supplemented by essay tests
whose content is then judged by humans, we wished to evaluate the adequacy of LSA' s
representation of knowledge in complete passages of text. We could not have LSA write
essays because it has no means for producing sentences. However, we were able to assess
the accum::y with which LSA could extract and represent the knowledge expressed in
essays written by students by simulating judgments about their content that were made by
human readers (Landauer, Laham, Rehder, & Schreiner, in press).
In these tests, students were asked to write short essays to cover an assigned topic or to
answer a posed question. In various experiments, the topics included anatomy axI
function of the heart, phenomena from introductory psychology, the history of the
Panama Canal, and tolerance of diversity in America. In each case, LSA was first trained
either on a large sample of instructional text from the same domain or, in the latter case,
on combined text from the very large number of essays themselves, to produce a highdimensional (100-350 dimensions in the various tests) semantic space. We then
represented each essay simply as the vector average of the vectors for the words it
contained. Two properties of these average vectors were then used to measure the quality
and quantity of knowledge conveyed by an essay: (1) the similarity (measured as the
cosine of the angle between vectors) of the student essay and one or more standard essays,
and (2) the total amount of domain specific content, measured as the vector length.
In each case, two human experts independently rated the overall quality of each essay on a
five or ten point scale. The judges were either university course instructors or professional
exam readers from Educational Testing Service. The LSA measures were calibrated with
respect to the judges' rating scale in several different ways, but because they gave nearly
the same results only one will be described here. In this method, each student essay was
compared to a large (90-200) set of essays previously scored by experts, and the ten most
similar (by cosine) identified. The target essay was then assigned a "quality" score
component consisting of the cosine-weighted average of the ten. A second, "relevant
quantity", score component was the vector length of the student essay. Finally, regression
on expert scores was used to weight the quality and quantity scores (However, the weights
in all cases were so close to equal that merely adding them would have given comparable
results). Calibration was performed on data independent of that used to evaluate the
relation between LSA and expert ratings.
The correlation between the LSA score for an essay and that assigned by the average of
the human readers was .80, .64, .XX and .84 for the four sets of exams. The comparable
correlation between one reader and the other was .83, .65, .XX and .82, respectively. In
the heart topic case, each student had also taken a carefully constructed "objective" test
over the same material (a short answer test with near perfect scoring agreement). The
correlation between the LSA essay score and the objective test was .81, the average
correlation for the two expert readers .74.
A striking aspect of these results is that the LSA representations were based on analyses
of the essays that took no account of word order, each essay was treated as a "bag of
words". In extracting meaning from a text, human readers presumably rely on syntax as
well as the mere combination of words it contains, yet they were no better at agreeing on
an essay's quality or in assigning a score that predicted a performance on a separate test of
knowledge. Apparently, either the relevant information conveyed by word order in
sentences is redundant with the information that can be inferred from the combination of
words in the essay, or the processes used by LSA and humans extract different but
compensatingly useful information.
Learning HumanKnowledgc by Singular Value Decomposition: A Progress Report
3.3
49
OTHER EVIDENCE
LSA has been compared with human knowledge in several additional ways, some
confrrming the correspondence, others indicating limitations. Here are some examples, all
based on encyclopedia corpus training.
(1) Overall LSA similarity between antonyms (mean cos = .18) was equivalent to that
between synonyms (mean cos .17) in triplets sampled from an antonym/synonym
dictionary (W & R Chambers, 1989), both of which significantly exceeded that for
unrelated pairs (mean cos = .01; ps < .0001). However for antonym (but not for
synonym) pairs a dominant dimension of difference could easily be extracted by
computing a one dimensional unfolding using the LSA cosines from a set of words listed
in Rogel's (1992) thesaurus as related respectively to the two members of the pair.
=
(2) Anglin (1970) asked children and adults to sort words varying in concept relations ad
parts of speech. LSA wont-word similarity correlates .50 with children and .32 with adults
for the number of times they sorted two words together. Conceptual structure is reflected,
but grammatical classification, strong in the adult data, is not.
(3) When people are asked to decide that a letter string is a wont, they do so faster if they
have just read a sentence that does not contain the word but implies a related concept (e.g.
Till, Mross & Kintsch, 1988). LSA mirrors this result with high similarities between the
same sentences and words. (Landauer & Dumais, 1997).
(4) People frequently make a logical error, called the conjunction error by Tversky ad
Kahneman (1974), in which they estimate that the probability that an object is a member
of a class is greater than that it is a member of a superset class when the description of
the object is "similar" to the description of the subset. For example, when told that
"Linda is a young woman who is single, outspoken ...deeply concerned with issues of
discrimination and social justice," over 80% of even statistically sophisticated subjects
rate it more likely that Linda is a feminist bank teller than that Linda is a bank teller
(Tversky & Kahneman, 1980). LSA similarities between descriptions of people ad
occupations of this kind taken from Shafir, Smith and Osherson (1990) were computed as
the cosine between the vector averages of words in the paired person-occupation
descriptions. Conjunction error statements were more similar to the subset than superset
statement in 12 out of 14 cases (p<.01), showing that LSA's representation of sentential
meaning reflected similarity relations of the sort that have been hypothesized to underlie
the conjunction fallacy in human judgment.
(5) A semantic subspace was constructed for words from natural kind and artifact
categories whose differential preservation is characteristic of agnosias due to local damage
from herpes simplex encephalitis (Warrington & Shallice, 1984). Principal components
analysis of the similarities among these words as represented by LSA revealed that
categories that tend to be lost contain words that are more highly inter-related than those
in preserved categories (Laham, in press).
Of course, LSA does not capture all of the human knowledge conveyed by text. Some of
the shortfall is probably due merely to the use of training corpora that are still imperfectly
representative of the language experience of a typical person, and to lack of knowledge
from non-textual sources. For example, in all these studies, less total text was used for
LSA training than even a single educated adult would have read However, a more
fundamental restriction is that the analysis does not reflect order relations between words,
and therefore cannot extract infonnation that depends on syntax. Because the analysis
discovers and represents only unsigned continuous similarities, it can be used to induce
only certain classes of structural relations, not including ones that express Boolean, causal
or other non-commutative logical relations. As we have seen, this lack does not prevent
accurate simulation of human cognition in many cases, possibly because humans also
50
T. K Landauer, D. Laham and P. Foltz
frequently rely on similarity rather than syntax-based. discrete logic (fversky axl
Kahneman, 1983); however, it does limit the utility of the results for populating the
symbolic data structures commonly used to represent knowledge in traditional AI. On the
other hand, as examples in the next section show, continuous-valued similarity relations
can be fruitfully applied if appropriate computational use is made of them.
4 SAMPLE APPLICATIONS
LSA has been used successfully in a variety of experimental applications, including the
essay scoring techniques described earlier. Here are some additional examples:
(1) The technique has been used to improve automatic information retrieval by 20-30%
over otherwise identical methods by allowing users' queries to match documents with the
desired conceptual meaning but expressed in different words (Dumais, 1991, 1994).
(2) By training on corpora of translated text in which the words of corresponding
paragraphs in the two languages are combined in the "bags of words", LSA has been able
to provide at least as good retrieval when queries and documents are in a different language
as when in the same language (Landauer and Littman, 1990).
(3) LSA-based measures of the similarity of student essays on a topic to instructional
texts can predict how much an individual student will learn from a particular text (Wolfe
et al., in press; Rehder et al., in press). To do this, the full set of student essays and the
texts in question are aligned along a single dimension that best accommodates the LSA
similarities among them. Estimates from one such experiment showed that using LSA to
choose the optimal one of four texts for each student (a text that is slightly more
sophisticated than the student) rather than assigning all students the overall best text
(which LSA also picked correctly) increased the average amount learned by over 40%.
(4) LSA-based measures of conceptual similarity between successive sentences accurately
predict differences in judged coherence and measured comprehensibility of text (Foltz,
Kintsch and Landauer, in press).
5 SUMMARY
SVD-based learning of the structure underlying the use of words in meaningful contexts
has been found capable of deriving and representing the similarity of meaning of words
and text passages in a manner that accurately simulates corresponding similarity relations
as reflected in several sorts of human judgments and behavior. The validity of the
resulting representation of meaning similarities has been established in a variety of ways,
and the utility of its knowledge representation illustrated by several educational axl
cognitive psychological research applications. It is obviously too early to assess whether
the particular computational model is a true analog of the process used by the human
brain to accomplish the same things. However, the basic process, the representation of
myriad local associative relations between components and larger contexts of experience
in a joint space of lower dimensionality, offers, for the first time, a candidate for such a
mechanism that has been shown sufficient to approximate human knowledge acquisition
from natural sources at natural scale.
Acknowledgments
We thank members of the LSA research group at the University of Colorado for valuable
collaboration and advice: Walter Kintsch, Bob Rehder, Mike Wolfe, & M. E. Shreiner.
We especially acknowledge two participants from the Spring 1997 LSA seminar at CU
whose unpublished work is described: Alan Sanfey (3.3.4) and Michael Emerson (3.3.1).
Thanks also to Susan T. Dumais of Bellcore.
Learning Human Knowledge by Singular Value Decomposition: A Progress Repol1
51
References
Anglin, J. M. (1970). The growth of word meaning. Cambridge, MA: MIT.
Berry, M. W. (1992). Large scale singular value computations. International Journal of
Supercomputer Applications, 6, 13-49.
Deerwester, S., Dumais, S. T., Furnas, G. W., Landauer, T. K., & Harshman, R. (1990).
Indexing By Latent Semantic Analysis. Journal of the American Society For
Information Science. 41, 391-407.
Dumais, S. T. (1991). Improving the retrieval of infonnation from external sources.
Behavior Research Methods, Instruments and Computers. 23,229-236.
Dumais, S. T. (1994). Latent semantic indexing (LSI) and TREC-2. In D. Harman (Ed.),
National Institute of Standards and Technology Text Retrieval Conference. NIST
special publication.
Foltz, P. W., Kintsch, W., & Landauer, T. K. (in press). Analysis of text coherence
using Latent Semantic Analysis. Discourse Processes.
Laham, D. (in press). Latent Semantic Analysis approaches to categorization. Proceedings
of the Cognitive Science Society. 1997.
Landauer, T. K., & Dumais, S. T. (1997). A solution to Plato's problem: The Latent
Semantic Analysis theory of the acquisition, induction, and representation of
knowledge. Psychological Review, 104,211-240.
Landauer, T. K., Foltz, P. W., & Laham, D. (1997). Latent Semantic Analysis passes
the test: knowledge representation and multiple-choice testing. Manuscript in
preparation.
Landauer, T. K., Laham, D., Rehder, B. & Schreiner, M .E. (in press). How well can
passage meaning be derived without using word order: A comparison of Latent
Semantic Analysis and humans. Proceedings of the Cognitive Science Society, 1997.
Landauer, T. K., & Littman, M. L. (1990). Fully automatic cross-language document
retrieval using latent semantic indexing. In Proceedings of the Sixth Annual
Conference of the UW Centre for the New Oxford English Dictionary and Text
Research (pp. 31-38). Waterloo, Ontario: UW Centre for the New OED.
Myers, D. G. (1995). Psychology. Fourth Edition. NY, NY: Worth.
Rehder, B., Schreiner, M. E., Wolfe, B. W., Laham, D., Landauer, T. K., & Kintsch, W.
(in press). Using Latent Semantic Analysis to assess knowledge: Some technical
considerations. Discourse Processes.
ShafIf, E., Smith, E. E., & Osherson, D. N. (1990). Typicality and reasoning judgments.
Memory & Cognition, 3, 229-239.
Till, R. E., Mross, E. F., & Kintsch. W. (1988). Time course of priming for associate
and inference words in discourse context. Memory and Cognition, 16, 283-299.
Tversky, A., & Kahneman, D. (1974). Judgment under uncertainty: Heuristics and biases.
Science, 18S, 1124-1131.
Tversky, A., & Kahneman, D. (1980). Judgments of and by representativeness. In D.
Kahneman, P. Slovic, & A. Tversky (Eds.), Judgment under uncertainty: Heuristics
and biases. New York: Cambridge University Press.
Tversky, A., & Kahneman, D. (1983). Extensional versus intuitive reasoning: The
conjunction fallacy in probability judgment. Psychological Review, 90,293-315.
Warrington, E. K., & Shallice, T. (1984). Category-specific semantic impairments.
Brain, 107, 829-853.
Wolfe, M. B., Schreiner, M. E., Rehder, B., Laham, D., Foltz, P. W., Kintsch, W., &
Landauer, T. K. (in press). Learning from text: Matching readers and text by Latent
Semantic Analysis. Discourse Processes.
| 1468 |@word kintsch:7 briefly:1 cu:1 justice:1 essay:24 simulation:2 decomposition:7 moment:1 reduction:1 contains:1 score:8 document:3 past:1 contextual:1 protection:1 activation:1 yet:1 assigning:2 written:1 extensional:1 interpretable:1 discrimination:1 half:1 intelligence:1 item:2 affair:1 rehder:6 smith:2 short:3 node:1 successive:2 five:2 admission:1 along:1 constructed:2 direct:1 differential:1 ik:2 qualitative:1 prove:1 introductory:2 paragraph:3 manner:2 inter:1 indeed:1 behavior:3 themselves:1 frequently:2 brain:2 automatically:1 xx:2 unrelated:1 underlying:1 mass:1 linda:3 kind:4 psych:1 string:2 textbook:1 axl:2 growth:1 uk:2 shafir:1 unit:1 underlie:1 lsa:40 appear:1 producing:1 harshman:1 educated:1 service:2 engineering:1 local:3 limit:1 oxford:1 approximately:1 co:6 factorization:1 limited:1 statistically:1 practical:1 acknowledgment:1 testing:3 lost:1 differs:1 emerson:1 got:1 significantly:1 matching:1 word:52 induce:2 pre:1 instructor:1 chomsky:1 symbolic:2 cannot:3 close:1 wad:1 judged:2 unsigned:1 context:5 restriction:1 equivalent:2 demonstrated:1 educational:3 independently:1 typicality:1 schreiner:4 insight:1 orthonormal:1 deriving:1 embedding:1 traditionally:1 constructor:1 laham:12 analogous:1 unconscious:1 target:3 colorado:3 user:1 us:1 hypothesis:1 agreement:1 associate:2 element:1 wolfe:4 predicts:1 mike:1 solved:1 capture:1 susan:1 valuable:1 deeply:1 asked:3 littman:2 oed:1 tversky:6 trained:5 myriad:1 serve:1 efficiency:1 basis:1 translated:1 kahneman:7 easily:1 joint:1 osherson:2 represented:5 various:2 america:1 walter:1 forced:1 describe:1 query:2 herpes:1 choosing:1 whose:4 heuristic:2 larger:2 posed:1 valued:1 otherwise:1 objectively:1 ability:2 statistic:1 wont:2 syntactic:1 jointly:1 associative:1 obviously:1 myers:2 net:2 essa:1 took:1 product:1 fr:1 relevant:2 aligned:1 till:2 ontario:1 representational:2 description:4 intuitive:1 constituent:1 regularity:1 darrell:1 p:1 produce:3 categorization:1 perfect:1 object:2 exam:2 measured:4 ij:3 school:1 wished:1 progress:4 strong:1 predicted:1 judge:2 beg:1 implies:1 anatomy:1 emulation:1 correct:1 human:33 material:1 argued:1 crux:1 assign:1 anglin:2 around:1 considered:1 presumably:1 cognition:3 predict:2 dictionary:2 early:1 purpose:1 estimation:1 bag:2 infonnation:2 waterloo:1 largest:2 successfully:1 weighted:1 unfolding:1 mit:1 rather:5 varying:1 publication:1 conjunction:4 linguistic:1 derived:2 likelihood:2 inference:3 foreign:1 typically:1 hidden:2 relation:14 transformed:1 overall:3 among:6 classification:1 issue:1 bellcore:1 special:1 mutual:2 equal:1 identical:2 represents:1 unsupervised:1 constitutes:1 nearly:1 simplex:1 report:3 others:2 fundamentally:1 inherent:1 serious:1 primarily:1 simultaneously:1 national:1 individual:3 intended:1 ourselves:1 consisting:1 highly:1 evaluation:1 accurate:1 capable:1 experience:2 desired:1 causal:1 theoretical:1 psychological:3 increased:1 column:2 earlier:1 boolean:1 cover:1 phrase:4 entry:1 subset:2 imperfectly:1 successful:1 fruitfully:1 too:1 reported:2 answer:3 accomplish:1 dumais:9 combined:2 calibrated:1 slovic:1 person:2 fundamental:1 thanks:1 international:1 shortfall:1 standing:1 told:1 michael:1 together:1 again:1 reflect:1 nm:1 satisfied:1 containing:1 choose:2 possibly:1 woman:1 sentential:1 cognitive:4 external:1 expert:6 american:1 style:1 account:1 potential:1 diversity:1 coding:2 student:18 representativeness:1 depends:2 ad:3 performed:1 try:1 view:1 picked:1 apparently:1 start:1 relied:1 sort:3 participant:1 ass:4 square:1 who:2 characteristic:1 judgment:9 accurately:2 produced:1 populating:1 mere:1 worth:1 bob:1 history:1 reach:1 ed:2 sixth:1 acquisition:2 frequency:3 pp:1 resultant:1 associated:1 gain:1 newly:1 fre:1 sampled:2 popular:1 ask:1 logical:2 knowledge:35 dimensionality:2 carefully:1 sophisticated:2 manuscript:1 exceeded:1 reflected:4 strongly:1 just:2 implicit:2 correlation:4 hand:2 warrington:2 lack:2 quality:5 artifact:1 scientific:1 hypothesized:2 contain:3 concept:2 validity:1 true:1 harman:1 assigned:3 read:2 semantic:19 freq:1 illustrated:1 mel:1 cosine:8 syntax:3 complete:2 passage:16 reasoning:2 meaning:15 consideration:1 discovers:1 recently:1 million:1 association:1 occurred:1 analog:1 interpret:1 significant:1 cambridge:2 ai:1 automatic:2 newswire:1 centre:2 language:8 had:2 access:1 calibration:1 similarity:23 subjectively:1 dominant:1 showed:1 perspective:1 certain:1 success:1 accomplished:2 scoring:2 seen:1 additional:3 greater:1 redundant:1 preservation:1 multiple:5 full:1 alan:1 technical:1 faster:1 match:1 offer:2 long:1 retrieval:5 cross:1 paired:1 underlies:1 basic:2 regression:1 represent:4 cell:2 preserved:1 objection:1 singular:10 source:6 country:1 comprehensibility:1 probably:1 pass:1 induced:1 subject:1 tend:1 simulates:1 thing:1 member:4 plato:1 grammatically:1 adequacy:1 extracting:1 unitary:1 near:1 structural:1 revealed:1 easy:1 superset:2 concerned:1 variety:3 psychology:5 gave:1 identified:2 idea:2 whether:1 expression:2 utility:2 peter:1 speech:1 speaking:1 passing:1 york:1 impairment:1 useful:2 listed:1 amount:3 encyclopedia:2 ten:3 induces:1 category:4 generate:1 lsi:1 canal:1 correctly:1 write:2 discrete:1 express:3 group:1 four:5 prevent:1 uw:2 vast:1 merely:2 deerwester:2 angle:3 letter:1 fourth:1 uncertainty:2 striking:1 place:1 reader:7 decide:1 thesaurus:1 coherence:2 acceptable:1 comparable:2 layer:2 correspondence:2 annual:1 constraint:5 deficiency:1 unlimited:1 aspect:1 simulate:1 spring:1 relatively:2 department:2 combination:2 slightly:1 agreeing:1 presently:1 indexing:3 boulder:2 instructional:2 heart:2 taken:2 previously:3 mechanism:2 know:1 subjected:1 instrument:1 appropriate:1 chamber:1 simulating:2 occurrence:4 alternative:3 professional:1 supercomputer:1 thomas:1 original:2 exploit:1 especially:1 society:3 move:1 objective:2 question:5 quantity:4 occurs:1 damage:1 diagonal:1 guessing:1 traditional:1 subspace:1 separate:1 thank:1 simulated:1 accommodates:1 topic:4 reason:1 induction:3 length:4 retained:2 relationship:2 insufficient:1 kk:2 acquire:1 mexico:1 ftrst:1 potentially:1 statement:2 design:1 shallice:2 allowing:1 observation:1 markov:1 finite:1 melded:1 acknowledge:1 nist:1 trec:1 inferred:1 rating:2 pair:3 unpublished:1 connection:1 sentence:8 learned:2 textual:1 established:1 adult:4 able:2 usually:1 panama:1 max:1 including:2 reciprocally:1 memory:2 event:1 satisfaction:2 natural:7 rely:3 treated:1 representing:1 improve:1 rated:3 technology:1 extract:4 utterance:1 text:30 review:2 teller:2 berry:2 multiplication:1 occupation:2 fully:1 limitation:2 versus:1 conjoint:1 degree:1 conveyed:3 sufficient:2 article:1 principle:1 bank:2 collaboration:1 row:1 course:3 summary:1 english:3 verbal:1 bias:2 foltz:9 institute:2 tolerance:1 grammatical:1 dimension:6 vocabulary:4 transition:2 stand:1 axi:1 made:3 commonly:1 replicated:1 far:1 social:1 correlate:1 approximate:1 preferred:1 logic:1 abstracted:2 global:1 corpus:4 conceptual:3 landauer:20 accum:1 continuous:2 latent:11 fallacy:2 triplet:1 learn:2 improving:1 bearing:1 priming:1 domain:4 synonym:3 scored:1 edition:1 child:2 convey:2 body:2 positively:1 advice:1 representative:2 ny:2 furnas:1 seminar:1 comprises:1 explicit:1 candidate:1 young:1 resembled:1 specific:2 showing:1 supplemented:1 agnosia:1 explored:1 list:1 multitude:1 evidence:2 adding:1 mirror:1 associational:1 antonym:3 commutative:1 simply:1 likely:1 expressed:2 contained:3 acquiring:1 discourse:4 extracted:2 ma:1 viewed:1 sized:1 sorted:1 crl:1 content:4 included:1 determined:1 typical:1 principal:1 called:2 total:2 svd:4 la:1 experimental:1 meaningful:1 indicating:1 college:1 highdimensional:1 people:5 latter:1 inability:1 preparation:1 evaluate:2 tested:3 phenomenon:1 correlated:1 |
512 | 1,469 | A Superadditive-Impairment Theory
of Optic Aphasia
Michael C. Mozer
Dept. of Computer Science
University of Colorado
Boulder; CO 80309-0430
Mark Sitton
Dept. of Computer Science
University of Colorado
Boulder; CO 80309-0430
Martha Farah
Dept. of Psychology
University of Pennsylvania
Phila., PA 19104-6196
Abstract
Accounts of neurological disorders often posit damage to a specific
functional pathway of the brain. Farah (1990) has proposed an alternative class of explanations involving partial damage to multiple pathways. We explore this explanation for optic aphasia, a disorder in which
severe perfonnance deficits are observed when patients are asked to
name visually presented objects, but surprisingly, performance is relatively nonnal on naming objects from auditory cues and on gesturing
the appropriate use of visually presented objects. We model this highly
specific deficit through partial damage to two pathways-one that maps
visual input to semantics, and the other that maps semantics to naming
responses. The effect of this damage is superadditive, meaning that
tasks which require one pathway or the other show little or no performance deficit, but the damage is manifested when a task requires both
pathways (i.e., naming visually presented objects). Our model explains
other phenomena associated with optic aphasia, and makes testable
experimental predictions.
Neuropsychology is the study of disrupted cognition resulting from damage to functional
systems in the brain. Generally, accounts of neuropsychological disorders posit damage to
a particular functional system or a disconnection between systems. Farah (1990) suggested an alternative class of explanations for neuropsychological disorders: partial damage to multiple systems, which is manifested through interactions among the loci of
damage. We explore this explanation for the neuropsychological disorder of optic aphasia.
Optic aphasia, arising from unilateral left posterior lesions, including occipital cortex
and the splenium of the corpus callosum (Schnider, Benson, & Scharre, 1994), is marked
by a deficit in naming visually presented objects, hereafter referred to as visual naming
(Farah, 1990). However, patients can demonstrate recognition of visually presented
objects nonverbally, for example, by gesturing the appropriate use of an object or sorting
visual items into their proper superordinate categories (hereafter, visual gesturing).
Patients can also name objects by nonvisual cues such as a verbal definition or typical
sounds made by the objects (hereafter, auditory naming). The highly specific nature of the
deficit rules out an explanation in terms of damage to a single pathway in a standard model
of visual naming (Figure 1), suggesting that a more complex model is required, involving
A Superadditive-Impainnent Theory of Optic Aphasia
FIGURE 1. A standard box-and-arrow
model of visual naming. The boxes denote
levels of representation, and the arrows
denote pathways mapping from one level of
representation to another. Although optic
aphasia cannot be explained by damage to
the vision-to-semantics pathway or the
semantics-ta-naming
pathway,
Farah
(1990) proposed an explanation in terms of
partial damage to both pathways (the X's).
67
visual
auditory
mUltiple semantic systems or multiple pathways to visual naming. However, a mere parsimonious account is suggested by Farah (1990): Optic aphasia might arise from partial
lesions to two pathways in the standard model-those connecting visual input to semantics, and semantics to naming-and the effect of damage to these pathways is superadditive, meaning that tasks which require only one of these pathways (e.g., visual gesturing,
or auditory naming) will be relatively unimpaired, whereas tasks requiring both pathways
(e.g., visual naming) will show a significant deficit.
1 A MODEL OF SUPERADDITIVE IMPAIRMENTS
We present a computational model of the superadditive-impairment theory of optic aphasia by elaborating the architecture of Figure 1. The architecture has four pathways: visual
input to semantics (V~S), auditory input to semantics (A~S), semantics to naming
(S~N), and semantics to gesturing (S~G). Each pathway acts as an associative memory.
The critical property of a pathway that is required to explain optic aphasia is a speed-accuracy trade off: The initial output of a pathway appears rapidly, but it may be inaccurate.
This "quick and dirty" guess is refined over time, and the pathway output asymptotically
converges on the best interpretation of the input.
We implement a pathway using the architecture suggested by Mathis and Mozer
(1996). In this architecture, inputs are mapped to their best interpretations by means of a
two-stage process (Figure 2). First, a quick, one-shot mapping is performed by a multilayer feedforward connectionist network to transform the input directly to its corresponding output. This is followed by a slower iterative clean-up process carried out by a
recurrent attractor network. This architecture shows a speed-accuracy trade off by virtue
of the .assumption that the feed forward mapping network does not have the capacity to
produce exactly the right output to every input, especially when the inputs are corrupted
by noise or are otherwise incomplete. Consequently, the clean up stage is required to produce a sensible interpretation of the noisy output of the mapping network.
Fully distributed attractor networks have been used for similar purposes (e.g., Plaut
& Shallice, 1993). For simplicity, we adopt a localist-attractor network with a layer of
state units and a layer of radial basis function (RBF) units, one RBF unit per attractor.
Each RBF or attractor unit measures the distance of the current state to the attractor that it
represents. The activity of attractor unit i, aj, is:
FIGURE 2. Connectionist implementation of a processing pathway. The pathway consists of feedforward mapping
network followed by a recurrent cleanup or attractor network. Circles denote
connectionist processing units and
arrows denote connections between
units or between layers of units.
clean up network
mapping network
M. C. Mozer, M. Sitton and M. Farah
68
(1)
(2)
where set) is the state unit activity vector at time t, ~i is the vector denoting the location of
attractor i, and ~i is the strength of the attractor. The strength detennines the region of the
state space over which an attractor will exert its pull, and also the rate at which the state
will converge to the attractor. The state units receive input from the mapping network and
from the attractor units and are updated as follows:
s;(t)
= d;(t)e;(t) + (I -
d;(t? La/t -I )~j;
(3)
j
where sift) is the activity of state unit i at time t, e; is the ith output of the mapping net,
is the ith element of attractor j, and d; is given by
d .(t)
I
= h[I __
e;(_t-..,.-_l)]
ej(t)
(4)
ej(t)
where h[.] is a linear threshold function that bounds activity between -1 and +1,
weighted time average of the ith output of the mapping net,
= cxej(t) + (I -CX)e;(t-I)
~j;
ej
IS
a
(5)
In all simulations, cx = .02.
The activity of the state units are governed by two forces-the external input from
the feedforward net (first tenn in Equation 3) and the attractor unit activities (second
tenn). The parameter d; acts as a kind of attentional mechanism that modulates the relative
influence of these two forces. The basic idea is that when the input coming from the mapping net is changing, the system should be responsive to the input and should not yet be
concerned with interpreting the input. In this case, the input is copied straight through to
the state units and hence dj should have a value close to I. When the input begins to stabilize, however, the focus shifts to interpreting the input and following the dynamics of the
attractor network. This shift corresponds to d j being lowered to zero. The weighted time
average in the update rule for d j is what allows for the smooth transition of the function to
its new value. For certain constructions of the function d, Zemel and Mozer (in preparation) have proven convergence of the algorithm to an attractor.
Apart from speed-accuracy trade off, these dynamics have another important consequence for the present model, particularly with respect to cascading pathways. If pathway
A feeds into pathway S, such as V~S feeding into S~N, then the state unit activities of A
act as the input to S. Because these activities change over time as the state approaches a
well-fonned state, the dynamics of pathway S can be quite complex as it is forced to deal
with an unstable input. This property is important in explaining several phenomena associated with optic aphasia.
1.1 PATTERN GENERATION
Patterns were constructed for each of the five representational spaces: visual and auditory
input, semantic, name and gesture responses. Each representational space was arbitrarily
made to be 200 dimensional. We generated 200 binary-valued (-1,+1) patterns in each
space, which were meant to correspond to known entities of that representational domain.
For the visual, auditory, and semantic spaces, patterns were partitioned into 50 similarity clusters with 4 ?siblings per cluster. Patterns were chosen randomly subject to two
constraints: patterns in different clusters had to be at least 80? apart, and siblings had to be
between 25? and 50? apart. Because similarity of patterns in the name and gesture spaces
was irrelevant to our modeling, we did not impose a similarity structure on these spaces.
A Superadditive-Impainnent Theory of Optic Aphasia
69
Instead, we generated patterns in these spaces at random subject to the constraint that
every pattern had to be at least 60? from every other.
After generating patterns in each of the representational spaces, we established arbitrary correspondences among the patterns such that visual pattern n. auditory pattern n,
semantic pattern n, name pattern n, and gesture pattern n all represented the same concept.
That is, the appropriate response in a visual-naming task to visual pattern n would be
semantic pattern n and name pattern n.
1.2 TRAINING PROCEDURE
The feedforward networks in the four pathways (V-"7S, A-"7S, S-"7N, and S-"70) were
independently trained on all 200 associations using back propagation. Each of these networks contained a single hidden layer of 150 units, and all units in the network used the
symmetric activation function to give activities in the range [-1,+1]. The amount of training was chosen such that performance on the training examples was not perfect; usually
several elements in the output would be erroneous-i.e., have the wrong sign-and others
would not be exactly correct-i.e., -lor + 1. This was done to embody the architectural
assumption that the feedforward net does not have the capacity to map every input to
exactly the right output, and hence, the clean-up process is required.
Training was not required for the clean-up network. Due to the localist representation
of attractors in the clean-up network, it was trivial to hand wire each clean-up net with the
200 attractors for its domain, along with one rest-state attractor. All attractor strengths
were initialized to the same value, ~= 15, except the rest-state attractor, for which P=5. The
rest-state attractor required a lower strength so that even a weak external input would be
sufficient to kick the attractor network out of the rest state.
1.3 SIMULATION METHODOLOGY
After each pathway had been trained, the model was damaged by "lesioning" or
removing a fraction y of the connections in the V -"7S and S-"7N mapping networks. The
lesioned connections were chosen at random and an equal fraction was removed from the
two pathways. The clean-up nets were not damaged. The architecture was damaged a total
of 30 different times, creating 30 simulated patients who were tested on each of the four
tasks and on all 200 input patterns for a task. The results we report come from averaging
across simulated patients and input patterns. Responses were detennined after the system
had been given sufficient time to relax into a name or gesture attractor, which was taken to
be the response. Each response was classified as one of the following mutually exclusive
response types: correct, perseveration (response is the same as that produced on any of the
three immediately preceding trials), visual error (the visual pattern corresponding to the
incorrect response is a sibling of the visual pattern corresponding to the correct response),
semantic error, visual+semantic error, or other error.
1.4 PRIMING MECHANISM
Priming-the increased availability of recently experienced stimuli-has been found
across a wide variety of tasks in normal subjects. We included priming in our model as a
strengthening (increasing the Pi parameter) of recently visited attractors (see Mathis &
Mozer 1996, for details, and Becker, Behrmann, & Moscovitch, 1993, for a related
approach). In the damaged model, this mechanism often gave rise to perseverations.
2 RESULTS
We have examined the model's behavior as we varied the amount of damage, quantified by
the parameter y. However, we report on the perfonnance of simulated patients with y = .30.
This intermediate amount of damage yielded no floor or ceiling effects, and also produced
error rates for the visual-naming task in the range of 30-40%, roughly the median performance of patients in the literature.
70
TABLE 1. Error rate of the
damaged model on various
tasks.
M. C. Mozer, M. Sitton and M. Farah
task
aUOltory gesturing
auditory naming
visual gesturing
visual naming
error rate
U.U"'/o
0.5%
8.7%
36.8%
Table 1 presents the error rates of the model on four tasks. The pattern of errors
shows a qualitative fit to human patient data. The model produced no errors on the auditory gesturing task because the two component pathways (A~S and S~G) were undamaged. Relatively few errors were made on the auditory-naming and visual-gesturing tasks,
each of which involved one damaged pathway, because the clean-up nets were able to
compensate for the damage. However, the error rate for the visual-naming task was quite
large, due t'O damage on both of its component pathways (V~S and S~N). The error rate
for visual naming cannot be accounted for by summing the effects of the damage to the
two component pathways because the sum of the error rates for auditory naming and
visual gesturing, each of which involves one of the two partially damaged pathways, is
nearly four times smaller. Rather, the effects of damage on these pathways interact, and
their interaction leads to superadditive impairments.
When a visual pattern is presented to the model, it is mapped by the damaged V~S
pathway into a corrupted semantic representation which is then cleaned up. While the corruption is sufficiently minor that clean up will eventually succeed, the clean up process is
slowed considerably by the corruption. During the period of time in which the semantic
clean-up network is searching for the correct attractor, the corrupted semantic representation is nonetheless fed into the damaged S~N pathway. The combined effect of the (initially) noisy semantic representation serving as input to a damaged pathway leads to
corruption of the naming representation past the point where it can be cleaned-up properly.
Interactions in the architecture are inevitable, and are not merely a consequence of
some arbitrary assumption that is built into our model. To argue this point, we consider
two modifications to the architecture that might eliminate the interaction in the damaged
model. First, if we allowed the V~S pathway to relax into a well-formed state before
feeding its output into the S~N pathway, there would be little interaction-the effects of
the damage would be additive. However, cortical pathways do not operate sequentially,
one stage finishing its computation and then turning on the next stage. Moreover, in the
undamaged brain, such a processing strategy is unadaptive, as cascading partial results
from one pathway to the next can speed processing without the introduction of errors
(McClelland, 1979). Second, the interaction might be eliminated by making the S~N
pathway continually responsive to changes in the output of the V ~S pathway. Then, the
rate of convergence of the V ~S pathway would be irrelevant to determining the eventual
output of the S~N pathway. However, because the output of the S~N pathway depends
not only on its input but its internal state (the state of the clean-up net), one cannot design
a pathway that is continually responsive to changes in the input and is also able to clean up
noisy responses. Thus, the two modifications one might consider to eliminate the interactions in the damaged model seriously weaken the computational power of the undamaged
model. We therefore conclude that the framework of our model makes it difficult to avoid
an interaction of damage in two pathways.
A subtle yet significant aspect of the model's performance is that the error rate on the
visual-gesturing task was reliably higher than the error rate on the auditory-naming task,
despite the fact that each task made use of one damaged pathway, and the pathways were
damaged to the same degree. The difference in performance is due to the fact that the damaged pathway for the visual-gesturing task is the first in a cascade of two, while the damaged pathway for the auditory-naming task is the second. The initially noisy response
from a damaged pathway early in the system propagates to subsequent pathways, and
A Superadditive-Impairment Theory of Optic Aphasia
71
although the damaged pathway will eventually produce the correct response, this is not
sufficient to ensure that subsequent pathways will do so as well.
2.1 DISTRIBUTION OF ERRORS FOR VISUAL OBJECT NAMING
Figure 2 presents the model's error distribution for the visual-naming task. Consistent with the patient data (Farah, 1990), the model produces many more semantic and perseveration errors than by chance. The chance error proportions were computed by
assuming that if the correct response was not made, then all other responses had an equal
probability of being chosen.
To understand the predominance of semantic errors, consider the effect of damage to
the V ~S pathway. For relatively small amounts of damage, the mapping produced will be
close to the correct mapping. "Close" here means that the Euclidean distance in the
semantic output space between the correct and perturbed mapping is small. Most of the
time, minor perturbation of the mapping will be compensated for by the clean-up net.
Occasionally, the perturbation will land the model in a different attractor basin, and a different response will be made. However, when the wrong attractor is selected, it will be one
"close" to the correct attractor, i.e., it will likely be a sibling in the same pattern cluster as
the correct attractor. In the case of the V ~S pathway, the siblings of the correct attract or
are by definition semantically related. A semantic error will be produced by the model
when a sibling semantic attractor is chosen, and then this pattern is correctly mapped to a
naming response in the S~N pathway.
In addition to semantic errors, the other frequent error type in visual naming is perseverations. The priming mechanism is responsible for the significant number of perseverations, although in the unlesioned model, it facilitates processing of repeated stimuli
without producing perseverations.
Just as important as the presence of perseverative and semantic errors is the absence
of visual errors, a feature of optic aphasia that contrasts sharply with visual agnosia
(Farah, 1990). The same mechanisms explain why the rate of visual errors is close to its
chance value and why visual+semantic errors are above chance. Visual-naming errors
occur because there is an error either in the V~S or S~N mappings, or both. Since the
erroneous outputs of these pathways show a strong tendency to be similar to the correct
output, and because semantic and name similarity does not imply visual similarity (the
patterns were paired randomly), visual errors should only occur by chance. When a visual
error does occur, though, there is a high probability that the error is also semantic because
of the strong bias that already exists toward producing semantic errors. This is the reason
why more visual+semantic errors occur than by chance and why the proportion of these
?
actual
?
cI1ance
FIGURE 3. Distribution
of error types made by
model on the V~N task
(black bars) relative to
chance (grey bars).
visual
vis+sem
Error type
pe~Yendwe
72
M. C. Mozer, M. Sitton and M Farah
errors is only slightly less than the proportion of visual errors.
Plaut and Shallice (1993) have proposed a connectionist model to account for the
distribution of errors made by optic aphasics. Although their model was not designed to
account for any of the other phenomena associated with the disorder, it has much in common with the model we are proposing. Unlike our model, however, theirs requires the
assumption that visually similar objects also share semantic similarity. This assumption
might be questioned, especially because our model does not require this assumption to
produce the correct distribution of error responses.
3 DISCUSSION
In demonstrating superadditive effects of damage, we have offered an account of optic
aphasia that explains the primary phenomenon: severe impairments in visual naming in
conjunction with relatively spared performance on naming from verbal description or gesturing the appropriate use of a visually presented object. The model also explains the distribution of errors on visual naming. Although we did not have the space in this brief
report to elaborate, the model accounts for several other distinct characteristics of optic
aphasia, including the tendency of patients to "home in" on the correct name for a visually
presented object when given sufficient time, and a positive correlation between the error
rates on naming and gesturing responses to a visual object (Sitton, Mozer, & Farah, 1998).
Further, the model makes several strong predictions which have yet to be tested experimentally. One such prediction, which was apparent in the results presented earlier, is that a
higher error rate should be observed on visual gesturing than on auditory naming when the
tasks are equated for difficulty, as our simulation does.
More generally, we have strengthened the plausibility of Farah's (1990) hypothesis
that partial damage to two processing pathways may result in close-to-normal performance on tasks involving one pathway or the other while yielding a severe performance
deficit on tasks involving both damaged pathways. The superadditive-impairment theory
thus may provide a more parsimonious account of various disorders that were previously
believed to require more complex architectures or explanations.
4 ACKNOWLEDGMENTS
This research was supported by grant 97-18 from the McDonnell-Pew Program in Cognitive Neuroscience.
5 REFERENCES
Becker, S., Behrmann, M., & Moscovitch, K. (1993). Word priming in attractor networks.
Proceedings of the Fifteenth Annual Conference of the Cognitive Science Society (pp.
231-236). Hillsdale, NJ: Erlbaum.
Farah, M. J. (1990). Visual agnosia. Cambridge, MA: MIT PresslBradford Books.
Mathis, D. W., & Mozer, M. C. (1996). Conscious and unconscious perception: A computational theory. In G. Cottrell (Ed.), Proceedings of the Eighteenth Annual Conference
of the Cognitive Science Society (pp. 324-328). Hillsdale, NJ: Erlbaum.
McClelland, J. L. (1979). On the time relations of mental processes: An examination of
systems of processes in cascade. Psychological Review, 86, 287-330.
Plaut, D., & Shallice, T. (1993). Perseverative and semantic influences on visual object
naming errors in optic aphasia: A connectionist approach. Journal of Cognitive Neuroscience, 5,89-112.
Schnider, A., Benson, D. E, and Scharre, D. W. (1994). Visual agnosia and optic aphasia:
Are they anatomically distinct? Cortex, 30, 445-457.
Sitton, M., Mozer, M. C., & Farah, M. (1998). Diffuse lesions in a modular connectionisl
architecture: An account of optic aphasia. Manuscript submitted for publication.
| 1469 |@word trial:1 proportion:3 grey:1 simulation:3 shot:1 initial:1 hereafter:3 denoting:1 seriously:1 elaborating:1 past:1 current:1 activation:1 yet:3 cottrell:1 additive:1 subsequent:2 designed:1 update:1 cue:2 tenn:2 guess:1 item:1 selected:1 ith:3 mental:1 plaut:3 location:1 five:1 lor:1 along:1 constructed:1 incorrect:1 consists:1 qualitative:1 pathway:65 roughly:1 behavior:1 embody:1 brain:3 little:2 actual:1 increasing:1 begin:1 moreover:1 what:1 kind:1 proposing:1 nj:2 every:4 act:3 exactly:3 wrong:2 unit:18 grant:1 producing:2 continually:2 before:1 positive:1 consequence:2 despite:1 might:5 black:1 exert:1 examined:1 quantified:1 co:2 range:2 neuropsychological:3 acknowledgment:1 responsible:1 implement:1 procedure:1 cascade:2 word:1 radial:1 lesioning:1 cannot:3 close:6 influence:2 map:3 quick:2 compensated:1 eighteenth:1 occipital:1 independently:1 simplicity:1 disorder:7 immediately:1 rule:2 undamaged:3 cascading:2 pull:1 searching:1 updated:1 construction:1 unconscious:1 colorado:2 damaged:19 hypothesis:1 pa:1 element:2 recognition:1 particularly:1 observed:2 region:1 trade:3 removed:1 neuropsychology:1 mozer:10 asked:1 lesioned:1 dynamic:3 trained:2 basis:1 represented:1 various:2 forced:1 distinct:2 zemel:1 refined:1 quite:2 apparent:1 modular:1 valued:1 relax:2 otherwise:1 transform:1 noisy:4 associative:1 net:10 interaction:8 coming:1 strengthening:1 frequent:1 rapidly:1 detennined:1 representational:4 description:1 convergence:2 cluster:4 produce:5 generating:1 perfect:1 converges:1 object:15 recurrent:2 minor:2 strong:3 involves:1 come:1 posit:2 detennines:1 correct:14 human:1 hillsdale:2 explains:3 require:4 feeding:2 sufficiently:1 normal:2 visually:8 cognition:1 mapping:16 adopt:1 early:1 purpose:1 visited:1 predominance:1 callosum:1 weighted:2 mit:1 rather:1 avoid:1 ej:3 publication:1 conjunction:1 focus:1 finishing:1 properly:1 contrast:1 spared:1 attract:1 inaccurate:1 eliminate:2 initially:2 hidden:1 relation:1 semantics:10 nonnal:1 among:2 equal:2 eliminated:1 represents:1 nearly:1 inevitable:1 connectionist:5 others:1 report:3 stimulus:2 few:1 randomly:2 attractor:33 highly:2 superadditive:11 severe:3 yielding:1 partial:7 perfonnance:2 incomplete:1 euclidean:1 initialized:1 circle:1 weaken:1 psychological:1 increased:1 modeling:1 earlier:1 localist:2 erlbaum:2 perturbed:1 corrupted:3 considerably:1 combined:1 disrupted:1 off:3 michael:1 connecting:1 external:2 creating:1 cognitive:4 book:1 account:9 suggesting:1 stabilize:1 availability:1 depends:1 vi:1 performed:1 farah:15 formed:1 accuracy:3 who:1 characteristic:1 correspond:1 weak:1 produced:5 mere:1 perseveration:6 corruption:3 straight:1 classified:1 submitted:1 explain:2 ed:1 definition:2 nonetheless:1 pp:2 involved:1 associated:3 auditory:15 subtle:1 back:1 appears:1 feed:2 manuscript:1 ta:1 higher:2 methodology:1 response:19 done:1 box:2 though:1 just:1 stage:4 correlation:1 hand:1 propagation:1 aj:1 name:9 effect:9 requiring:1 concept:1 hence:2 symmetric:1 semantic:25 deal:1 during:1 demonstrate:1 interpreting:2 meaning:2 recently:2 common:1 functional:3 association:1 interpretation:3 theirs:1 significant:3 cambridge:1 pew:1 dj:1 unimpaired:1 had:6 lowered:1 cortex:2 similarity:6 posterior:1 irrelevant:2 apart:3 occasionally:1 certain:1 manifested:2 binary:1 arbitrarily:1 impose:1 preceding:1 floor:1 converge:1 period:1 multiple:4 sound:1 smooth:1 gesture:4 plausibility:1 compensate:1 believed:1 naming:36 paired:1 prediction:3 involving:4 basic:1 multilayer:1 vision:1 patient:10 fifteenth:1 receive:1 whereas:1 addition:1 median:1 rest:4 operate:1 unlike:1 subject:3 facilitates:1 kick:1 presence:1 feedforward:5 intermediate:1 concerned:1 variety:1 fit:1 psychology:1 gave:1 pennsylvania:1 architecture:10 idea:1 sibling:6 shift:2 becker:2 unilateral:1 questioned:1 impairment:7 generally:2 amount:4 conscious:1 category:1 mcclelland:2 sign:1 neuroscience:2 arising:1 per:2 correctly:1 serving:1 four:5 threshold:1 demonstrating:1 changing:1 clean:15 asymptotically:1 merely:1 fraction:2 sum:1 architectural:1 parsimonious:2 home:1 bound:1 layer:4 followed:2 perseverative:2 copied:1 correspondence:1 presslbradford:1 yielded:1 activity:9 annual:2 strength:4 occur:4 optic:20 constraint:2 sharply:1 diffuse:1 aspect:1 speed:4 relatively:5 mcdonnell:1 across:2 smaller:1 slightly:1 partitioned:1 modification:2 making:1 benson:2 explained:1 slowed:1 anatomically:1 boulder:2 taken:1 ceiling:1 equation:1 mutually:1 previously:1 eventually:2 mechanism:5 aphasia:19 locus:1 fed:1 appropriate:4 responsive:3 alternative:2 slower:1 dirty:1 ensure:1 testable:1 especially:2 society:2 already:1 damage:25 strategy:1 exclusive:1 primary:1 distance:2 deficit:7 mapped:3 attentional:1 capacity:2 entity:1 sensible:1 simulated:3 argue:1 unstable:1 trivial:1 toward:1 reason:1 assuming:1 difficult:1 rise:1 implementation:1 design:1 proper:1 shallice:3 reliably:1 wire:1 varied:1 perturbation:2 arbitrary:2 required:6 cleaned:2 connection:3 established:1 able:2 suggested:3 superordinate:1 usually:1 pattern:28 bar:2 perception:1 program:1 built:1 including:2 memory:1 explanation:7 power:1 critical:1 difficulty:1 force:2 examination:1 turning:1 brief:1 imply:1 carried:1 review:1 literature:1 determining:1 relative:2 fully:1 generation:1 unadaptive:1 proven:1 degree:1 offered:1 sufficient:4 consistent:1 basin:1 propagates:1 pi:1 land:1 share:1 accounted:1 surprisingly:1 supported:1 verbal:2 bias:1 disconnection:1 understand:1 explaining:1 wide:1 distributed:1 cortical:1 transition:1 forward:1 made:8 fonned:1 equated:1 sequentially:1 corpus:1 summing:1 conclude:1 iterative:1 why:4 table:2 nature:1 sem:1 interact:1 complex:3 mathis:3 priming:5 domain:2 did:2 arrow:3 noise:1 arise:1 lesion:3 allowed:1 repeated:1 referred:1 elaborate:1 strengthened:1 experienced:1 governed:1 pe:1 behrmann:2 removing:1 erroneous:2 specific:3 sift:1 agnosia:3 virtue:1 exists:1 unlesioned:1 modulates:1 sorting:1 cx:2 explore:2 likely:1 visual:52 contained:1 neurological:1 partially:1 corresponds:1 chance:7 ma:1 succeed:1 marked:1 consequently:1 rbf:3 eventual:1 absence:1 martha:1 change:3 included:1 typical:1 except:1 experimentally:1 semantically:1 averaging:1 total:1 experimental:1 la:1 tendency:2 internal:1 mark:1 moscovitch:2 cleanup:1 meant:1 preparation:1 dept:3 tested:2 phenomenon:4 |
513 | 147 | 297
A NETWORK FOR IMAGE SEGMENTATION
USING COLOR
Anya Hurlbert and Tomaso Poggio
Center for Biological Information Processing at Whitaker College
Department of Brain and Cognitive Science
and the MIT AI Laboratory
Cambridge, MA 02139
(hur lbert@wheaties.ai.mit.edu)
ABSTRACT
We propose a parallel network of simple processors to find
color boundaries irrespective of spatial changes in illumination, and to spread uniform colors within marked reglOns.
.
INTRODUCTION
To rely on color as a cue in recognizing objects, a visual system must have at least
approximate color constancy. Otherwise it might ascribe different characteristics to
the same object under different lights. But the first step in using color for recognition, segmenting the scene into regions of different colors, does not require color
constancy. In this crucial step color serves simply as a means of distinguishing
one object from another in a given scene. Color differences, which mark material
boundaries, are essential, while absolute color values are not. The goal of segmentation algorithms is to achieve this first step toward object recognition by finding
discontinuities in the image irradiance that mark material boundaries.
The problems that segmentation algorithms must solve is how to choose color labels, how to distinguish material boundaries from other changes in the image that
give rise to color edges, and how to fill in uniform regions with the appropriate
color labels. (Ideally, the color labels should remain constant under changes in the
illumination or scene composition and color edges should occur only at material
boundaries.) Rubin and Richards (1984 ) show that algorithms can solve the second problem under some conditions by comparing the image irradiance signal in
distinct spectral channels on either side of an edge.
The goal of the segmentation algorithms we discuss here is to find boundaries between regions of different surface spectral reflectances and to spread uniform colors
within them, without explicitly requiring the colors to be constant under changes
in illumination. The color labels we use are analogous to the CIE chromaticity
coordinates x and y. Under the single source assumption, they change across space
298
Hurlbert and Poggio
only when the surface spectral reflectance changes, except when strong specularities are present. (The algorithms therefore require help at a later stage to identify
between color label changes due to specularities, which we have not yet explicitly
incorporated.) The color edges themselves are localised with the help of luminance
edges, by analogy with psychophysics of segmentation and filling-in. The Koftka
Ring illusion, for example, indicates that color is attributed to surfaces by an interaction between an edge-finding operator and a filling-in operator. 1 The interaction
is justified by the fact that in the real world changes in surface spectral reflectance
are almost always accompanied by changes in brightness.
Color Labels
We assume that surfaces reflect light according to the neutral-interface-reflection
model. In this model (Lee, 1986 , Shaefer, 1984 [3]) the image irradiance I(X,y,A)
is the sum of two components, the surface reflection and the body reflection:
I(x, y, A)
=
L(r(x, y), A)[a(r, A)g(6(r)) + bh(6(r))],
where A labels wavelength and r( x, y) is the point on the 3D surface to which
the image coordinates (x, y) correspond. L(r(x, y), A) is the illumination on the
surface. a(r, A) is the spectral reflectance factor of the body reflection component
and g(6(r)) its magnitude, which depends on the viewing geometry parameters
lumped together in 6(r). The spectral reflectance factor of the specular, or surface
reflection, component b is assumed to be constant with respect to A, as is true
for inhomogeneous materials such as paints and plastics. For most materials, the
magnitude of the specular component h depends strongly on the viewing geometry.
Using the single source assumption, we may factor the illumination L into separate
spatial and spectral components (L(r, A)
L(r)c(A)). Multiplying I by the
spectral sensitivities of the color sensors i = 1,2,3 and integrating over wavelength
yields the triplet of color values (R, G, B), where
and so forth and where the a i and bi are the reflectance factors in the spectral
channels defined by the sensor spectral sensitivities.
We define the hues u and v as
R
u= - -__- -
R+G+B
and
1 Note that Land's original retinex algorithm, which thresholds and swns the differences in image
irradiance between adjacent points along many paths, accounts for the contribution of edges to
color, without introducing a separate luminance edge detector.
A Network for Image Segmentation Using Color
G
v=-----
R+G+B
at each pixel.
In Lambertian reflection, the specular reflectance factor b is zero. In this case, u and
v are piecewise constant: they change in the image only when the ai(x,y) change.
Thus u or v mark discontinuities in the surface spectral reflectance function, e.g
they mark material boundaries. Conversely, image regions of constant u correspond
to regions of constant surface color. Synthetic images generated with standard
computer graphics algorithms (using, for example, the Phong reflectance model)
behave in this way: u is constant across the visible surface of a shaded sphere.
Across specularities, u in general changes but often not much. Thus one approach
to the segmentation problem is to find regions of "constant" u and their boundaries .
The difficulty with this approach is that real u data are noisy and unreliable: u is
the quotient of numbers that are not only noisy themselves but also, at least for
biological photosensor spectral sensitivities, very close to one another. The goals of
segmentation algorithms are therefore to enhance discontinuities in u and, within
the regions marked by the discontinuities, to smoothe over the noise and fill in the
data where they are unreliable. We have explored several methods of meeting these
goals.
Segmentation Algorithms
One method is to regularize - to eliminate the noise and fill in the data, while
preserving the discontinuities. Using an algorithm based on Markov Random Field
techniques, we have obtained encouraging results on real images (see Poggio et
al., 1988) . The MRF technique exploits the constraint that u should be piecewise
constant within the discontinuity contours and uses image brightness edges as guides
in finding the contours.
An alternative to the MRF approach is a cooperative network that fills in data
and filters out noise while enforcing the constraint of piecewise constancy. The
network, a type of Hopfield net, is similar to the cooperative stereo network of
Marr and Poggio (1976). Another approach consists of a one-pass winner-take-all
scheme. Both algorithms involve loading the initial hue values into discrete bins, an
undesirable and biologically unlikely feature . Although they produce good results
on noisy synthetic images and can be improved by modification (see Hurlbert, 1989),
another class of algorithms which we now describe are simple and effective, especially
on parallel computers such as the Connection Machine.
Averaging Network
One way to avoid small step changes in hue across a uniform surface resulting
from initial loading into discrete bins is to relax the local requirement for piecewise
299
300
Hurlbert and Poggio
b.
41
97
"
74
Figure 1: (a) Image of a Mondrian-textured sphere - the red channel. (b) Vertical
slice through the specularity in a 75 x 75 pixel region of the three-channel image
(R + G + B) of the same sphere.
constancy and instead require only that hue be smooth within regions delineated by
the edge input. We will see that this local smoothness requirement actually yields
an iterative algorithm that provides asymptotically piecewise constant hue regions.
To implement the local smoothness criterion we use an averaging scheme that simply
replaces the value of each pixel in the hue image with the average of its local
surround, iterating many times over the whole image.
The algorithm takes as input the hue image (either the u-image or the v-image)
and one or two edge images, either luminance edges alone, or luminance edges plus
u or v edges, or u edges plus v edges. The edge images are obtained by performing
Canny edge detection or by using a thresholded directional first derivative. On each
iteration, the value at each pixel in the hue image is replaced by the average of its
value and those in its contributing neighborhood. A neighboring pixel is allowed
to contribute if (i) it is one of the four pixels sharing a full border with the central
pixel (ii) it shares the same edge label with the central pixel in all input edge images
(iii) its value is non-zero and (iv) its value is within a fixed range of the central pixel
value. The last requirement simply reinforces the edge label requirement when a
hue image serves as an input edge image - the edge label requirement allows only
those pixels that lie on the same side of an edge to be averaged, while the other
insures that only those pixels with similar hues are averaged.
More formally
A Network for Image Segmentation Using Color
where Cn(hf,j) is the set of N(C n ) pixels among the next neighbors of i,j that
differ from h~. less than a specified amount and are not crossed by an edge in the
edge map(s) (on the assumption that the pixel (i,j) does not belong to an edge).
The iteration of this operator is similar to nonlinear diffusion and to discontinuous
regularization of the type discussed by Blake and Zisserman (1987), Geman and
Geman (1984) and Marroquin (9]. The iterative scheme of the above equation can
be derived from minimization via gradient descent of the energy function
E
= L:Ei,j
with
=
where V(x, y) V(x - y) is a quadratic potential around 0 and constant for
above a certain value.
Ix - yl
The local averaging smoothes noise in the hue values and spreads uniform hues
across regions marked by the edge inputs. On images with shading but without
strong specularities the algorithm performs a clean segmentation into regions of
different hues.
Conclusions
The averaging scheme finds constant hue regions under the assumptions of a single
source and no strong specularities. A strong highlight may originate an edge that
could then "break" the averaging operation. In our limited experience most specularities seem to average out and disappear from the smoothed hue map, largely
because even strong specularities in the image are much reduced in the initial hue
image. The iterative averaging scheme completely eliminates the remaining gradients in hue. It is possible that more powerful discrimination of specularities will
require specialized routines and higher-level knowledge (Hurlbert, 1989).
Yet this simple network alone is sufficient to reproduce some psychophysical phenomena. In particular, the interaction between brightness and color edges enables
the network to mimic such visual "illusions" as the Koftka Ring. We replicate the
illusion in the following way. A black-and-white Koft'ka Ring (a uniform grey annulus against a rectangular bipartite background, one side black and the other white)
(Hurlbert and Poggio, 1988b) is filtered through the lightness filter estimated in
301
302
Hurlbert and Poggio
h.
a.
c.
9.49679872
9.1989
9
72
9.1122449
9
299
Figure 2: (a) A 75x75 pixel region of the u image, including the specularity. (b) The
image obtained after 500 iterations of the averaging network on (a), using as edge
input the Canny edges of the luminance image. A threshold on differences in the v
image allows only similar v values to be averaged. (c) Vertical slice through center
of (a). (d) Vertical slice at same coordinates through (b) (note different scales of
(c) and (d?.
A Network for Image Segmentation Using Color
the way described elsewhere (Hurlbert and Poggio, 1988a). (For black-and-white
images this step replaces the operation of obtaining u and v: in both cases the goal
is to eliminate spatial gradients of in the effective illumination.) The filtered Koffka
Ring is then fed to the averaging network together with the brightness edges. When
in the input image the boundary between the two parts of the background continues
across the annulus, in the output image (after 2000 iterations of the averaging network) the annulus splits into two semi-annuli of different colors in the output image,
dark grey against the white half, light grey against the black half (Hurlbert, 1989).
When the boundary does not continue across the annulus, the annulus remains a
uniform grey. These results agree with human perception.
Acknowledgements
This report describes research done within the Center for Biological Information
Processing, in the Department of Brain and Cognitive Sciences, and at the Artificial Intelligence Laboratory. This research is sponsored by a grant from the Office
of Naval Research (ONR), Cognitive and Neural Sciences Division; by the Artificial
Intelligence Center of Hughes Aircraft Corporation; by the Alfred P. Sloan Foundation; by the National Science Foundation; by the Artificial Intelligence Center of
Hughes Aircraft Corporation (SI-801534-2); and by the NATO Scientific Affairs Division (0403/87). Support for the A. I. Laboratory's artificial intelligence research is
provided by the Advanced Research Projects Agency of the Department of Defense
under Army contract DACA76-85-C-001O, and in part by ONR contract NOOOI485-K-0124. Tomaso Foggio is supported by the Uncas and Helen Whitaker Chair
at the Massachusetts Institute of Technology, Whitaker College.
References
John Rubin and Whitman Richards. Colour VISIon: representing material categories. Artificial Intelligence Laboratory Memo 764, Massachusetts Institute of
Technology, 1984.
Hsien-Che Lee. Method for computing the scene-illuminant chromaticity from specular highlights. Journal of the Optical Society of America, 3:1694-1699, 1986.
Steven A. Shafer. Using color to separate reflection components. Color Research
and Applications, 10(4):210-218, 1985.
Tomaso Poggio, J. Little, E. Gamble, W. Gillett, D. Geiger, D. Weinshall, M. Villalba, N. Larson, T. Cass, H. Biilthoff, M. Drumheller, P. Oppenheimer,
W. Yang, and A. Hurlbert. The MIT Vision Machine. In Proceedings Image
Understanding Workshop, Cambridge, MA, April 1988. Morgan Kaufmann, San
Mateo, CA.
David Marr and Tomaso Poggio. Cooperative computation of stereo disparity. Science, 194:283-287, 1976.
Anya C. Hurlbert. The Computation of Color. PhD thesis, Massachusetts Institute
of Technology, Cambridge, MA, 1989.
Jose L. Marroquin. Probabilistic Solution of Inverse Problems. PhD thesis, Massachusetts Institute of Technology, Cambridge, MA, 1985.
303
304
Hurlbert and Poggio
Andrew Blake and Andrew Zisserman. Visual Reconstruction. MIT Press, Cambridge, Mass, 1987.
Stuart Geman and Don Geman. Stochastic relaxation, Gibbs distributions, and
the Bayesian restoration of images. IEEE Transactions on Pattern A nalysis and
Machine Intelligence, PAMI-6:721-741, 1984.
Anya C. Hurlbert and Tomaso A. Poggio. Learning a color algorithm from examples.
In Dana Z. Anderson, editor, Neural Information Processing Systems. American
Institute of Physics, 1988.
A. C. Hurlbert and T. A. Poggio. Synthesizing a color algorithm from examples.
Science, 239:482-485, 1988.
| 147 |@word aircraft:2 loading:2 replicate:1 grey:4 brightness:4 shading:1 initial:3 disparity:1 ka:1 comparing:1 si:1 yet:2 must:2 john:1 visible:1 enables:1 sponsored:1 discrimination:1 alone:2 cue:1 half:2 intelligence:6 affair:1 filtered:2 provides:1 contribute:1 along:1 consists:1 tomaso:5 themselves:2 brain:2 encouraging:1 little:1 provided:1 project:1 mass:1 weinshall:1 finding:3 corporation:2 grant:1 segmenting:1 local:5 path:1 pami:1 might:1 plus:2 black:4 mateo:1 conversely:1 shaded:1 limited:1 bi:1 range:1 averaged:3 hughes:2 implement:1 illusion:3 integrating:1 close:1 undesirable:1 operator:3 bh:1 map:2 center:5 helen:1 rectangular:1 fill:4 regularize:1 marr:2 coordinate:3 analogous:1 distinguishing:1 us:1 recognition:2 continues:1 richards:2 geman:4 cooperative:3 constancy:4 steven:1 region:14 agency:1 ideally:1 mondrian:1 bipartite:1 division:2 textured:1 completely:1 whitman:1 hopfield:1 america:1 distinct:1 describe:1 effective:2 artificial:5 neighborhood:1 solve:2 relax:1 otherwise:1 noisy:3 net:1 propose:1 reconstruction:1 interaction:3 canny:2 neighboring:1 achieve:1 forth:1 requirement:5 produce:1 ring:4 object:4 help:2 andrew:2 strong:5 quotient:1 differ:1 inhomogeneous:1 discontinuous:1 filter:2 stochastic:1 human:1 viewing:2 material:8 bin:2 require:4 biological:3 daca76:1 around:1 blake:2 label:10 x75:1 minimization:1 mit:4 sensor:2 always:1 avoid:1 office:1 derived:1 naval:1 indicates:1 eliminate:2 unlikely:1 reproduce:1 pixel:14 among:1 spatial:3 psychophysics:1 field:1 stuart:1 filling:2 mimic:1 report:1 piecewise:5 koffka:1 national:1 replaced:1 geometry:2 detection:1 light:3 edge:33 experience:1 poggio:13 iv:1 reglons:1 restoration:1 introducing:1 neutral:1 uniform:7 recognizing:1 graphic:1 synthetic:2 sensitivity:3 lee:2 yl:1 contract:2 probabilistic:1 physic:1 enhance:1 together:2 thesis:2 reflect:1 central:3 choose:1 cognitive:3 american:1 derivative:1 account:1 potential:1 accompanied:1 explicitly:2 sloan:1 depends:2 crossed:1 later:1 break:1 red:1 hf:1 parallel:2 contribution:1 kaufmann:1 characteristic:1 largely:1 correspond:2 identify:1 yield:2 directional:1 bayesian:1 plastic:1 annulus:6 multiplying:1 processor:1 detector:1 sharing:1 against:3 energy:1 hurlbert:14 attributed:1 massachusetts:4 hur:1 color:37 knowledge:1 segmentation:12 routine:1 marroquin:2 irradiance:4 actually:1 higher:1 zisserman:2 improved:1 april:1 done:1 strongly:1 anderson:1 stage:1 ei:1 nonlinear:1 ascribe:1 scientific:1 requiring:1 true:1 regularization:1 laboratory:4 white:4 chromaticity:2 adjacent:1 lumped:1 larson:1 criterion:1 performs:1 interface:1 reflection:7 image:42 specialized:1 winner:1 phong:1 belong:1 discussed:1 composition:1 cambridge:5 surround:1 ai:3 gibbs:1 smoothness:2 surface:13 certain:1 onr:2 continue:1 meeting:1 preserving:1 morgan:1 signal:1 ii:1 semi:1 full:1 smooth:1 sphere:3 mrf:2 vision:2 iteration:4 cie:1 justified:1 background:2 source:3 crucial:1 eliminates:1 photosensor:1 seem:1 yang:1 iii:1 split:1 specular:4 cn:1 defense:1 colour:1 stereo:2 iterating:1 involve:1 amount:1 dark:1 hue:17 category:1 reduced:1 estimated:1 reinforces:1 smoothe:1 discrete:2 alfred:1 four:1 threshold:2 clean:1 thresholded:1 diffusion:1 nalysis:1 luminance:5 asymptotically:1 relaxation:1 sum:1 jose:1 inverse:1 powerful:1 almost:1 smoothes:1 geiger:1 distinguish:1 replaces:2 quadratic:1 occur:1 constraint:2 scene:4 chair:1 performing:1 optical:1 department:3 according:1 remain:1 across:7 describes:1 delineated:1 biologically:1 modification:1 equation:1 agree:1 remains:1 discus:1 fed:1 serf:2 operation:2 lambertian:1 appropriate:1 spectral:12 alternative:1 original:1 remaining:1 whitaker:3 exploit:1 reflectance:9 especially:1 disappear:1 society:1 psychophysical:1 paint:1 che:1 gradient:3 separate:3 originate:1 toward:1 enforcing:1 localised:1 memo:1 rise:1 synthesizing:1 vertical:3 markov:1 descent:1 behave:1 hsien:1 incorporated:1 smoothed:1 david:1 specified:1 connection:1 discontinuity:6 perception:1 pattern:1 including:1 difficulty:1 rely:1 advanced:1 representing:1 scheme:5 technology:4 lightness:1 irrespective:1 understanding:1 acknowledgement:1 contributing:1 anya:3 highlight:2 specularity:2 analogy:1 dana:1 foundation:2 sufficient:1 rubin:2 editor:1 share:1 land:1 elsewhere:1 koftka:2 supported:1 last:1 side:3 guide:1 institute:5 neighbor:1 absolute:1 slice:3 boundary:10 world:1 contour:2 san:1 transaction:1 approximate:1 biilthoff:1 nato:1 unreliable:2 assumed:1 don:1 iterative:3 triplet:1 channel:4 ca:2 obtaining:1 spread:3 whole:1 noise:4 border:1 shafer:1 allowed:1 body:2 lie:1 ix:1 swns:1 explored:1 essential:1 workshop:1 phd:2 magnitude:2 illumination:6 specularities:8 simply:3 wavelength:2 insures:1 army:1 visual:3 ma:4 marked:3 goal:5 change:13 except:1 averaging:9 pas:1 gamble:1 formally:1 college:2 mark:4 retinex:1 support:1 illuminant:1 phenomenon:1 |
514 | 1,470 | Bach in a Box - Real-Time Harmony
Randall R. Spangler and Rodney M. Goodman*
Computation and Neural Systems
California Institute of Technology, 136-93
Pasadena, CA 91125
Jim Hawkins t
88B Milton Grove
Stoke Newington, London N16 8QY, UK
Abstract
We describe a system for learning J. S. Bach's rules of musical harmony. These rules are learned from examples and are expressed
as rule-based neural networks. The rules are then applied in realtime to generate new accompanying harmony for a live performer.
Real-time functionality imposes constraints on the learning and
harmonizing processes, including limitations on the types of information the system can use as input and the amount of processing
the system can perform. We demonstrate algorithms for generating and refining musical rules from examples which meet these
constraints. We describe a method for including a priori knowledge into the rules which yields significant performance gains. We
then describe techniques for applying these rules to generate new
music in real-time. We conclude the paper with an analysis of
experimental results.
1
Introduction
The goal of this research is the development of a system to learn musical rules from
examples of J.S. Bach's music, and then to apply those rules in real-time to generate
new music in a similar style. These algorithms would take as input a melody such
*rspangle@micro.caltech.edu, rogo@micro.caltech.edu
tjhawkins@cix.compulink.co.uk
R. R. Spangler; R. M. Goodman and J Hawkins
958
I~II- JIJ
Figure 1: Melody for Chorale #1 "Aus meines Herzens Grunde"
Figure 2: J. S. Bach's Harmony For Chorale #1
as Figure 1 and produce a complete harmony such as Figure 2. Performance of this
harmonization in real-time is a challenging problem. It also provides insight into
the nature of composing music.
We briefly review the representation of input data and the process of rule base
generation. Then we focus on methods of increasing the performance of rule-based
systems. Finally we present our data on learning the style of Bach.
1.1
Constraints Imposed by Real-Time Functionality
A program which is to provide real-time harmony to accompany musicians at live
performances faces two major constraints.
First, the algorithms must be fast enough to generate accompaniment without detectable delay between the musician playing the melody and the algorithm generating the corresponding harmony. For musical instrument sounds with sharp attacks
(plucked and percussive instruments, such as the harp or piano), delays of even a
few tens of milliseconds between the start of the melody note and the start of the
harmony notes are noticeable and distracting. This limits the complexity of the
algorithm and the amount of information it can process for each timestep.
Second, the algorithms must base their output only on information from previous timesteps. This differentiates our system from HARMONET (Hild, Feulnzer
and Menzel, 1992) which required knowledge of the next note in the future before
generating harmony for the current note.
1.2
Advantages of a Rule-Based Algorithm
A rule-based neural network algorithm was chosen over a recurrent network or a
non-linear feed-forward network. Neural networks have been previously used for
harmonizing music with some success (Mozer, 1991)(Todd, 1989). However, rulebased algorithms have several advantages when dealing with music. Almost all
music has some sort of rhythm and is tonal, meaning both pitch and duration of
individual notes are quantized. This presents problems in the use of continuous
networks, which must be overtrained to reasonably approximate discrete behavior.
959
Bach in a Box-Real-Time Harmony
Rule-based systems are inherently discrete, and do not have this problem.
Furthermore it is very difficult to determine why a non-linear multi-layer network
makes a given decision or to extract the knowledge contained in such a network.
However, it is straightforward to determine why a rule-based network produced
a given result by examining the rules which fired. This aids development of the
algorithm, since it is easier to determine where mistakes are being made. It allows
comparison of the results to existing knowledge of music theory as shown below, and
may provide insight into the theory of musical composition beyond that currently
available.
Rule-based neural networks can also be modified via segmentation to take advantage
of additional a priori knowledge.
2
Background
2.1
Representation of Input Data
The choice of input representation greatly affects the ability of a learning algorithm
to generate meaningful rules. The learning and inferencing algorithms presented
here speak an extended form of the classical figured bass representation common
in Bach's time. Paired with a melody, figured bass provides a sufficient amount of
information to reconstruct the harmonic content of a piece of music.
Figured bass has several characteristics which make it well-disposed to learning
rules. It is a symbolic format which uses a relatively small alphabet of symbols.
It is also hierarchical - it specifies first the chord function that is to be played at
the current note/timestep, then the scale step to be played by the bass voice, then
additional information as needed to specify the alto and tenor scale steps. This
allows our algorithm to fire sets of rules sequentially, to first determine the chord
function which should be associated with a new melody note, and then to use that
chord function as an input attribute to subsequent rulebases which determine the
bass, alto, and tenor scale steps. In this way we can build up the final chord from
simpler pieces, each governed by a specialized rulebase.
2.2
Generation of Rulebases
Our algorithm was trained on a set of 100 harmonized Bach chorales. These were
translated from MIDI format into our figured bass format by a preprocessing program which segmented them into chords at points where any voice changed pitch.
Chord function was determined by simple table lookup in a table of 120 common
Bach chords based on the scale steps played by each voice in the chord. The algorithm was given information on the current timestep (MelO-TeO), and the previous
two timesteps (Mell-Func2). This produced a set of 7630 training examples, a
subset of which are shown below:
MelO
D
E
F
G
FuncO
V
17
IV
V
800
82
81
80
80
BaO
Bl
B3
Bl
BO
AIO
A2
AO
A2
Al
TeO
TO
T2
Tl
T2
Mell
E
D
E
F
Funcl
I
V
17
IV
801
81
82
81
80
Bal
BO
Bl
B3
Bl
All
AO
A2
AO
A2
Tel
T2
TO
T2
Tl
Me12
C
E
D
E
Func2
I
I
V
17
R. R. Spangler; R. M. Goodman and 1. Hawkins
960
A rulebase is a collection of rules which predict the same right hand side (RHS)
attribute (for example, FunctionO). All rules have the form IF Y=y... THEN
X=x. A rule's order is the number of terms on its left hand side (LHS).
Rules are generated from examples using a modified version of the ITRULE algorithm. (Goodman et al., 1992) All possible rules are considered and ranked by a
measure of the information contained in each rule defined as
J(X; Y
= y) = p(y)
[P(x1Y)log
(p;~~~)) + (I -
p(xly))log
(11-!;~~~)) ]
(1)
This measure trades off the amount of information a rule contains against the probability of being able to use the rule. Rules are less valuable if they contains little
information. Thus, the J-measure is low when p{xly) is not much higher than p(x) .
Rules are also less valuable if they fire only rarely (p(y) is small) since those rules
are unlikely to be useful in generalizing to new data.
A rulebase generated to predict the current chord's function might start with the
following rules:
1. IF
HelodyO
2. IF Function1
AND Helody1
AND HelodyO
3. IF Function1
AND HelodyO
2.3
p(corr) J-meas
0.621
0.095
E
THEN FunctionO
I
V
THEN FunctionO
V7
0.624
0.051
THEN FunctionO
V7
0.662
0.049
D
D
V
D
Inferencing Using Rulebases
Rule based nets are a form of probabilistic graph model. When a rulebase is used
to infer a value, each rule in the rule base is checked in order of decreasing rule
J-measure. A rule can fire if it has not been inhibited and all the clauses on its LHS
are true. When a rule fires, its weight is added to the weight of the value which it
predicts, After all rules have had a chance to fire, the result is an array of weights
for all predicted values.
2.4
Process of Harmonizing a Melody
Input is received a note at a time as a musician plays a melody on a MIDI keyboard.
The algorithm initially knows the current melody note and the data for the last two
timesteps. The system first uses a rule base to determine the chord function which
should be played for the current melody note. For example, given the melody note
"e" , "it might playa chord function "IV", corresponding to an F -Major chord. The
program then uses additional rulebases to specify how the chord will be voiced.
In the example, the bass, alto, and tenor notes might be set to "BO", "AI", and
"T2" , corresponding to the notes "F", "A", and "e". The harmony notes are then
converted to MIDI data and sent to a synthesizer, which plays them in real-time to
accompany the melody.
Bach in a Box-Real-Time Harmony
3
961
Improvement of Rulebases
The J-measure is a good measure for determining the information-theoretic worth of
rules. However, it is unable to take into account any additional a priori knowledge
about the nature of the problem - for example, that harmony rules which use the
current melody note as input are more desirable because they avoid dissonance
between the melody and harmony.
3.1
Segmentation
A priori knowledge of this nature is incorporated by segmenting rulebases into moreand less-desirable rules based on the presence or absence of a desired LHS attribute
such as the current melody note (MelodyO). Rules lacking the attribute are removed
from the primary set of rules and placed in a second "fallback" set. Only in the
event that no primary rules are able to fire is the secondary set allowed to fire. This
gives greater impact to the primary rules (since they are used first) without the loss
of domain size (since the less desirable rules are not actually deleted).
Rulebase segmentation provides substantial improvements in the speed of the algorithm in addition to improving its inferencing ability. When an unsegmented
rule base is fired, the algorithm has to compare the current input data with the LHS
of every rule in the rulebase. However, processing for a segmented rulebase stops
after the first segment which fires a rule on the input data. The algorithm does
not need to spend time examining rules in lower-priority segments of that rulebase.
This increase in efficiency allows segmented rule bases to contain more rules without
impacting performance. The greater number of rules provides a richer and more
robust knowledge base for generating harmony.
3.2
Realtime Dependency Pruning
When rules are used to infer a value, the rules weights are summed to generate probabilities. This requires that all rules which are allowed to fire must be independent
of one another. Otherwise, one good rule could be overwhelmed by the combined
weight of twenty mediocre but virtually identical rules. To prevent this problem,
each segment of a rulebase is analyzed to determine which rules are dependent with
other rules in the same segment. Two rules are considered dependent if they fire
together on more than half the training examples where either rule fires.
For each rule, the algorithm maintains a list of lower rank rules which are dependent
with the rule. This list is used in real-time dependency pruning. Whenever a rule
fires on a given input, all rules dependent on it are inhibited for the duration of the
input. This ensures that all rules which are able to fire for an input are independent.
3.3
Conflict Resolution
When multiple rules fire and predict different values, an algorithm must be used to
resolve the conflict. Simply picking the value with the highest weight, while most
likely to be correct, leads to monotonous music since a given melody then always
produces the same harmony.
To provide a more varied harmony, our system exponentiates the accumulated rule
R. R. Spangler, R. M Goodman and J Hawkins
962
Table 1: Rulebase Segments
RHS
FunctionO
SopranoO
BassO
AltoO
TenorO
REQUIRED LHS FOR SEGMENT
MelodyO, Functionl, Function2
MelodyO,Functionl
MelodyO
MelodyO, FunctionO
FunctionO, SopranoO
(none)
SopranoO, BassO
(none)
SopranoO, BassO, AltoO. FunctionO
SopranoO, Bas80, AltoO
(none)
RULES
llO
380
346
74
125
182
267
533
52
164
115
Table 2: Rulebase Performance
RHS
FunctionO
SopranoO
Bas80
AltoO
TenorO
RULEBASE
un8egmented
segmented
unsegmented # 2
un8egmented
unsegmented
8egmented
unsegmented #2
un8egmented
segmented
unsegmented #2
un8egmented
segmented
unsegmented #2
RULES
1825
816
428
74
307
307
162
800
800
275
331
331
180
AVG EVAL
1825
428
428
74
307
162
162
800
275
275
331
180
180
CORRECT
55%
56%
50%
95%
70%
70%
65%
63%
63%
59%
73%
74%
67%
weights for the possible outcomes to produce probabilities for each value, and the
final outcome is chosen randomly based on those probabilities. It is because we use
the accumulated rule weights to determine these probabilities that all rules which
are allowed to fire must be independent of each other.
If no rules at all fire, the system uses a first-order Bayes classifier to determine the
RlIS value based on the current melody note. This ensures that the system will
always return an outcome compatible with the melody.
4
Results
Rulebases were generated for each attribute. Up to 2048 rules were kept in each
rule base. Rules were retained if they were correct at least 30% of the time they
fired, and had a J-measure greater than 0.001. The rulebases were then segmented.
These rulebases were tested on 742 examples derived from 27 chorales not used in
the training set. The number of examples correctly inferenced is shown for each
rule base before and after segmentation. Also shown is the average number of rules
evaluated per test example; the speed of inferencing is proportional to this number.
To determine whether segmentation was in effect only removing lower J-measure
rules, we removed low-order rules from the unsegmented rule bases until they had
the same average number of rules evaluated as the segmented rule bases.
In all cases, segmenting the rulebases reduced the average rules fired per example
without lowering the accuracy of the rule bases (in some cases, segmentation even
increased accuracy). Speed gains from segmentation ranged from 80% for TenorO up
to 320% for FunctionO. In comparison, simply reducing the size of the unsegmented
963
Bach in a Box-Real-Time Harmony
rulebase to match the speed of the segmented rulebase reduced the number of
correctly inferred examples by 4% to 6%.
The generated rules for harmony have a great deal of similarity to accepted harmonic
transitions (Ottman, 1989). For example, high-priority rules specify common chord
transitions such as V-V7-I (a classic way to end a piece of music).
5
Remarks
The system described in this paper meets the basic objectives described in Section 1.
It learns harmony rules from examples of the music of J.S. Bach. The system is then
able to harmonize melodies in real-time. The generated harmonies are sometimes
surprising (such as the diminished 7th chord near the end of "Happy Birthday"),
yet are consistent with Bach harmony.
I
1\
..
I
I
I
Figure 3: Algorithm's Bach-Like Harmony for "Happy Birthday"
Rulebase segmentation is an effective method for incorporating a priori knowledge
into learned rulebases. It can provides significant speed increases over unsegmented
rule bases with no loss of accuracy.
Acknowledgements
Randall R. Spangler is supported in part by an NSF fellowship.
References
J. Bach (Ed.: A. Riemenschneider) (1941) 371 Harmonized Chorales and 96 Chorale
Melodies. Milwaukee, WI: G. Schirmer.
H. Hild, J. Feulner & W. Menzel. (1992) HARMONET: A Neural Net for Harmonizing Chorales in the Style of J. S. Bach. In J. Moody (ed.), Advances in Neural
Information Processing Systems 4,267-274. San Mateo, CA: Morgan Kaufmann.
M. Mozer, T. Soukup. {1991} Connectionist Music Composition Based on Melodic
and Stylistic Constraints. In R. Lippmann (ed.), Advances in Neural Information
Processing Systems 3. San Mateo, CA: Morgan Kaufmann.
P. Todd. (1989) A Connectionist Approach to Algorithmic Composition. Computer
Music Joumal13(4}:27-43.
R. Goodman, P. Smyth, C. Higgins, J. Miller. {1992} Rule-Based Neural Networks
for Classification and Probability Estimation. Neural Computation 4(6}:781-804.
R. Ottman. (1989) Elementary Harmony. Englewood Cliffs, NJ: Prentice Hall.
| 1470 |@word version:1 briefly:1 llo:1 contains:2 accompaniment:1 existing:1 feulner:1 current:10 surprising:1 synthesizer:1 yet:1 must:6 subsequent:1 half:1 rulebase:15 provides:5 quantized:1 attack:1 simpler:1 harmonize:1 behavior:1 multi:1 decreasing:1 resolve:1 little:1 increasing:1 alto:3 riemenschneider:1 nj:1 every:1 classifier:1 uk:2 segmenting:2 before:2 todd:2 limit:1 mistake:1 cliff:1 meet:2 birthday:2 might:3 au:1 mateo:2 challenging:1 co:1 melodic:1 symbolic:1 mediocre:1 prentice:1 live:2 applying:1 function2:1 imposed:1 musician:3 straightforward:1 duration:2 resolution:1 rule:94 insight:2 array:1 higgins:1 stoke:1 classic:1 play:2 speak:1 smyth:1 us:4 predicts:1 ensures:2 bass:7 trade:1 removed:2 chord:15 valuable:2 highest:1 substantial:1 mozer:2 complexity:1 trained:1 segment:6 efficiency:1 translated:1 alphabet:1 fast:1 describe:3 london:1 effective:1 outcome:3 richer:1 spend:1 reconstruct:1 otherwise:1 ability:2 final:2 advantage:3 net:2 jij:1 basso:3 fired:4 bao:1 produce:3 generating:4 recurrent:1 inferencing:4 received:1 noticeable:1 predicted:1 xly:2 correct:3 functionality:2 attribute:5 func2:2 melody:20 ao:3 itrule:1 harmonization:1 elementary:1 mell:2 accompanying:1 hild:2 hawkins:4 considered:2 hall:1 great:1 algorithmic:1 predict:3 major:2 a2:4 estimation:1 harmony:24 currently:1 teo:2 always:2 modified:2 harmonizing:4 avoid:1 derived:1 focus:1 refining:1 improvement:2 rank:1 greatly:1 dependent:4 accumulated:2 unlikely:1 initially:1 pasadena:1 classification:1 priori:5 impacting:1 development:2 summed:1 harmonet:2 dissonance:1 identical:1 future:1 t2:5 connectionist:2 micro:2 few:1 inhibited:2 randomly:1 individual:1 fire:16 spangler:5 englewood:1 eval:1 analyzed:1 grove:1 lh:5 iv:3 desired:1 monotonous:1 increased:1 aio:1 subset:1 delay:2 examining:2 menzel:2 dependency:2 combined:1 probabilistic:1 off:1 picking:1 together:1 moody:1 priority:2 v7:3 style:3 return:1 account:1 converted:1 lookup:1 piece:3 start:3 sort:1 maintains:1 bayes:1 rodney:1 voiced:1 accuracy:3 musical:5 characteristic:1 kaufmann:2 miller:1 yield:1 produced:2 none:3 worth:1 whenever:1 checked:1 ed:3 against:1 associated:1 gain:2 stop:1 knowledge:9 segmentation:8 actually:1 feed:1 higher:1 specify:3 evaluated:2 box:4 furthermore:1 until:1 hand:2 unsegmented:9 fallback:1 disposed:1 b3:2 effect:1 contain:1 true:1 ranged:1 deal:1 rhythm:1 bal:1 distracting:1 complete:1 demonstrate:1 theoretic:1 meaning:1 harmonic:2 common:3 specialized:1 clause:1 function1:2 significant:2 cix:1 composition:3 ai:1 had:3 similarity:1 base:13 playa:1 milwaukee:1 keyboard:1 success:1 caltech:2 morgan:2 additional:4 greater:3 performer:1 determine:10 ii:1 multiple:1 sound:1 desirable:3 infer:2 segmented:9 match:1 bach:16 melo:2 paired:1 impact:1 pitch:2 basic:1 sometimes:1 qy:1 background:1 addition:1 fellowship:1 goodman:6 accompany:2 virtually:1 sent:1 near:1 presence:1 enough:1 affect:1 timesteps:3 whether:1 tonal:1 remark:1 useful:1 amount:4 ten:1 reduced:2 generate:6 specifies:1 nsf:1 millisecond:1 correctly:2 per:2 discrete:2 harp:1 deleted:1 prevent:1 figured:4 kept:1 lowering:1 timestep:3 graph:1 almost:1 stylistic:1 realtime:2 decision:1 layer:1 played:4 newington:1 constraint:5 rulebased:1 speed:5 format:3 relatively:1 percussive:1 wi:1 randall:2 previously:1 detectable:1 differentiates:1 needed:1 know:1 milton:1 instrument:2 end:2 available:1 apply:1 hierarchical:1 voice:3 overtrained:1 music:14 soukup:1 build:1 classical:1 bl:4 objective:1 added:1 primary:3 unable:1 retained:1 happy:2 difficult:1 twenty:1 perform:1 extended:1 incorporated:1 jim:1 varied:1 sharp:1 inferred:1 required:2 conflict:2 california:1 learned:2 beyond:1 able:4 below:2 program:3 including:2 event:1 ranked:1 chorale:7 technology:1 extract:1 review:1 piano:1 acknowledgement:1 determining:1 lacking:1 loss:2 generation:2 limitation:1 proportional:1 sufficient:1 consistent:1 imposes:1 playing:1 compatible:1 changed:1 placed:1 last:1 supported:1 side:2 institute:1 face:1 transition:2 forward:1 made:1 collection:1 preprocessing:1 avg:1 san:2 approximate:1 pruning:2 lippmann:1 midi:3 dealing:1 sequentially:1 conclude:1 tenor:3 continuous:1 why:2 table:4 learn:1 nature:3 robust:1 ca:3 composing:1 reasonably:1 inherently:1 tel:1 funco:1 improving:1 domain:1 rh:3 allowed:3 tl:2 aid:1 governed:1 learns:1 removing:1 symbol:1 meas:1 list:2 incorporating:1 corr:1 overwhelmed:1 easier:1 generalizing:1 simply:2 likely:1 expressed:1 contained:2 bo:3 chance:1 goal:1 harmonized:2 absence:1 content:1 diminished:1 determined:1 reducing:1 secondary:1 accepted:1 experimental:1 meaningful:1 rarely:1 tested:1 |
515 | 1,471 | A Model of Early Visual Processing
Laurent Itti, Jochen Braun, Dale K. Lee and Christof Koch
{itti, achim, jjwen, koch}Gklab.caltech.edu
Computation & Neural Systems, MSC 139-74
California Institute of Technology, Pasadena, CA 91125, U.S.A.
Abstract
We propose a model for early visual processing in primates. The
model consists of a population of linear spatial filters which interact through non-linear excitatory and inhibitory pooling. Statistical estimation theory is then used to derive human psychophysical
thresholds from the responses of the entire population of units. The
model is able to reproduce human thresholds for contrast and orientation discrimination tasks, and to predict contrast thresholds in
the presence of masks of varying orientation and spatial frequency.
1
INTRODUCTION
A remarkably wide range of human visual thresholds for spatial patterns appears to
be determined by the earliest stages of visual processing, namely, orientation- and
spatial frequency-tuned visual filters and their interactions [18, 19, 3, 22, 9]. Here we
consider the possibility of quantitatively relating arbitrary spatial vision thresholds
to a single computational model. The success of such a unified account should
reveal the extent to which human spatial vision indeed reflects one particular stage
of processing. Another motivation for this work is the controversy over the neural
circuits that generate orientation and spatial frequency tuning in striate cortical
neurons (13, 8, 2]. We think it is likely that behaviorally defined visual filters
and their interactions reveal at least some of the characteristics of the underlying
neural circuitry. Two specific problems are addressed: (i) what is the minimal set
of model components necessary to account for human spatial vision, (ii) is there
a general decision strategy which relates model responses to behavioral thresholds
and which obviates case-by-case assumptions about the decision strategy in different
behavioral situations. To investigate these questions, we propose a computational
model articulated around three main stages: first, a population of bandpass linear
filters extracts visual features from a stimulus; second, linear filters interact through
non-linear excitatory and inhibitory pooling; third, a noise model and decision
strategy are assumed in order to relate the model's output to psychophysical data.
174
2
L Itti, 1. Braun, D. K. Lee and C. Koch
MODEL
We assume spatial visual filters tuned for a variety of orientations e E e and
spatial periods A E A. The filters have overlapping receptive fields in visual space.
Quadrature filter pairs, p{~(r and F{~d, are used to compute a phase-independent
linear energy response, E>.,6, to a visual stimulus S. A small constant background
activity, f, is added to the linear energy responses:
E>. ,6 = \I
. /(peven
>' ,6
* S)2 + (podd
* S)2 + f
>.,6
Filters have separable Gaussian tuning curves in orientation and spatial frequency.
Their corresponding shape in visual space is close to that of Gabor filters, although
not separable along spatial dimensions.
2.1
Pooling: self excitation and divisive inhibition
A model based on linear filters alone would not correctly account for the non-linear
response characteristics to stimulus contrast which have been observed psychophysically [19]. Several models have consequently introduced a non-linear transducer
stage following each linear unit [19]. A more appealing possibility is to assume a
non-linear pooling stage [6, 21, 3, 22]. In this study, we propose a pooling strategy
inspired by Heeger's model for gain control in cat area VI [5, 6]. The pooled response R>.,6 of a unit tuned for (A, 0) is computed from the linear energy responses
of the entire population:
E'Y
R>. >',6
+ 1]
(1)
,6 - So + L>'I,61W>.,6(N, OI)E~/,61
where the sum is taken over the entire population and W>.,6 is a two-dimensional
Gaussian weighting function centered around (A,O), and 1] a background activity.
The numerator in Eq. 1 represents a non-linear self-excitation term. The denominator represents a divisive inhibitory term which depends not only on the activity
of the unit (A,O) of interest, but also on the responses of other units . We shall see
in Section 3 that, in contrast to Heeger's model for electrophysiological data in
which all units contribute equally to the pool, it is necessary to assume that only a
subpopulation of units with tuning close to (A, 0) contribute to the pool in order to
account for psychophysical data. Also, we assume, > 15 to obtain a power law for
high contrasts [7], as opposed to Heeger's physiological model in which, = 15 = 2
to account for neuronal response saturation at high contrasts.
Several interesting properties result from this pooling model. First, a sigmoidal
transducer function - in agreement with contrast discrimination psychophysics - is
naturally obtained through pooling and thus need not be introduced post-hoc. The
transducer slope for high contrasts is determined by ,-15, the location of its inflexion
point by 5, and the slope at this point by the absolute value of, (and 15). Second, the
tuning curves of the pooled units for orientation and spatial period do not depend
of stimulus contrast, in agreement with physiological and psychophysical evidence
[14]. In comparison, a model which assumes a non-linear transducer but no pooling
exhibits sharper tuning curves for lower contrasts. Full contrast independence of
the tuning is achieved only when all units participate in the inhibitory pool; when
only sub-populations participate in the pool, some contrast dependence remains.
2.2
Noise model: Poisson lX
It is necessary to assume the presence of noise in the system in order to be able to
derive psychophysical performance from the responses of the population of pooled
A Model of Early Visual Processing
175
units. The deterministic response of each unit then represents the mean of a randomly distributed "neuronal" response which varies from trial to trial in a simulated
psychophysical experiment .
Existing models usually assume constant noise variance in order to simplify the
subsequent decision stage [18]. Using the decision strategy presented below, it is
however possible to derive psychophysical performance with a noise model whose
variance increases with mean activity, in agreement with electrophysiology [16].
In what follows, Poisson cx noise will be assumed and approximated by a Gaussian
random variable with variance mean cx (0' is a constant close to unity).
=
2.3
Decision strategy
We use tools from statistical estimation theory to compute the system's behavioral
response based on the responses of the population of pooled units. Similar tools
have been used by Seung and Sompolinsky [12] under the simplifying assumption of
purely Poisson noise and for the particular task of orientation discrimination in the
limit of an infinite population of oriented units. Here, we extend this framework
to the more general case in which any stimulus attribute may differ between the
two stimulus presentations to be discriminated by the model. Let's assume that we
want to estimate psychophysical performance at discriminating between two stimuli
which differ by the value of a stimulus parameter ((e.g . contrast, orientation,
spatial period).
The central assumption of our decision strategy is that the brain implements an
unbiased efficient statistic T(R; (), which is an estimator of the parameter ( based
{R).,I/; A E A, () E 0}. The efficient statistic is
on the population response R
the one which, among all possible estimators of (, has the property of minimum
variance in the estimated value of ( . Although we are not suggesting any putative
neuronal correlate for T, it is important to note that the assumption of efficient
statistic does not require T to be prohibitively complex; for instance, a maximum
likelihood estimator proposed in the decision stage of several existing models is
asymptotically (with respect to the number of observations) a efficient statistic.
=
Because T is efficient, it achieves the Cramer-Rao bound [1]. Consequently, when
the number of observations (i .e. simulated psychophysical trials) is large,
E[T] = (
and
var[T] = 1/3(()
where E[.] is the mean over all observations, var[.] the variance, and 3(() is the
Fisher information. The Fisher information can be computed using the noise model
assumption and tuning properties of the pooled units: for a random variable X
with probability density f(x; (), it is given by [1]:
J(() = E
[:c
In/(X;()r
For our Poisson cx noise model and assuming that different pooled units are independent [15], this translates into:
One unit R). ,I/:
All independent units:
The Fisher information computed for each pooled unit and three types of stimulus
parameters ( is shown in Figure 1. This figure demonstrates the importance of
using information from all units in the population rather than from only one unit
optimally tuned for the stimulus: although the unit carrying the most information
about contrast is the one optimally tuned to the stimulus pattern, more information
L. lui, 1 Braun, D. K. Lee and C. Koch
176
about orientation or spatial frequency is carried by units which are tuned to flanking
orientations and spatial periods and whose tuning curves have maximum slope for
the stimulus rather than maximum absolute sensitivity. In our implementation,
the derivatives of pooled responses used in the expression of Fisher information are
computed numerically.
orientation
spatial frequency
Figure 1: Fisher information computed for contrast, orientation and spatial frequency.
Each node in the tridimensional meshes represents the Fisher information for the corresponding pooled unit (A, B) in a model with 30 orientations and 4 scales. Arrows indicate
the unit (A, B) optimally tuned to the stimulus. The total Fisher information in the population is the sum of the information for all units.
Using the estimate of ( and its variance from the Fisher information, it is possible to derive psychophysical performance for a discrimination task between two
stimuli with parameters (1 ~ (2 using standard ideal observer signal discrimination
techniques [4] . For such discrimination, we use the Central Limit Theorem (in the
limit of large number of trials) to model the noisy responses of the system as two
Gaussians with means (1 and (2, and variances lTi
1/:1((d and lTi
1/:1((2)
respectively. A decision criterion D is chosen to minimize the overall probability of
error; since in our case lT1 =f. lT2 in general, we derive a slightly more complicated
expression for performance P at a Yes/No (one alternative forced choice) task than
what is commonly used with models assuming constant noise [18]:
=
D =
(2 lT
i - (llT~ -
=
lT1lT2J((1 - (2)2 + 2(lTr - lTi) log(lT!/lT2)
2
2
lT1 - lT2
P=
~+~erf((2-D)
2
4
lT2..J2
+
~erf(D-(l)
4
lT1..J2
where erf is the Normal error function. The expression for D extends by continuity
to D = ((2 - (1)/2 when lT1 = lT2 . This decision strategy provides a unified, taskindependent framework for the computation of psychophysical performance from the
deterministic responses of the pooled units. This strategy can easily be extended to
allow the model to perform discrimination tasks with respect to additional stimulus
parameters, under exactly the same theoretical assumptions.
3
3.1
RESULTS
Model calibration
The parameters of the model were automatically adjusted to fit human psychophysical thresholds measured in our laboratory [17] for contrast and orientation discrimination tasks (Figure 2). The model used in this experiment consisted of 60
orientations evenly distributed between 0 and 180deg. One spatial scale at 4 cycles
per degree (cpd) was sufficient to account for the data. A multidimensional simplex
method with simulated annealing overhead was used to determine the best fit of
the model to the data [10]. The free parameters adjusted during the automatic
A Model of Early VlSUal Processing
177
fits were: the noise level a, the pooling exponents 'Y and &, the inhibitory pooling
constant 5, and the background firing rates, E and rJ.
The error function minimized by the fitting algorithm was a weighted average of
three constraints: 1) least-square error with the contrast discrimination data in
Figure 2.a; 2) least-square error with the orientation discrimination data in Figure 2.h; 3) because the data was sparse in the "dip-shaped" region of the curve
in Figure 2.a, and unreliable due to the limited contrast resolution of the display used for the psychophysics, we added an additional constraint favoring a more
pronounced "dip", as has been observed by several other groups [11, 19, 22] .
Data fits used for model calibration:
a
. -_ _ _ _ _---..:a:::..,
iii
~u ~
~
0-
~ ~ 10-2
c:Ch
Q)~
E :5 10-3 L..-_ _ _ _ _ _--...J
oc:
Q)
10
.-
?2
10
0
0.2
0.4
stimulus contrast
mask contrast
Transducer function:
-
~50.----________c~
Q)
C
o
a.
~
~
~
0.5
Q)
>
uQ)
8.
d
Q)
Ch
o
a.
Ch
c:
(5
I;
Orientation tuning:
0.5
stimulus contrast
~
~
O~~----~----=-~
-100
0
100
stimulUS orientation (deg)
Figure 2: The model (solid lines) was calibrated using data from two psychophysical
experiments: (a) discrimination between a pedestal contrast (a.a) and the same pedestal
plus an increment contrast (a.{3); (b) discrimination between two orientations near vertical
(b.a and b.{3). After calibration, the transducer function of each pooled unit (c) correctly
exhibits an accelerating non-linearity near threshold (contrast ~ 1%) and compressive
non-linearity for high contrasts (Weber's law). We can see in (d) that pooling among
units with similar tuning properties sharpens their tuning curves. Model parameters were:
a ~ 0.75,,), ~ 4,?5 ~ 3.5,E ~ 1%, '1 ~ 1.7Hz,S such that transducer inflexion point is
at 4x detection threshold contrast, orientation tuning FWHM=68deg (full width at half
maximum), orientation pooling FWHM=40deg.
Two remaining parameters are the orientation tuning width, (7'8, of the filters and
the width, (7'We, of the pool. It was not possible from the data in Figure 2 alone
to unambiguously determine these parameters. However, for any given (7'8, (7'W8
is uniquely determined by the following two qualitative constraints: first, a small
pool size is not desirable because it yields contrast-dependent orientation tuning;
it however appears from the data in Figure 2.h that this tuning should not vary
much over a wide range of contrasts. The second constraint is qualitatively derived
from Figure 3.a: for large pool sizes, the model predicted significant interference
between mask and test patterns even for large orientation differences. Such inter-
178
L Itti, 1. Braun, D. K. Lee and C. Koch
ference was not observed in the data for orientation differences larger than 45deg .
It consequently seems that a partial inhibitory pool, composed only of a fraction of
the population of oriented filters with tuning similar to the central excitatory unit,
accounts best for the psychophysical data. Finally, (76 was fixed so as to yield a
correct qualitative curve shape for Figure 3.a.
3.2
Predictions
We used complex stimuli from masking experiments to test the predictive value
of the model (Figure 3). Although it was necessary to use some of the qualitative properties of the data seen in Figure 3.a to calibrate the model as detailed
above, the calibrated model correctly produced a quantitative fit of this data. The
calibrated model also correctly predicted the complex data of Figure 3.h.
a
c::10
0
a
~
>
Q)
~
>
Q)
CD 5
Q)
~
"C
(5
.r;
(/J
Q)
~
.r;
......
c::
0
no
mask orientation (deg) mask
0
30
60
90
"C
(5
b
6
4
2
.r;
(/J
Q)
~
.r;
......
2
4
8
mask spatial freq. (cpd)
Figure 3: Prediction of psychophysical contrast thresholds in the presence of an oblique
mask. The mask was a 50%-contrast stochastic oriented pattern (a). and the superimposed test pattern was a sixth-derivative of Gaussian bar (j3). In (a), threshold elevation
(i.e. ratio of threshold in the presence of mask to threshold in the absence of mask) was
measured for varying mask orientation, for mask and test patterns at 4 cycles per degree
(cpd). In (b), orientation difference between test and mask was fixed to 15deg, and threshold elevation was measured as a function of mask spatial frequency. Solid lines represent
model predictions, and dashed lines represent unity threshold elevation.
4
DISCUSSION AND CONCLUSION
We have developed a model of early visual processing in humans which accounts for
a wide range of measured spatial vision thresholds and which predicts behavioral
thresholds for a potentially unlimited number of spatial discriminations. In addition to orientation- and spatial-frequency-tuned units, we have found it necessary to
assume two types of interactions between such units: (i) non-linear self-excitation
of each unit and (ii) divisive normalization of each unit response relative to the
responses of similarly tuned units. All model parameters are constrained by psychophysical data and an automatic fitting procedure consistently converged to the
same parameter set regardless of the initial position in parameter space.
Our two main contributions are the small number of model components and the un i.fied, task-independent decision strategy. Rather than making different assumptions
about the decision strategy in different behavioral tasks, we combine the information contained in the responses of all model units in a manner that is optimal for
any behavioral task. We suggest that human observers adopt a similarly optimal
decision procedure as they become familiar with a particular task (" task set"). Although here we apply this decision strategy only to the discrimination of stimulus
contrast, orientation, and spatial frequency, it can readily be generalized to arbitrary discriminations such as, for example, the discrimination of vernier targets.
A Model of Early Vzsual Processing
179
So far we have considered only situations in which the same decision strategy is
optimal for every stimulus presentation. We are now studying situations in which
the optimal decision strategy varies unpredictably from trial to trial (" decision
uncertainty"). For example, situations in which the observer attempts to detect an
increase in either the spatial frequency or the contrast of stimulus. In this way, we
hope to learn the extent to which our model reflects the decision strategy adopted
by human observers in an even wider range of situations. We have also assumed
that the model's units were independent, which is not strictly true in biological
systems (although the main source of correlation between neurons is the overlap
between their respective tuning curves, which is accounted for in the model). The
mathematical developments necessary to account for fixed or variable covariance
between units are currently under study.
In contrast to other models of early visual processing [5, 6], we find that the psychophysical data is consistent only with interactions between similarly tuned units
(e.g., "near-orientation inhibition")' not with interactions between units of very different tuning (e.g., "cross-orientation inhibition") . Although such partial pooling
does not render tuning functions completely contrast-independent, an additional degree of contrast-independence could be provided by pooling across different spatial
locations. This issue is currently under investigation.
In conclusion, we have developed a model based on self-excitation of each unit,
divisive normalization [5, 6] between similarly tuned units, and an ideal observer
decision strategy. It was able to reproduce a wide range of human visual thresholds.
The fact that such a simple and idealized model can account quantitatively for
a wide range of psychophysical observations greatly strengthens the notion that
spatial vision thresholds reflect processing at one particular neuroanatomical level.
Acknowledgments: This work was supported by NSF-Engineering Research Center
(ERC), NIMH, ONR, and the Sloan Center for Theoretical Neurobiology.
References
[1] Cover TM, Thomas JA. Elem Info Theo, Wiley & Sons, 1991
[2] Ferster D, Chung S, Wheat H. Nature 1996;380(6571):249-52
[3] Foley JM. J Opt Soc A 1994;11(6):1710-9
[4] Green DM, Swets JA. Signal Detectability and Psychophys. Wiley & Sons, 1966.
[5] Heeger DJ. Comput Models of Vis Processing, MIT Press, 1991
[6] Heeger DJ . Vis Neurosci 1992;9:181-97
[7] Nachmias J, Sansbury RV. Vis Res 1974;14:1039-42
[8] Nelson S, Toth L, Sheth B, Sur M. Science 1994;265(5173):774-77
[9] Perona P, Malik J. J Opt Soc A 1990;7(5):923-32
[10] Press WH, Teukolsky SA, et al. Num Rec in C. Cambridge University Press, 1992
[ll] Ross J, Speed HD. Proc R Soc B 1991;246:61-9
[12] Seung HS, Sompolinksy H. Proc Natl Acad Sci USA 1993;90:10749-53.
[13] Sillito AM. Progr Brain Res 1992;90:349-84
[14] Skottun BC, Bradley A, Sclar G et al. J Neurophys 1987;57(3):773-86
[15] Snippe HP, Koenderink JJ. Bioi Cybern 1992;67:183-90
[16] Teich MC, Thrcott RG, Siegel RM. IEEE Eng Med Bioi 1996;Sept-Oct,79-87
[17] Wen J, Koch C , Braun J. Proc ARVO 1997;5457
[18] Wilson HR, Bergen JR. Vis Res 1979; 19: 19-32
[19] Wilson HR. Bioi Cybern 1980;38: 171-8
[20] Wilson HR, McFarlane DK, Phillips GC. Vis Res 1983;23;873-82.
[21] Wilson HR, Humanski R. Vis Res 1993;33(8):1133-50
[22] Zenger B, Sagi D. Vis Res 1996;36(16):2497-2513.
| 1471 |@word h:1 trial:6 sharpens:1 seems:1 teich:1 simplifying:1 covariance:1 eng:1 solid:2 initial:1 tuned:11 bc:1 existing:2 bradley:1 neurophys:1 readily:1 mesh:1 subsequent:1 shape:2 discrimination:16 alone:2 half:1 oblique:1 num:1 provides:1 contribute:2 location:2 lx:1 node:1 sigmoidal:1 mathematical:1 along:1 become:1 transducer:7 consists:1 qualitative:3 overhead:1 fitting:2 behavioral:6 combine:1 manner:1 swets:1 inter:1 mask:14 indeed:1 brain:2 inspired:1 automatically:1 jm:1 provided:1 zenger:1 underlying:1 linearity:2 circuit:1 what:3 developed:2 compressive:1 unified:2 w8:1 quantitative:1 multidimensional:1 every:1 braun:5 exactly:1 prohibitively:1 demonstrates:1 rm:1 control:1 unit:41 christof:1 engineering:1 sagi:1 limit:3 acad:1 laurent:1 firing:1 plus:1 limited:1 range:6 fwhm:2 acknowledgment:1 implement:1 procedure:2 area:1 gabor:1 subpopulation:1 suggest:1 close:3 cybern:2 deterministic:2 center:2 regardless:1 resolution:1 estimator:3 hd:1 population:13 notion:1 increment:1 target:1 vzsual:1 agreement:3 approximated:1 strengthens:1 rec:1 predicts:1 observed:3 region:1 wheat:1 cycle:2 sompolinsky:1 nimh:1 seung:2 controversy:1 depend:1 carrying:1 predictive:1 purely:1 completely:1 easily:1 cat:1 articulated:1 forced:1 whose:2 larger:1 arvo:1 statistic:4 erf:3 think:1 noisy:1 hoc:1 propose:3 interaction:5 j2:2 pronounced:1 wider:1 derive:5 measured:4 sa:1 eq:1 soc:3 predicted:2 indicate:1 differ:2 correct:1 attribute:1 filter:13 stochastic:1 snippe:1 centered:1 human:10 require:1 ja:2 investigation:1 elevation:3 opt:2 biological:1 adjusted:2 strictly:1 koch:6 around:2 cramer:1 normal:1 considered:1 predict:1 circuitry:1 achieves:1 early:7 vary:1 adopt:1 estimation:2 proc:3 currently:2 ross:1 tool:2 reflects:2 weighted:1 hope:1 mit:1 behaviorally:1 gaussian:4 rather:3 varying:2 wilson:4 earliest:1 derived:1 consistently:1 likelihood:1 superimposed:1 greatly:1 contrast:35 detect:1 am:1 dependent:1 bergen:1 entire:3 achim:1 pasadena:1 perona:1 favoring:1 reproduce:2 overall:1 among:2 orientation:33 issue:1 sheth:1 exponent:1 development:1 ference:1 spatial:28 constrained:1 psychophysics:2 field:1 shaped:1 represents:4 jochen:1 simplex:1 minimized:1 stimulus:22 simplify:1 quantitatively:2 cpd:3 wen:1 randomly:1 oriented:3 composed:1 inflexion:2 familiar:1 phase:1 attempt:1 detection:1 interest:1 possibility:2 investigate:1 natl:1 partial:2 necessary:6 respective:1 re:6 theoretical:2 minimal:1 instance:1 rao:1 cover:1 calibrate:1 optimally:3 varies:2 psychophysically:1 calibrated:3 density:1 sensitivity:1 discriminating:1 lee:4 pool:8 central:3 reflect:1 opposed:1 derivative:2 itti:4 chung:1 koenderink:1 account:10 suggesting:1 pedestal:2 pooled:11 sloan:1 vi:8 depends:1 idealized:1 observer:5 complicated:1 masking:1 slope:3 contribution:1 minimize:1 oi:1 square:2 variance:7 characteristic:2 yield:2 yes:1 produced:1 mc:1 converged:1 llt:1 sixth:1 energy:3 frequency:11 dm:1 naturally:1 gain:1 wh:1 electrophysiological:1 appears:2 unambiguously:1 response:21 stage:7 msc:1 correlation:1 overlapping:1 continuity:1 reveal:2 usa:1 consisted:1 unbiased:1 true:1 progr:1 laboratory:1 freq:1 ll:1 numerator:1 self:4 during:1 width:3 uniquely:1 excitation:4 oc:1 criterion:1 generalized:1 weber:1 discriminated:1 extend:1 relating:1 numerically:1 elem:1 significant:1 cambridge:1 phillips:1 tuning:19 automatic:2 hp:1 similarly:4 erc:1 dj:2 calibration:3 inhibition:3 onr:1 success:1 caltech:1 seen:1 minimum:1 additional:3 determine:2 period:4 signal:2 ii:2 relates:1 full:2 desirable:1 rj:1 dashed:1 rv:1 cross:1 post:1 equally:1 prediction:3 j3:1 denominator:1 vision:5 poisson:4 represent:2 normalization:2 achieved:1 background:3 remarkably:1 want:1 addition:1 addressed:1 annealing:1 source:1 pooling:14 hz:1 med:1 near:3 presence:4 ideal:2 iii:1 variety:1 independence:2 fit:5 tm:1 translates:1 expression:3 accelerating:1 render:1 jj:1 detailed:1 generate:1 nsf:1 inhibitory:6 estimated:1 correctly:4 per:2 detectability:1 shall:1 group:1 threshold:19 lti:3 asymptotically:1 fraction:1 sum:2 uncertainty:1 extends:1 putative:1 decision:19 bound:1 display:1 activity:4 constraint:4 unlimited:1 speed:1 separable:2 nachmias:1 jr:1 across:1 slightly:1 son:2 unity:2 appealing:1 primate:1 making:1 lt2:5 flanking:1 interference:1 taken:1 remains:1 studying:1 adopted:1 gaussians:1 apply:1 uq:1 alternative:1 thomas:1 obviates:1 assumes:1 remaining:1 neuroanatomical:1 humanski:1 psychophysical:18 malik:1 question:1 added:2 strategy:16 receptive:1 dependence:1 striate:1 exhibit:2 simulated:3 sci:1 participate:2 evenly:1 nelson:1 extent:2 assuming:2 sur:1 ratio:1 sharper:1 relate:1 potentially:1 vernier:1 info:1 implementation:1 perform:1 vertical:1 neuron:2 observation:4 situation:5 extended:1 neurobiology:1 gc:1 arbitrary:2 introduced:2 namely:1 pair:1 california:1 able:3 bar:1 usually:1 pattern:6 below:1 psychophys:1 mcfarlane:1 saturation:1 green:1 power:1 overlap:1 hr:4 technology:1 carried:1 extract:1 foley:1 sept:1 relative:1 law:2 interesting:1 var:2 degree:3 ltr:1 sufficient:1 consistent:1 cd:1 excitatory:3 accounted:1 supported:1 free:1 theo:1 allow:1 institute:1 wide:5 absolute:2 sparse:1 distributed:2 curve:8 dimension:1 cortical:1 dip:2 dale:1 commonly:1 qualitatively:1 far:1 correlate:1 unreliable:1 deg:7 assumed:3 un:1 sillito:1 learn:1 nature:1 toth:1 ca:1 interact:2 complex:3 main:3 neurosci:1 arrow:1 motivation:1 noise:11 quadrature:1 fied:1 neuronal:3 siegel:1 lt1:4 wiley:2 sub:1 position:1 heeger:5 bandpass:1 comput:1 third:1 weighting:1 theorem:1 specific:1 dk:1 physiological:2 evidence:1 importance:1 rg:1 cx:3 electrophysiology:1 lt:2 likely:1 visual:15 contained:1 sclar:1 ch:3 teukolsky:1 oct:1 bioi:3 presentation:2 consequently:3 ferster:1 fisher:8 absence:1 determined:3 infinite:1 lui:1 unpredictably:1 total:1 divisive:4 |
516 | 1,472 | Hierarchical Non-linear Factor Analysis
and Topographic Maps
Zoubin Ghahramani and Geoffrey E. Hinton
Dept. of Computer Science, University of Toronto
Toronto, Ontario, M5S 3H5, Canada
http://www.cs.toronto.edu/neuron/
{zoubin,hinton}Ocs.toronto.edu
Abstract
We first describe a hierarchical, generative model that can be
viewed as a non-linear generalisation of factor analysis and can
be implemented in a neural network. The model performs perceptual inference in a probabilistically consistent manner by using
top-down, bottom-up and lateral connections. These connections
can be learned using simple rules that require only locally available information. We then show how to incorporate lateral connections into the generative model. The model extracts a sparse,
distributed, hierarchical representation of depth from simplified
random-dot stereograms and the localised disparity detectors in
the first hidden layer form a topographic map. When presented
with image patches from natural scenes, the model develops topographically organised local feature detectors.
1
Introduction
Factor analysis is a probabilistic model for real-valued data which assumes that
the data is a linear combination of real-valued uncorrelated Gaussian sources (the
factors). After the linear combination, each component of the data vector is also
assumed to be corrupted by additional Gaussian noise. A major advantage of this
generative model is that, given a data vector, the probability distribution in the
space of factors is a multivariate Gaussian whose mean is a linear function of the
data. It is therefore tractable to compute the posterior distribution exactly and to
use it when learning the parameters of the model (the linear combination matrix
and noise variances). A major disadvantage is that factor analysis is a linear model
that is insensitive to higher order statistical structure of the observed data vectors.
One way to make factor analysis non-linear is to use a mixture of factor analyser
modules, each of which captures a different linear regime in the data [3]. We can
view the factors of all of the modules as a large set of basis functions for describing
the data and the process of selecting one module then corresponds to selecting
an appropriate subset of the basis functions. Since the number of subsets under
consideration is only linear in the number of modules, it is still tractable to compute
Hierarchical Non-linear Factor Analysis and Topographic Maps
487
the full posterior distribution when given a data point. Unfortunately, this mixture
model is often inadequate. Consider, for example, a typical image that contains
multiple objects. To represent the pose and deformation of each object we want
a componential representation of the object's parameters which could be obtained
from an appropriate factor analyser. But to represent the multiple objects we need
several of these componential representations at once, so the pure mixture idea is
not tenable. A more powerful non-linear generalisation of factor analysis iF to have
a large set of factors and to allow any subset of the factors to be selected. This
can be achieved by using a generative model in which there is a high probability of
generating factor activations of exactly zero.
2
Rectified Gaussian Belief Nets
The Rectified Gaussian Belief Net (RGBN) uses multiple layers of units with states
that are either positive real values or zero [5]. Its main disadvantage is that computing the posterior distribution over the factors given a data vector involves Gibbs
sampling. In general, Gibbs sampling can be very time consuming, but in practice
10 to 20 samples per unit have proved adequate and there are theoretical reasons
for believing that learning can work well even when the Gibbs sampling fails to
reach equilibrium [10].
We first describe the RGBN without considering neural plausibility. Then we show
how lateral interactions within a layer can be used to perform probabilistic inference correctly using locally available information. This makes the RGBN far more
plausible as a neural model than a sigmoid belief net [9, 8] because it means that
Gibbs sampling can be performed without requiring units in one layer to see the
total top-down input to units in the layer below.
The generative model for RGBN's consists of multiple layers of units each of which
has a real-valued unrectified state, Yj, and a rectified state, [Yj]+, which is zero if
Yj is negative and equal to Yj otherwise. This rectification is the only non-linearity
in the network. 1 The value of Yj is Gaussian distributed with a standard deviation
(Jj and a mean, ih that is determined by the generative bias, gOj, and the combined
effects of the rectified states of units, k, in the layer above:
Yj = gOj
+ Lgkj[Yk]+
(1)
k
The rectified state [Yj]+ therefore has a Gaussian distribution above zero, but all
of the mass of the Gaussian that falls below zero is concentrated in an infinitely
dense spike at zero as shown in Fig. la. This infinite density creates problems if we
attempt to use Gibbs sampling over the rectified states, so, following a suggestion
by Radford Neal, we perform Gibbs sampling on the unrectified states.
Consider a unit, j, in some intermediate layer of a multilayer RGBN. Suppose
that we fix the unrectified states of all the other units in the net. To perform Gibbs
sampling, we need to stochastically select a value for Yj according to its distribution
given the unrectified states of all the other units. If we think in terms of energy
functions, which are equal to negative log probabilities (up to a constant), the
rectified states of the units in the layer above contribute a quadratic energy term
by determining Yj. The unrectified states of units, i, in the layer below contribute a
constant if [Yj]+ is 0, and if [Yj]+ is positive they each contribute a quadratic term
1 The key arguments presented in this paper hold for general nonlinear belief networks
as long as the noise is Gaussian; they are not specific to the rectification nonlinearity.
Z Ghahramani and G. E. Hinton
488
a
c
b
I
I
I
W
/
~----..J,'
, Top-down
,
'-_ .. "
-3-2-1 0 1 2 3 -3-2-1 0 1 2 3
Y
because of the effect of [Yj] + on
Y
Figure 1: a) Probability density in which all the mass of a
Gaussian below zero has been
replaced by an infinitely dense
spike at zero. b) Schematic
of the density of a unit's unrectified state. c) Bottomup and top-down energy functions corresponding to b.
Yi.
(2)
where h is an index over all the units in the same layer as j including j itself. Terms
that do not depend on Yj have been omitted from Eq. 2. For values of Yj below zero
there is a quadratic energy function which leads to a Gaussian distribution. The
same is true for values of Yj above zero, but it is a different quadratic (Fig . Ic) . The
Gaussian distributions corresponding to the two quadratics must agree at Yj
0
(Fig. Ib). Because this distribution is piecewise Gaussian it is possible to perform
Gibbs sampling exactly.
=
Given samples from the posterior, the generative weights of a RGBN can be learned
by using the online delta rule to maximise the log probability of the data. 2
(3)
The variance of the local Gaussian noise of each unit, o}, can also be learned by
an online rule, D-.o}
f [(Yj - Yj)2 - o}]. Alternatively, o} can be fixed at I for
all hidden units and the effective local noise level can be controlled by scaling the
generative weights.
=
3
The Role of Lateral Connections in Perceptual Inference
In RGBNs and other layered belief networks, fixing the value of a unit in one layer
causes correlations between the parents of that unit in the layer above. One of
the main reasons why purely bottom-up approaches to perceptual inference have
proven inadequate for learning in layered belief networks is that they fail to take
into account this phenomenon, which is known as "explaining away."
Lee and Seung (1997) introduced a clever way of using lateral connections to handle
explaining away effects during perceptual inference. Consider the network shown
in Fig. 2. One contribution, Ebelow, to the energy of the state of the network is
the squared difference between the unrectified states of the units in one layer, Yj,
a.nd the top-down expectations generated by the states of units in the layer above.
Assuming the local noise models for the lower layer units all have unit variance, and
2 If Gibbs sampling has not been run long enough to reach equilibrium, the delta rule
follows the gradient of the penalized log probability of the data [10]. The penalty term is
the Kullback-Liebler divergence between the equilibrium distribution and the distribution
produced by Gibbs sampling. Other things being equal, the delta rule therefore adjusts
the parameters that determine the equilibrium distribution to reduce this penalty, thus
favouring models for which Gibbs sampling works quickly.
Hierarchical Non-linear Factor Analysis and Topographic Maps
489
ignoring biases and constant terms that are unaffected by the states of the units
Ebe\ow
= ~ l:)Yj -
= ~ I)Yj -
Yj)2
j
= gkj and mkl = - Lj gkjglj
Rearranging this expression and setting rjk
Ebe\ow
= ~ LyJ j
2:k[Yk]+9kj)2.
(4)
j
L[Yk]+ LYjrjk k
~ L[Yk]+ L[y!l+mkl .
k
j
we get
(5)
I
This energy function can be exactly implemented in a network with recognition
weights, rjk, and symmetric lateral interactions, mkl. The lateral and recognition
connections allow a unit, k, to compute how Ebe\ow for the layer below depends on
its own state and therefore they allow it to follow the gradient of E or to perform
Gibbs sampling in E .
Figure 2: A small segment of a network,
showing the generative weights (dashed) and
the recognition and lateral weights (solid)
which implement perceptual inference and
correctly handle explaining away effects.
Seung's trick can be used in an RGBN and it eliminates the most neurally implausible aspect of this model which is that a unit in one layer appears to need to send
both its state Y and the top-down prediction of its state Y to units in the layer above.
Using the lateral connections, the units in the layer above can, in effect, compute
all they need to know about the top-down predictions. In computer simulations, we
can simply set each lateral connection mkl to be the dot product - 2: j gkjglj. It is
also possible to learn these lateral connections in a more biologically plausible way
by driving units in the layer below with unit-variance independent Gaussian noise
and using a simple anti-Hebbian learning rule. Similarly, a purely local learning
rule can learn recognition weights equal to the generative weights . .If units at one
layer are driven by unit-variance, independent Gaussian noise, and these in turn
drive units in the layer below using the generative weights, then Hebbian learning
between the two layers will learn the correct recognition weights [5].
4
Lateral Connections in the Generative Model
When the generative model contains only top-down connections, lateral connections
make it possible to do perceptual inference using locally available information. But
it is also possible, and often desirable, to have lateral connections in the generative
model. Such connections can cause nearby units in a layer to have a priori correlated
activities, which in turn can lead to the formation of redundant codes and, as we
will see, topographic maps.
Symmetric lateral interactions between the unrectified states of units within a layer
have the effect of adding a quadratic term to the energy function
EMRF
=
~ L: L
k
Mkl YkYI,
(6)
I
which corresponds to a Gaussian Markov Random Field (MRF). During sampling,
this term is simply added to the top-down energy contribution. Learning is more
difficult. The difficulty sterns from the need to know the derivatives of the partition
function of the MRF for each data vector. This partition function depends on the
490
Z Ghahramani and G. E. Hinton
top-down inputs to a layer so it varies from one data vector to the next, even if the
lateral connections themselves are non-adaptive . Fortunately, since both the MRF
and the top-down prediction define Gaussians over the states of the units in a layer,
these derivatives can be easily calculated. Assuming unit variances,
tlYj; = , ([Yj]+(Y; -
ii;) + [Yj]+ ~ [M(I + M)-ll;. ii.)
(7)
where M is the MRF matrix for the layer including units i and k, and I is the identity
matrix. The first term is the delta rule (Eq. 3); the second term is the derivative
of the partition function which unfortunately involves a matrix inversion. Since
the partition function for a multivariate Gaussian is analytical it is also possible to
learn the lateral connections in the MRF.
Lateral interactions between the rectified states of units add the quadratic term
~ Lk Ll Mkl [Yk]+[YzJ+? The partition function is no longer analytical, so computing the gradient of the likelihood involves a two-phase Boltzmann-like procedure:
!19ji =
f
([Yj]+Yi) * - ([Yj]+Yi
r) ,
(8)
where 0* averages with respect to the posterior distribution of Yi and Yj, and 0averages with respect to the posterior distribution of Yj and the prior of Yi given
units in the same layer as j. This learning rule suffers from all the problems of
the Boltzmann machine, namely it is slow and requires two-phases. However, there
is an approximation which results in the familiar one-phase delta rule that can
be described in three equivalent ways: (1) it treats the lateral connections in the
generative model as if they were additional lateral connections in the recognition
model; (2) instead of lateral connections in the generative model it assumes some
fictitious children with clamped values which affect inference but whose likelihood
is not maximised during learning; (3) it maximises a penalized likelihood of the
model without the lateral connections in the generative model.
5
Discovering depth in simplified stereograms
Consider the following generative process for stereo pairs. Random dots of uniformly
distributed intensities are scattered sparsely on a one-dimensional surface, and the
image is blurred with a Gaussian filter. This surface is then randomly placed at one
of two different depths, giving rise to two possible left-to-right disparities between
the images seen by each eye. Separate Gaussian noise is then added to the image
seen by each eye. Some images generated in this manner are shown in Fig. 3a.
Figure 3: a) Sample data from the stereo
disparity problem. The left and right column
of each 2 x 32 image are the inputs to the left
and right eye, respectively. Periodic boundary conditions were used. The value of a pixel
is represented by the size of the square, with
white being positive and black being negative. Notice that pixel noise makes it difficult
to infer the disparity, i.e. the vertical shift
between the left and right columns, in some
images. b) Sample images generated by the
model after learning.
We trained a three-layer RGBN consisting of 64 visible units, 64 units in the first
hidden layer and 1 unit in the second hidden layer on the 32-pixel wide stereo
Hierarchical Non-linear Factor Analysis and Topographic Maps
491
disparity problem. Each of the hidden units in the first hidden layer was connected
to the entire array of visible units, i.e. it had inputs from both eyes. The hidden
units in this layer were also laterally connected in an MRF over the unrectified
units. Nearby units excited each other and more distant units inhibited each other,
with the net pattern of excitation/inhibition being a difference of two Gaussians.
This MRF was initialised with large weights which decayed exponentially to zero
over the course of training. The network was trained for 30 passes through a data
set of 2000 images. For each image we used 16 iterations of Gibbs sampling to
approximate the posterior distribution over hidden states. Each iteration consisted
of sampling every hidden unit once in a random order. The states after the fourth
iteration of Gibbs sampling were used for learning, with a learning rate of 0.05 and
a weight decay parameter of 0.001. Since the top level of the generative process
makes a discrete decision between left and right global disparity we used a trivial
extension of the RGBN in which the top level unit saturates both at 0 and 1.
a
__
IEI[I:I_1II_-=-_.-:rr::JI...___I:IIUI::JI-L1D-.--:tIl::Jl-=-::l .-'-'
_______
OW''--o--.,u'-'-''__=_-..._.-'-"._
._--="TI:~?:I=-[J
b
c
Figure 4: Generative weights of a three-layered RGBN after being trained on the stereo
disparity problem. a) Weights from the top layer hidden unit to the 64 middle-layer hidden
units. b) Biases of the middle-layer hidden units, and c) weights from the hidden units to
the 2 x 32 visible array.
Thirty-two of the hidden units learned to become local left-disparity detectors, while
the other 32 became local right-disparity detectors (Fig. 4c). The unit in the second
hidden layer learned positive weights to the left-disparity detectors in the layer
below, and negative weights to the right detectors (Fig. 4a). In fact, the activity
of this top unit discriminated the true global disparity of the input images with
99% accuracy. A random sample of images generated by the model after learning is
shown in Fig. 3b. In addition to forming a hierarchical distributed representation
of disparity, units in the hidden layer self-organised into a topographic map. The
MRF caused high correlations between nearby units early in learning, which in
turn resulted in nearby units learning similar weight vectors. The emergence of
topography depended on the strength of the MRF and on the speed with which it
decayed. Results were relatively insensitive to other parametric changes.
We also presented image patches taken from natural images [1] to a network with
units in the first hidden layer arranged in laterally-connected 2D grid. The network
developed local feature detectors, with nearby units responding to similar features
(Fig. 5). Not all units were used, but the unused units all clustered into one area.
6
Discussion
Classical models of topography formation such as Kohonen's self-organising map [6]
and the elastic net [2, 4] can be thought of as variations on mixture models where
additional constraints have been placed to encourage neighboring hidden units to
have similar generative weights . The problem with a mixture model is that it cannot
handle images in which there are several things going on at once. In contrast, we
492
Z. Ghahramani and G. E. Hinton
Figure 5: Generative weights of an
RGBN trained on 12 x 12 natural
image patches: weights from each
of the 100 hidden units which were
arranged in a 10 x 10 sheet with
toroidal boundary conclitions.
have shown that topography can arise in much richer hierarchical and componential
generative models by inducing correlations between neighboring units.
There is a sense in which topography is a necessary consequence of the lateral
connection trick used for perceptual inference. It is infeasible to interconnect all
pairs of units in a cortical area. If we assume that direct lateral interactions (or
interactions mediated by interneurons) are primarily local, then widely separated
units will not have the apparatus required for explaining away. Consequently the
computation of the posterior distribution will be incorrect unless the generative
weight vectors of widely separated units are orthogonal. If the generative weights
are constrained to be positive, the only way two vectors can be orthogonal is for
each to have zeros wherever the other has non-zeros. Since the redundancies that
the hidden units are trying to model are typically spatially localised, it follows
that widely separated units must attend to different parts of the image and units
can only attend to overlapping patches if they are laterally interconnected. The
lateral connections in the generative model assist in the formation of the topography
required for correct perceptual inference.
Acknowledgements. We thank P. Dayan, B. Frey, G. Goodhill, D. MacKay, R. Neal
and M. Revow. The research was funded by NSERC and ITRC. GEH is the Nesbitt-Burns
fellow of CIAR.
References
[1] A. Bell & T. J. Sejnowski. The 'Independent components' of natural scenes are edge
filters. Vision Research, In Press .
[2] R. Durbin & D. Willshaw. An analogue approach to the travelling salesman problem
using an elastic net method. Nature, 326(16):689-691, 1987.
[3] Z. Ghahramani & G. E. Hinton. The EM algorithm for mixtures of factor analyzers.
Univ. Toronto Technical Report CRG-TR-96-1, 1996.
[4] G. J . Goodhill & D. J . Willshaw. Application of the elatic net algorithm to the
formation of ocular dominance stripes. Network: Compo in Neur. Sys ., 1:41-59, 1990.
[5] G. E. Hinton & Z. Ghahramani. Generative models for cliscovering sparse clistributed
representations. Philos. Trans. Roy. Soc . B, 352:1177-1190, 1997.
[6] T. Kohonen. Self-organized formation of topologically correct feature maps. Biological
Cybernetics, 43:59-69, 1982.
[7] D. D. Lee & H. S. Seung. Unsupervised learning by convex and conic cocling. In
M. Mozer, M. Jordan, & T. Petsche, eds., NIPS 9. MIT Press, Cambridge, MA, 1997.
[8] M. S. Lewicki & T. J. Sejnowski. Bayesian unsupervised learning of higher order
structure. In NIPS 9. MIT Press, Cambridge, MA, 1997.
[9] R. M. Neal. Connectionist learning of belief networks. Arti/. Intell., 56:71-113, 1992.
[10] R. M. Neal & G. E. Hinton. A new view of the EM algorithm that justifies incremental
and other variants. Unpublished Manuscript, 1993.
| 1472 |@word middle:2 inversion:1 nd:1 simulation:1 excited:1 arti:1 tr:1 solid:1 contains:2 disparity:12 selecting:2 favouring:1 activation:1 must:2 distant:1 visible:3 partition:5 generative:27 selected:1 discovering:1 maximised:1 sys:1 compo:1 contribute:3 toronto:5 organising:1 direct:1 become:1 incorrect:1 consists:1 manner:2 themselves:1 considering:1 ocs:1 linearity:1 mass:2 developed:1 fellow:1 every:1 ti:1 exactly:4 laterally:3 toroidal:1 willshaw:2 unit:72 positive:5 maximise:1 attend:2 local:9 treat:1 apparatus:1 depended:1 consequence:1 frey:1 black:1 burn:1 thirty:1 yj:28 practice:1 implement:1 procedure:1 area:2 bell:1 thought:1 zoubin:2 get:1 cannot:1 clever:1 layered:3 sheet:1 www:1 equivalent:1 map:9 send:1 convex:1 pure:1 rule:10 adjusts:1 array:2 handle:3 variation:1 suppose:1 us:1 trick:2 roy:1 recognition:6 stripe:1 sparsely:1 bottom:2 observed:1 module:4 role:1 capture:1 connected:3 yk:5 mozer:1 stereograms:2 seung:3 trained:4 depend:1 segment:1 topographically:1 purely:2 creates:1 basis:2 easily:1 represented:1 separated:3 univ:1 describe:2 effective:1 sejnowski:2 analyser:2 formation:5 whose:2 richer:1 widely:3 valued:3 plausible:2 otherwise:1 topographic:7 think:1 emergence:1 itself:1 online:2 advantage:1 rr:1 net:8 analytical:2 interaction:6 product:1 interconnected:1 kohonen:2 neighboring:2 ontario:1 inducing:1 parent:1 generating:1 incremental:1 object:4 fixing:1 pose:1 eq:2 soc:1 implemented:2 c:1 involves:3 correct:3 filter:2 require:1 fix:1 clustered:1 biological:1 crg:1 extension:1 hold:1 ic:1 equilibrium:4 driving:1 major:2 early:1 omitted:1 mit:2 gaussian:20 probabilistically:1 believing:1 likelihood:3 contrast:1 sense:1 inference:10 itrc:1 dayan:1 interconnect:1 lj:1 entire:1 typically:1 hidden:20 going:1 pixel:3 priori:1 constrained:1 mackay:1 equal:4 once:3 field:1 sampling:16 unsupervised:2 report:1 connectionist:1 develops:1 piecewise:1 inhibited:1 primarily:1 randomly:1 divergence:1 intell:1 resulted:1 familiar:1 replaced:1 phase:3 consisting:1 attempt:1 interneurons:1 mixture:6 edge:1 encourage:1 necessary:1 orthogonal:2 unless:1 deformation:1 theoretical:1 column:2 disadvantage:2 deviation:1 subset:3 inadequate:2 varies:1 corrupted:1 periodic:1 combined:1 density:3 decayed:2 probabilistic:2 lee:2 quickly:1 squared:1 stochastically:1 derivative:3 til:1 account:1 iei:1 blurred:1 caused:1 depends:2 performed:1 view:2 contribution:2 square:1 accuracy:1 became:1 variance:6 bayesian:1 l1d:1 produced:1 rectified:8 cybernetics:1 drive:1 m5s:1 liebler:1 unaffected:1 detector:7 implausible:1 reach:2 suffers:1 ed:1 energy:8 initialised:1 ocular:1 proved:1 organized:1 appears:1 manuscript:1 higher:2 follow:1 arranged:2 correlation:3 nonlinear:1 overlapping:1 mkl:6 effect:6 requiring:1 true:2 consisted:1 spatially:1 symmetric:2 neal:4 white:1 ll:2 during:3 self:3 excitation:1 trying:1 performs:1 image:18 consideration:1 sigmoid:1 ji:3 discriminated:1 insensitive:2 exponentially:1 jl:1 componential:3 cambridge:2 gibbs:14 philos:1 grid:1 similarly:1 nonlinearity:1 analyzer:1 had:1 dot:3 funded:1 longer:1 surface:2 inhibition:1 add:1 multivariate:2 posterior:8 own:1 driven:1 yi:5 seen:2 additional:3 fortunately:1 determine:1 redundant:1 dashed:1 ii:2 full:1 multiple:4 neurally:1 desirable:1 infer:1 hebbian:2 technical:1 plausibility:1 ebe:3 long:2 dept:1 controlled:1 schematic:1 prediction:3 mrf:9 variant:1 multilayer:1 vision:1 expectation:1 iteration:3 represent:2 achieved:1 addition:1 want:1 source:1 eliminates:1 pass:1 thing:2 jordan:1 unused:1 intermediate:1 enough:1 affect:1 reduce:1 idea:1 shift:1 ciar:1 expression:1 assist:1 penalty:2 stereo:4 cause:2 jj:1 adequate:1 locally:3 concentrated:1 http:1 goj:2 notice:1 delta:5 per:1 correctly:2 discrete:1 geh:1 dominance:1 key:1 redundancy:1 tenable:1 run:1 powerful:1 fourth:1 topologically:1 patch:4 decision:1 scaling:1 layer:42 quadratic:7 durbin:1 activity:2 strength:1 constraint:1 scene:2 nearby:5 aspect:1 speed:1 argument:1 relatively:1 according:1 neur:1 combination:3 em:2 biologically:1 wherever:1 taken:1 rectification:2 agree:1 describing:1 turn:3 fail:1 know:2 tractable:2 travelling:1 salesman:1 available:3 gaussians:2 hierarchical:8 away:4 appropriate:2 petsche:1 top:15 assumes:2 responding:1 giving:1 ghahramani:6 classical:1 added:2 spike:2 parametric:1 gradient:3 ow:4 separate:1 thank:1 lateral:25 trivial:1 reason:2 rjk:2 assuming:2 code:1 index:1 difficult:2 unfortunately:2 localised:2 negative:4 rise:1 stern:1 boltzmann:2 perform:5 maximises:1 vertical:1 neuron:1 markov:1 unrectified:9 anti:1 hinton:8 saturates:1 canada:1 intensity:1 introduced:1 namely:1 pair:2 required:2 unpublished:1 connection:22 learned:5 nip:2 trans:1 below:9 pattern:1 goodhill:2 regime:1 including:2 belief:7 analogue:1 natural:4 difficulty:1 eye:4 conic:1 lk:1 mediated:1 extract:1 kj:1 prior:1 acknowledgement:1 determining:1 topography:5 suggestion:1 organised:2 fictitious:1 proven:1 geoffrey:1 consistent:1 uncorrelated:1 course:1 penalized:2 placed:2 infeasible:1 bias:3 allow:3 fall:1 explaining:4 wide:1 sparse:2 distributed:4 boundary:2 depth:3 calculated:1 cortical:1 adaptive:1 simplified:2 far:1 approximate:1 kullback:1 global:2 assumed:1 consuming:1 alternatively:1 bottomup:1 why:1 learn:4 nature:1 rearranging:1 elastic:2 ignoring:1 main:2 dense:2 noise:10 arise:1 child:1 fig:9 scattered:1 slow:1 fails:1 lyj:1 clamped:1 perceptual:8 ib:1 down:11 specific:1 showing:1 decay:1 ih:1 adding:1 justifies:1 simply:2 infinitely:2 forming:1 nserc:1 lewicki:1 radford:1 corresponds:2 ma:2 viewed:1 identity:1 consequently:1 revow:1 change:1 infinite:1 generalisation:2 typical:1 determined:1 uniformly:1 total:1 la:1 select:1 incorporate:1 h5:1 phenomenon:1 correlated:1 |
517 | 1,473 | Linear concepts and hidden variables:
An empirical study
Adam J. Grove
Dan Rothe
NEC Research Institute
4 Independence Way
Princeton NJ 08540
grove@research.nj.nec.com
Department of Computer Science
University of Illinois at Urbana-Champaign
1304 W. Springfield Ave. Urbana 61801
danr@cs.uiuc.edu
Abstract
Some learning techniques for classification tasks work indirectly, by first trying
to fit a full probabilistic model to the observed data. Whether this is a good idea
or not depends on the robustness with respect to deviations from the postulated
model. We study this question experimentally in a restricted, yet non-trivial and
interesting case: we consider a conditionally independent attribute (CIA) model
which postulates a single binary-valued hidden variable z on which all other
attributes (i.e., the target and the observables) depend. In this model, finding the
most likely value of anyone variable (given known values for the others) reduces
to testing a linear function of the observed values.
We learn CIA with two techniques: the standard EM algorithm, and a new
algorithm we develop based on covariances. We compare these, in a controlled
fashion, against an algorithm (a version of Winnow) that attempts to find a good
linear classifier directly. Our conclusions help delimit the fragility of using the
CIA model for classification: once the data departs from this model, performance
quickly degrades and drops below that of the directly-learned linear classifier.
1 Introduction
We consider the classic task of predicting a binary (0/1) target variable zo, based on
the values of some n other binary variables ZI ??? Zft,. We can distinguish between two
styles of learning approach for such tasks. Parametric algorithms postulate some form of
probabilistic model underlying the data, and try to fit the model's parameters. To classify an
example we can compute the conditional probability distribution for Zo given the values of
the known variables, and then predict the most probable value. Non-parametric algorithms
do not assume that the training data has a particular form. They instead search directly in
the space of possible classification functions, attempting to find one with small error on the
training set of examples.
An important advantage of parametric approaches is that the induced model can be used to
support a wide range of inferences, aside from the specified classification task. On the other
hand, to postulate a particular form of probabilistic model can be a very strong assumption.
"Partly supported by ONR grant NOOOI4-96-1-0550 while visiting Harvard University.
Linear Concepts and Hidden Variables: An Empirical Study
501
So it is important to understand how robust such methods are when the real world deviates
from the assumed model.
In this paper, we report on some experiments that test this issue. We consider the specific
case of n + 1 conditionally independent attributes Zi together with a single unobserved
variable z, also assumed to be binary valued, on which the Zi depend (henceforth, the
binary CIA model); see Section 2. In fact, such models are plausible in many domains
(for instance, in some language interpretation tasks; see [GR96]). We fit the parameters of
the CIA model using the well-known expectation-maximization (EM) technique [DLR77],
and also with a new algorithm we have developed based on estimating covariances; see
Section 4. In the nonparametric case, we simply search for a good linear separator. This is
because the optimal predictors for the binary CIA model (i.e., for predicting one variable
given known values for the rest) are also linear. This means that our comparison is "fair"
in the sense that neither strategy can choose from classifiers with more expressive power
than the other. As a representative of the non-parametric class of algorithms, we use the
Winnow algorithm of [Lit881, with some modifications (see Section 6). Winnow works
directly to find a "good" linear separator. It is guaranteed to find a perfect separator if one
exists, and empirically seems to be fairly successful even when there is no perfect separator
[GR96, Blu9?]. It is also very fast.
Our experimental methodology is to first generate synthetic data from a true CIA model
and test performance; we then study various deviations from the model. There are various
interesting issues involved in constructing good experiments, including the desirability of
controlling the inherent "difficulty" of learning a model. Since we cannot characterize the
entire space, we consider here only deviations in which the data is drawn from a CIA model
in which the hidden variable can take more than two values. (Note that the optimal classifier
given Zo is generally not linear in this case.)
Our observations are not qualitatively surprising. CIA does well when the assumed model
is correct, but performance degrades when the world departs from the model. But as we
discuss, we found it surprising how fragile this model can sometimes be, when compared
against algorithms such as Winnow. This is even though the data is not linearly separable
either, and so one might expect the direct learning techniques to degrade in performance as
well. But it seems that Wmnow and related approaches are far less fragile. Thus the main
contribution of this work is that our results shed light on the specific tradeoff between fitting
parameters to a probabilistic model, versus direct search for a good classifier. Specifically,
they illustrate the dangers of predicting using a model that is even "slightly" simpler than
the distribution actually generating the data, vs. the relative robustness of directly searching
for a good predictor. This would seem to be an important practical issue, and highlights the
need for some better theoretical understanding of the notion of "robustness".
2 Conditionally Independent Attributes
Throughout we assume that each example is a binary vector z E {O, 1}n+l, and that each
example is generated independently at random according to some unknown distribution on
{O, 1}n+l. We use Xi to denote the i'th attribute, considered as a random variable, and Zi
to denote a value for Xi. In the conditionally independent attribute (CIA) model, examples
are generated as follows. We postulate a "hidden" variable Z with Ie values, which takes
values z for 0 $ z < Ie with probability a. ~ O. Since we must have E::~ a.
1
there are Ie - 1 independent parameters. Having randomly chosen a value z for the hidden
variable, we choose the value Zi for each observable Xi: the value is 1 with probability
p~.}, and 0 otherwise. Here p~.} E [0,1). The attributes' values are chosen independently
of each other, although z remains fixed. Note that there are thus (n + 1)1e probability
parameters p~.). In the following, let l' denote the set of all (n + 1)1e + Ie - 1 parameters
in the model. From this point, and until Section 7, we always assume that Ie = 2 and in this
case, to simplify notation, we write al as a, ao (= 1 - a) as ai, p! as Pi and p~ as qi.
=
A. 1. Grove and D. Roth
502
3
The Expectation-Maximization algorithm (EM)
One traditional unsupervised approach to learning the parameters of this model is to find
the maximum-likelihood parameters of the distribution given the data. That is, we attempt
to find the set of parameters that maximizes the probability of the data observed.
Finding the maximum likelihood parameterization analytically appears to be a difficult
problem, even in this rather simple setting. However, a practical approach is to use the wellknown Expectation-Maximization algorithm (EM) [DLR77], which is an iterative approach
that always converges to a local maximum of the likelihood function. In our setting, the
procedure is as follows. We simply begin with a randomly chosen parameterization p, and
then we iterate until (apparent) convergence: 1
Expectation: For all zi, compute Ui = p-p(zi 1\ Z = 1) and Vi = p-p(zi 1\ Z = 0).
=
=
Maximization: Reestimate P as follows (writing U Ei Ui and V
Ei Vi):
a f- E:=I Ui/(U + V) P; f- E{i::i~=I} u;./U qj f- E{i::i~=I} Vi/V.
After convergence has been detected all we kno'w is that we are near a [ocdi minima of the
likelihood function. Thus it is prudent to repeat the process with many different restarts.
(All our experiments were extremely conservative concerning the stopping criteria at each
iteration, and in the number of iterations we tried.) But in practice, we are never sure that
the true optimum has been located.
4
Covariances-Based approach
Partly in response to concern just expressed, we also developed another heuristic technique
for learning P. The algorithm, which we call COY, is based on measuring the covariance
between pairs of attributes. Since we do not see Z, attributes will appear to be correlated. In
fact, if the CIA model is correct, it is easy to show that covariance between Xi and X j (defined as Yi,; = ~,; - ~I-'; where~, 1-';, ~,; are the expectations of Xi, Xj, (Xi and Xj),
respectively), will be Yi,j
aa'did; where di denotes Pi - qi. We also know that the
expected value of Xi is ~ = aPi + a'qi. Furthermore, we will be able to get very accurate
estimates of ~ just by observing the proportion of samples in which Zi is 1. Thus, if we
could estimate both a and di it would be trivial to solve for estimates of Pi and qi.
To estimate di, suppose we have computed all the pairwise covariances using the data;
we use fli,; to denote our estimate of Yi,j' For any distinct j, Ie i= i we clearly have
aa l = IV'rd.r"/o1 so we could estimate d; using this equation. A better estimate would be
"/o
to consider all pairs j, Ie and average the individual estimates. However, not all individual
estimates are equally good. It can be shown that the smaller Y;,II is, the less reliable
we should expect the estimate to be (and in the limit, where X; and XII are perfectly
uncorrelated, we get no valid estimate at all). This suggests that we use a weighted average,
with the weights proportional to Yj,II. Using these weights leads us to the next equation for
determining 5i , which, after simplification, is:
=
5;
E j ,II:j;t1l;ti IYi,jYi,II I
(E;:#i IYi,; 1)2 - E;:;;ti if,;
E;,II:;;tll;ti IY;,II I
E;,II:;;tll IY;,II I - 2 Ej:j;ti IYj,i I
dl.
By substituting the estimates 'Oi,; we get an estimate for aa'
This estimate can be
computed in linear time except for the determination of Ej,II:j;tll IYj,II I which, although
quadratic, does not depend on i and so can be computed once and for all. Thus it takes
O(n2) time in total to estimate aa'd; for all i.
It remains only to estimate a and the signs of the di'S. Briefly, to determine the signs we first
stipulate that do is positive. (Because we never see z, one sign can be chosen at random.)
IThe maximization phase works as though we were estimating parameters by taking averages
based on weighted labeled data (Le., in which we see z). If i i is a sample point, these fictional data
1) with weight Ui/U and (ii, z 0) with weight Vi/V.
points are (ii,Z
=
=
Linear Concepts and Hidden Variables: An Empirical Study
503
In principle, then, the sign of 0; will then be equal to the sign of Yo,;, which we have an
estimate for. In practice, this can statistically unreliable for small sample sizes and so we
use a more involved ''voting'' procedure (details omitted here). Finally we estimate Q. We
have found no better method of doing this than to simply search for the optimal value, using
likelihood as the search criterion. However, this is only a I-dimensional search and it turns
out to be quite efficient in practice.
5
Linear Separators and CIA
Given a fully parameterized CIA model, we may be interested in predicting the value of
one variable, say Xo, given known values for the remaining variables. One can show that
in fact the optimal prediction region is given by a linear separator in the other variables,
although we omit details of this derivationhere. 2 This suggest an obvious learning strategy:
simply try to find the line which minimizes this loss on the training set. Unfortunately, in
general the task of finding a linear separator that minimizes disagreements on a collection
of examples is known to be NP-hard [HS92]. So instead we use an algorithm called Winnow
that is known to produce good results when a linear separator exists, as well as under certain
more relaxed assumptions [Lit9I], and appears to be quite effective in practice.
6 Learning using a Winnow-based algorithm
The basic version of the Winnow algorithm [Lit88] keeps an n-dimensional vector w =
(1011" .1On ) of positive weights (Le.,
is the weight associated with the ith feature),
which it updates whenever a mistake is made. Initially, the weight vector is typically
set to assign equal positive weight to all features. The algorithm has 3 parameters, a
promotion parameter Q > I, a demotion parameter 0 < f3 < 1 and a threshold 8. For a
> 8. If the
given instance (:1:1, ? "1 :l: n ) the algorithm predicts that :1:0 = 1 iff E~l
algorithm predicts 0 and the label (Le., :1:0) is 1 (positive example) then the weights which
correspond to active attributes (:1:, = 1) are promoted-the weight 10, is replaced by a larger
weight Q ? Wi. Conversely, if algorithm predicts 1 and the received label is 0, then the
weights which correspond to active features are demoted by factor {3. We allow for negative
weights as follows. Given an example (:1:1" "1 :l: n ), we rewrite it as an example over 2n
variables (Y1, 'Y21 ?.. I 'Y2n) where y, = :1:, and Yn+,
1 - :1:,. We then apply Winnow just
as above to learn 2n (positive) weights. If wi is the weight associated with :1:, and wi is
the weight associated with :l:n+i (Le., 1 - :1:,), then the prediction rule is simply to compare
E~=l(wi:l:, + wi(1 - :1:,)) with the threshold.
w,
W,:I:,
=
In the experiments described here we have made two significant modifications to the basic
algorithm. To reduce variance, our final classifier is a weighted average of several classifiers;
each is trained using a subs ample from the training set, and its weight is based based on how
well it was doing on that sample. Second, we biased the algorithm so as to look for "thick"
classifiers. To understand this, consider the case in which the data is perfectly linearly
separable. Then there will generally be many linear concepts that separate the training data
we actually see. Among these, it seems plausible that we have a better chance of doing
well on the unseen test data if we choose a linear concept that separates the positive and
negative training examples as "widely" as possible. The idea of having a wide separation
is less clear when there is no perfect separator, but we can still appeal to the basic intuition.
To bias the search towards "thick" separators, we change Wmnow's training rule somewhat.
We now have a new margin parameter T. As before, we always update when our current
hypothesis makes a mistake, but now we also update if I E~=l Wi:l:, - 8 I is less than T,
even if the prediction is correct. In our experiments, we found that performance when using
this version of Winnow is better than that of the basic algorithm, so in this paper we present
results for the former.
2A derivation
for the slightly different case, for predicting z, can be found in [MP69J.
A. J Grove and D. Roth
504
7 Experimental Methodology
Aside from the choice of algorithm used, the number of attributes n, and the sample
size 8, our experiments also differed in two other dimensions. These are the type of
process generating the data (we will be interested in various deviations from CIA), and
the "difficulty" of the problem. These features are determined by the data model we use
(i.e., the distribution over {O, I} ft used to generate data sets).
Our first experiments consider the case where the data really is drawn from a binary CIA
distribution. We associated with any such distribution a "difficulty" parameter B, which is
the accuracy with which one could predict the value of Z if one actually knew the correct
model. (Of course, even with knowledge of the correct model we should not expect 100%
accuracy.) The ability to control B allows us to select and study models with different
qualitative characteristics. In particular, this has allowed us concentrated most of our
experiments on fairly "hard" instances 3 , and to more meaningfully compare trials with
differing numbers of attributes. We denote by CIA(n, 2, b) the class of all data models
which are binary CIA distributions over n variables with difficulty b. 4 The next family of
data models we used are also CIA models, but now using more than two values for the
hidden variable. We denote the family using Ie values as CIA(n, Ie, b) where n and b are as
before. When Ie > 2 there are more complex correlation patterns between the Xi than when
Ie 2. Furthermore, the optimal predictor is not necessarily linear. The specific results we
discuss in the next section have concentrated on this case.
Given any set of parameters, including a particular class of data models, our experiments
are designed with the goal of good statistical accuracy. We repeatedly (typically 100 to
300 times) choose a data model at random from the chosen class, choose a sample of the
appropriate size from this model, and then run all our algorithms. Each algorithm produces
a (linear) hypothesis. We measure the success rate Salg (i.e., the proportion of times a
hypothesis makes the correct prediction of :1:0) by drawing yet more random samples from
the data model being used. In the test phase we always draw enough new samples so that
the confidence interval for Salg, for the results on a single model, has width at most ? 1%.
We use the Salg values to construct a normalized measure of performance (denoted T) as
follows. Let Sbest be the best possible accuracy attainable for predicting:l:o (i.e., the accuracy
achieved by the actual model generating the data). Let Sconst denote the performance of
the best possible constant prediction rule (i.e., the rule that predicts the most likely a priori
value for :1:0). Note that Sconst and Sbest can vary from model to model. For each model we
compute :alg--;onst ,and our normalized statistic T is the average of these values. It can be
best- const
thought of as measuring the percentage of the possible predictive power, over a plausible
baseline, that an algorithm achieves.
=
8
Results
We only report on a small, but representative, selection of our experiments in any detail.
For instance, although we have considered many values of n ranging from 10 to 500, here
we show six graphs giving the learning curves for CIA(n, Ie, 0.90) for n = 10,75, and for
Ie = 2,3,5; as noted, we display the T statistic. The error bars show the standard error,s
providing a rough indication of accuracy. Not surprisingly, when the data model is binary
3Note that if one simply chooses parameters of a CIA model independently at random, without
examining the difficulty of the model or adjusting for n, one will get many trivial problems, in which
it is easy to predict Z with nearly 100% accuracy, and thus predict optimally for Xo.
41t is nontrivial to efficiently select random models from this class. Briefly, our scheme is to choose
each parameter in a CIA model independently from a symmetric beta distribution. Thus, the model
parameters will have expected value 0.5. We choose the parameter of the beta distribution (which
determines concentration about 0.5) so that the average B value, of the models thus generated, equals
b. Finally, we use rejection sampling to find CIA models with B values that are exactly b ? 1%.
5Computed as the observed standard deviation, divided by the square root of the number of trials.
Linear Concepts and Hidden Variables: An Empirical Study
505
CIA, the EM algorithm does extremely well, learning significantly (if not overwhelmingly)
faster than Winnow. But as we depart from the binary CIA assumption, the performance of
EM quickly degrades.
CI"(10.2.0.1IO)
,""
Cl"(78.2.0.1IO)
'00
--_..... -. ..
'00
loo
Joo
I ?? z
t 40
J:
..
m
- - oa>I
......,
..
,0
...
'00
.oIT~~
,
,1,1
40
,
-EM
......
00
lOO
~i'.~/
j:
~ _~~! ~.!-,-4?
)......,
-....
- - oa>I
00
'0
}oo
m
- E..
- - oa>I
..
'00
'000
...
CIA(78,',O.1O)
- - oa>I
I""
J. . .
t
0
......,
-
'000
Figure 3: CIA(1O,3,0.9)
,0
.
...
'00
.01 Tralring ~
'000
Figure 4: CIA(75,3,0.9)
CIA(7????o..a)
CIAC10.a,O.1O)
10
. ........
40
l
loo '-----'
,
140
too
I
...
-EM
00
l40
.dT"INng~
'00
.oIT,.......~
? ?" m
00
l
,0
..
'000
.
'00
......
? "m
Figure 2: CIA(75,2,0.9)
OA,(10 ?? ,O.IO)
I""
of ,.?f
....
-EM
Figure 1: CIA(10,2,0.9)
j40
.-
,,
j
20
a
-r'
)-
0
" -1--
......,
-20
..
......,~,o~----~~~,~~~-----=*~~,_
.oIT'. . . . ExemPM
-
_---t--r I
m
...
- E..
- - oa>I
'0
eo
'00
.oIT~~
'ODD
Figure 5: CIA(10,5,0.9)
Figure 6: CIA(75,5,0.9)
When Ie = 3 performances is, on the whole, very similar for Winnow and EM. But when
Ie = 5 Winnow is already superior to EM; significantly and uniformly so for n = 10. For
fixed Ie the difference seems to become somewhat less dramatic as n increases; in Figure 6
(for n = 75) Winnow is less obviously dominant, and in fact is not better than EM until the
sample size has reached 100. (But when 8 ~ n, meaning that we have fewer samples than
attributes, the performance is unifonnly dismal anyway.)
Should we attribute this degradation to the binary CIA assumption, or to the EM itself?
This question is our reason for also considering the covariance algorithm. We see that the
results for COY are generally similar to EM's, supporting our belief that the phenomena
we see are properties inherent to the model rather than to the specific algorithm being used.
Similarly (the results are omitted) we have tried several other algorithms that try to find
good linear separators directly, including the classic Perceptron algorithm [MP69); our
version of Winnow was the best on the experiments we tried and thus we conjecture that its
performance is (somewhat) indicative of what is possible for any such approach.
As the comparison between n
10 and n = 75 illustrates, there is little qualitative differ-
=
506
A. J. Grove and D. Roth
ence between the phenomena observed as the number of attributes increases. Nevertheless,
as n grows it does seem that Winnow needs more examples before its performance surpasses
that of the other algorithms (for any fixed k). As already noted, this may be due simply to
the very "noisy" nature of the region 8 $ n. We also have reasons to believe that this is
partially an artifact of way we select models.
As previously noted, we also experimented with varying "difficulty" (B) levels. Although
we omit the corresponding figures we mentioned that the main difference is that Winnow is a
little faster in surpassing EM when the data deviates from the assumed model, but when the
data model really is binary CIA, and EM converge even faster to an optimal performance.
These patterns were confinned when we tried to compare the approaches on real data. We
have used data that originates from a problem in which assuming a hidden "context" variable
seems somewhat plausible. The data is taken from the context-sensitive spelling correction
domain. We used one data set from those that were used in [GR96]. For example, given
sentences in which the word passed or past appear, the task is to determine, for each such
occurrence, which of the two it should be. This task may be modeled by thinking of the
"context" as a hidden variable in our sense. Yet when we tried to learn in this case under
the CIA model, with a binary valued hidden variable, the results were no better than just
predicting the most likely classification (around 70%). Winnow, in contrast, performed
extremely well and exceeds 95% on this task. We hesitate to read much into our limited
real-data experiments, other than to note that so far they are consistent witli the more careful
experiments on synthetic data.
9 Conclusion
By restricting to a binary hidden variable, we have been able to consider a "fair" comparison
between probabilistic model construction, and more traditional algorithms that directly
learn a classification-at least in the sense that both have the same expressive power. Our
conclusions concerning the fragility of the former should not be surprising but we believe
that given the importance of the problem it is valuable to have some idea of the true
significance of the effect. As we have indicated, in many real-world cases, where a model
of the sort we have considered here seems plausible, it is impossible to nail down more
specific characterizations of the probabilistic model. Our results exhibit how important it
is to use the correct model and how sensitive are the results to deviations from it, when
attempting to learn using model construction. The purpose of this paper is not to advocate
that in practice one should use either Winnow or binary CIA in exactly the form considered
here. A richer probabilistic model should be used along with a model selection phase.
However, studying the problem in a restricted and controlled environment in crucial so as
to understand the nature and significance of this fundamental problem.
References
[Blu97] A. Blum. Empirical support for winnow and weighted majority based algorithms: results on
a calendar scheduling domain. Machine Learning, 26: 1-19, 1997.
[DLR77] A. P. Dempster, N. M. Laird, and D. B. Rubin. Maximum likelihood from incomplete data
via the EM algorithm. Royal Statistical SOCiety B, 39: 1-38, 1977.
[GR96] A. R. Golding and D. Roth. Applying winnow to context-sensitive spelling correcton. In
Proc. 13th International Conference on Machine Learning (ML' 96), pages 182-190, 1996.
[HS92] K. HOffgen and H. Simon. Robust trainability of single neurons. In Proc. 5th Annu. Workshop
on Comput. Learning Theory, pages 428-439, New York, New York, 1992. ACM Press.
[Lit88] N. Littlestone. Learning quickly when irrelevant attributes abound: A new linear-threshold
algorithm. Machine Learning, 2:285-318,1988.
[Lit91] N. Littlestone. Redundant noisy attributes, attribute errors, and linear threshold learning
using Winnow. In Proc. 4th Annu. Workshop on Comput. Learning Theory, pages 147-156, San
Mateo, CA, 1991. Morgan Kaufmann.
[MP69] M. L. Minsky and S. A. Papert. Perceptrons. MIT Press, Cambridge, MA, 1969.
| 1473 |@word trial:2 briefly:2 version:4 hoffgen:1 seems:6 proportion:2 tried:5 covariance:7 attainable:1 dramatic:1 past:1 current:1 com:1 surprising:3 yet:3 must:1 drop:1 designed:1 update:3 aside:2 v:1 fewer:1 parameterization:2 indicative:1 ith:1 characterization:1 simpler:1 demoted:1 along:1 direct:2 beta:2 become:1 qualitative:2 dan:1 fitting:1 advocate:1 pairwise:1 expected:2 uiuc:1 actual:1 little:2 considering:1 abound:1 begin:1 estimating:2 underlying:1 notation:1 maximizes:1 what:1 minimizes:2 developed:2 differing:1 finding:3 unobserved:1 nj:2 ti:4 voting:1 shed:1 exactly:2 classifier:8 control:1 originates:1 grant:1 omit:2 appear:2 yn:1 positive:6 before:3 local:1 limit:1 mistake:2 api:1 io:3 might:1 mateo:1 suggests:1 conversely:1 confinned:1 limited:1 range:1 statistically:1 practical:2 testing:1 yj:1 practice:5 procedure:2 danger:1 empirical:5 thought:1 significantly:2 confidence:1 word:1 suggest:1 get:4 cannot:1 selection:2 scheduling:1 context:4 impossible:1 writing:1 applying:1 roth:4 delimit:1 independently:4 rule:4 classic:2 searching:1 notion:1 anyway:1 target:2 controlling:1 suppose:1 construction:2 hypothesis:3 harvard:1 located:1 predicts:4 labeled:1 observed:5 ft:1 region:2 valuable:1 mentioned:1 intuition:1 environment:1 dempster:1 ui:4 trained:1 depend:3 rewrite:1 ithe:1 predictive:1 observables:1 various:3 derivation:1 zo:3 distinct:1 fast:1 effective:1 detected:1 apparent:1 heuristic:1 quite:2 valued:3 plausible:5 solve:1 say:1 otherwise:1 larger:1 widely:1 ability:1 drawing:1 statistic:2 unseen:1 calendar:1 itself:1 noisy:2 final:1 laird:1 obviously:1 advantage:1 indication:1 stipulate:1 iff:1 convergence:2 optimum:1 produce:2 generating:3 adam:1 perfect:3 converges:1 help:1 illustrate:1 develop:1 oo:1 odd:1 received:1 strong:1 c:1 differ:1 thick:2 correct:7 attribute:18 kno:1 assign:1 ao:1 really:2 probable:1 correction:1 around:1 considered:4 predict:4 substituting:1 vary:1 achieves:1 omitted:2 purpose:1 proc:3 label:2 sensitive:3 weighted:4 mit:1 promotion:1 clearly:1 rough:1 always:4 desirability:1 rather:2 ej:2 varying:1 tll:3 overwhelmingly:1 yo:1 likelihood:6 contrast:1 ave:1 baseline:1 sense:3 inference:1 stopping:1 entire:1 typically:2 initially:1 hidden:13 springfield:1 interested:2 issue:3 classification:6 among:1 y2n:1 prudent:1 denoted:1 priori:1 fairly:2 equal:3 once:2 never:2 having:2 f3:1 construct:1 sampling:1 look:1 unsupervised:1 nearly:1 thinking:1 others:1 report:2 simplify:1 inherent:2 np:1 randomly:2 individual:2 replaced:1 phase:3 minsky:1 attempt:2 danr:1 light:1 accurate:1 grove:5 unifonnly:1 iv:1 incomplete:1 littlestone:2 theoretical:1 instance:4 classify:1 ence:1 measuring:2 maximization:5 deviation:6 surpasses:1 predictor:3 successful:1 examining:1 too:1 loo:3 characterize:1 optimally:1 synthetic:2 chooses:1 fundamental:1 international:1 ie:16 probabilistic:7 together:1 quickly:3 iy:2 postulate:4 choose:7 henceforth:1 iyj:2 style:1 postulated:1 lit91:1 depends:1 vi:4 performed:1 try:3 root:1 observing:1 doing:3 reached:1 sort:1 simon:1 contribution:1 square:1 oi:1 accuracy:7 variance:1 characteristic:1 efficiently:1 kaufmann:1 correspond:2 reestimate:1 whenever:1 against:2 involved:2 obvious:1 associated:4 di:4 adjusting:1 noooi4:1 knowledge:1 actually:3 appears:2 dt:1 restarts:1 methodology:2 response:1 though:2 furthermore:2 just:4 until:3 correlation:1 hand:1 ei:2 expressive:2 artifact:1 indicated:1 grows:1 believe:2 effect:1 concept:6 true:3 normalized:2 former:2 analytically:1 sbest:2 read:1 symmetric:1 conditionally:4 width:1 noted:3 criterion:2 trying:1 ranging:1 meaning:1 superior:1 empirically:1 interpretation:1 dismal:1 surpassing:1 significant:1 cambridge:1 ai:1 rd:1 similarly:1 illinois:1 language:1 joo:1 fictional:1 iyi:2 dominant:1 winnow:21 irrelevant:1 wellknown:1 certain:1 binary:16 onr:1 success:1 yi:3 morgan:1 minimum:1 relaxed:1 promoted:1 somewhat:4 eo:1 determine:2 converge:1 redundant:1 ii:12 full:1 reduces:1 champaign:1 exceeds:1 faster:3 determination:1 divided:1 concerning:2 equally:1 controlled:2 qi:4 prediction:5 basic:4 y21:1 expectation:5 iteration:2 sometimes:1 achieved:1 hesitate:1 interval:1 crucial:1 biased:1 rest:1 jyi:1 sure:1 induced:1 ample:1 meaningfully:1 seem:2 call:1 near:1 easy:2 enough:1 iterate:1 independence:1 fit:3 zi:9 xj:2 perfectly:2 reduce:1 idea:3 tradeoff:1 oit:4 fragile:2 qj:1 golding:1 whether:1 six:1 passed:1 york:2 repeatedly:1 generally:3 clear:1 nonparametric:1 concentrated:2 generate:2 percentage:1 sign:5 xii:1 write:1 demotion:1 threshold:4 nevertheless:1 blum:1 drawn:2 neither:1 graph:1 run:1 parameterized:1 throughout:1 family:2 nail:1 separation:1 draw:1 guaranteed:1 distinguish:1 simplification:1 display:1 quadratic:1 nontrivial:1 anyone:1 extremely:3 attempting:2 separable:2 conjecture:1 department:1 according:1 smaller:1 slightly:2 em:17 wi:6 modification:2 restricted:2 xo:2 taken:1 equation:2 remains:2 previously:1 discus:2 turn:1 know:1 studying:1 apply:1 indirectly:1 disagreement:1 appropriate:1 cia:37 occurrence:1 robustness:3 denotes:1 remaining:1 onst:1 const:1 giving:1 society:1 question:2 already:2 depart:1 degrades:3 parametric:4 strategy:2 concentration:1 traditional:2 spelling:2 visiting:1 exhibit:1 t1l:1 separate:2 oa:6 majority:1 degrade:1 trivial:3 reason:2 assuming:1 o1:1 modeled:1 providing:1 difficult:1 unfortunately:1 negative:2 unknown:1 observation:1 neuron:1 urbana:2 supporting:1 y1:1 pair:2 specified:1 dlr77:3 sentence:1 learned:1 able:2 bar:1 below:1 pattern:2 including:3 reliable:1 royal:1 belief:1 power:3 difficulty:6 predicting:7 scheme:1 deviate:2 understanding:1 determining:1 relative:1 fully:1 expect:3 highlight:1 loss:1 interesting:2 proportional:1 versus:1 lit88:2 rothe:1 consistent:1 rubin:1 principle:1 uncorrelated:1 pi:3 course:1 supported:1 repeat:1 surprisingly:1 bias:1 allow:1 understand:3 perceptron:1 institute:1 wide:2 taking:1 curve:1 dimension:1 world:3 valid:1 qualitatively:1 collection:1 made:2 san:1 far:2 observable:1 unreliable:1 keep:1 ml:1 active:2 assumed:4 knew:1 xi:8 search:7 iterative:1 learn:5 nature:2 fragility:2 robust:2 ca:1 alg:1 cl:1 complex:1 separator:11 constructing:1 domain:3 fli:1 did:1 necessarily:1 main:2 significance:2 linearly:2 whole:1 n2:1 fair:2 allowed:1 representative:2 fashion:1 differed:1 sub:1 papert:1 richer:1 comput:2 down:1 departs:2 annu:2 specific:5 appeal:1 experimented:1 concern:1 dl:1 exists:2 workshop:2 restricting:1 importance:1 ci:1 nec:2 illustrates:1 margin:1 rejection:1 simply:7 likely:3 expressed:1 gr96:4 partially:1 aa:4 chance:1 determines:1 acm:1 ma:1 conditional:1 goal:1 careful:1 towards:1 experimentally:1 hard:2 change:1 specifically:1 except:1 determined:1 uniformly:1 degradation:1 conservative:1 total:1 called:1 partly:2 experimental:2 trainability:1 perceptrons:1 select:3 support:2 princeton:1 phenomenon:2 correlated:1 |
518 | 1,474 | Multiresolution Tangent Distance for
Affine-invariant Classification
Nuno Vasconcelos
Andrew Lippman
MIT Media Laboratory, 20 Ames St, E15-320M,
Cambridge, MA 02139, {nuno,lip }@media.mit.edu
Abstract
The ability to rely on similarity metrics invariant to image transformations is an important issue for image classification tasks such as face or
character recognition. We analyze an invariant metric that has performed
well for the latter - the tangent distance - and study its limitations when
applied to regular images, showing that the most significant among these
(convergence to local minima) can be drastically reduced by computing
the distance in a multiresolution setting. This leads to the multi resolution
tangent distance, which exhibits significantly higher invariance to image transformations, and can be easily combined with robust estimation
procedures.
1 Introduction
Image classification algorithms often rely on distance metrics which are too sensitive to
variations in the imaging environment or set up (e.g. the Euclidean and Hamming distances),
or on metrics which, even though less sensitive to these variations, are application specific
or too expensive from a computational point of view (e.g. deformable templates).
A solution to this problem, combining invariance to image transformations with computational simplicity and general purpose applicability was introduced by Simard et al in [7].
The key idea is that, when subject to spatial transformations, images describe manifolds in a
high dimensional space, and an invariant metric should measure the distance between those
manifolds instead of the distance between other properties of (or features extracted from)
the images themselves. Because these manifolds are complex, minimizing the distance between them is a difficult optimization problem which can, nevertheless, be made tractable
by considering the minimization of the distance between the tangents to the manifolds -the
tangent distance (TO) - instead of that between the manifolds themselves. While it has led
to impressive results for the problem of character recognition [8] , the linear approximation
inherent to the TO is too stringent for regular images, leading to invariance over only a very
narrow range of transformations.
844
N. Vasconcelos and A. Lippman
In this work we embed the distance computation in a multi resolution framework [3],
leading to the multiresolution tangent distance (MRTD). Multiresolution decompositions
are common in the vision literature and have been known to improve the performance of
image registration algorithms by extending the range over which linear approximations
hold [5, 1]. In particular, the MRTD has several appealing properties: 1) maintains
the general purpose nature of the TD; 2) can be easily combined with robust estimation
procedures, exhibiting invariance to moderate non-linear image variations (such as caused
by slight variations in shape or occlusions); 3) is amenable to computationally efficient
screening techniques where bad matches are discarded at low resolutions; and 4) can be
combined with several types of classifiers. Face recognition experiments show that the
MRTD exhibits a significantly extended invariance to image transformations, originating
improvements in recognition accuracy as high as 38%, for the hardest problems considered.
2 The tangent distance
Consider the manifold described by all the possible linear transformations that a pattern
lex) may be subject to
(1)
Tp [lex)] = 1('ljJ(x, p)),
where x are the spatial coordinates over which the pattern is defined, p is the set of
parameters which define the transformation, and 'ljJ is a function typically linear on p, but
not necessarily linear on x. Given two patterns M(x) and N(x), the distance between the
associated manifolds - manifold distance (MD) - is
T(M, N) = min IITq[M(x)] - Tp[N(x)]W.
p,q
(2)
For simplicity, we consider a version of the distance in which only one of the patterns is
subject to a transformation, i.e.
T(M, N)
= min
IIM(x) p
Tp[N(x)]lf,
(3)
but all results can be extended to the two-sided distance. Using the fact that
\7 p Tp[N(x)]
= \7pN('ljJ(x, p)) = \7 p '?(x, p)\7xN('?(x, p)),
(4)
where \7pTp is the gradient of Tp with respect to p, Tp[N(x)] can, for small p, be
approximated by a first order Taylor expansion around the identity transformation
Tp[N(x)] = N(x)
+ (p -
If\7p 'ljJ(x,p)\7 x N(x).
This is equivalent to approximating the manifold by a tangent hyper-plane, and leads to the
TD. Substituting this expression in equation 3, setting the gradient with respect to p to zero,
and solving for p leads to
p
~ [~'VP;6(X' P ) 'Vx N(x) 'V); N(X)'V~;6(x, P)]-' ~ D(x)'Vp;6(x, P l'VxN(x) + I,
(5)
where D(x) = M(x) - N(x). Given this optimal p, the TD between the two patterns
is computed using equations I and 3. The main limitation of this formulation is that it
relies on a first-order Taylor series approximation, which is valid only over a small range
of variation in the parameter vector p .
2.1
Manifold distance via Newton's method
The minimization of the MD of equation 3 can also be performed through Newton's method,
which consists of the iteration
pn+1
= pn _ 0: [\7~ T/p=pn] -I \7 p Tlp=pn
(6)
845
Multiresolution Tangent Distancefor Affine-invariant Classification
where \7 p / and \7~ / are, respectively, the gradient and Hessian of the cost function of
equation 3 with respect to the parameter p,
\7p/ = 2
L
[M(x) - Tp[N(x)]) V'pTp[N(x)]
x
V'~ /
= 2
L
[-V'pTp[N(x)]
\7~Tp[N(x)] + [M(x) -
N(x)]
V'~Tp[N(x)]]
.
x
Disregarding the term which contains second-order derivatives (V'~Tp[N(x)]), choosing
pO
I and Q:
1, using 4, and substituting in 6 leads to equation 5. I.e. the TO
corresponds to a single iteration of the minimization of the MD by a simplified version of
Newton's method, where sec!ond-orderderivatives are disregarded. This reduces the rate of
convergence of Newton's method, and a single iteration may not be enough to achieve the
local minimum, even for simple functions. It is, therefore, possible to achieve improvement
if the iteration described by equation 6 is repeated until convergence.
=
=
3 The multiresolution tangent distance
The iterative minimization of equation 6 suffers from two major drawbacks [2]: 1) it may
require a significant number of iterations for convergence and 2), it can easily get trapped
in local minima. Both these limitations can be, at least partially, avoided by embedding
the computation of the MD in a multiresolution framework, leading to the multiresolution
manifold distance (MRMD). For its computation, the patterns to classify are first subject to
a multiresolution decomposition, and the MD is then iteratively computed for each layer,
using the estimate obtained from the layer above as a starting point,
where, Dl(x) = M(x) - Tp~ [N(x)]. If only one iteration is allowed at each imageresolution, the MRMD becomes the multiresolution extension of the TO, i.e. the multi resolution
tangent distance (MRTO).
To illustrate the benefits of minimization over different scales consider the signal J(t) =
E{;=1 sin(wkt ), and the manifold generated by all its possible translations J'(t,d) =
J(t + d). Figure 1 depicts the multiresolution Gaussian decomposition of J(t), together
with the Euclidean distance to the points on the manifold as a function of the translation
associated with each of them (d). Notice that as the resolution increases, the distance
function has more local minima, and the range of translations over which an initial guess
is guaranteed to lead to convergence to the global minimum (at d = 0) is smaller. I.e.,
at higher resolutions, a better initial estimate is necessary to obtain the same performance
from the minimization algorithm.
Notice also that, since the function to minimize is very smooth at the lowest resolutions,
the minimization will require few iterations at these resolutions if a procedure such as
Newton's method is employed. Furthermore, since the minimum at one resolution is a good
guess for the minimum at the next resolution, the computational effort required to reach
that minimum will also be small. Finally, since a minimum at low resolutions is based on
coarse, or global, information about the function or patterns to be classified, it is likely to
be the global minimum of at least a significant region of the parameter space, if not the true
global minimum.
846
N. Vasconcelos and A. Lippman
?R B .~5ISa {\Z\Z\]
-UJj -F \lJ -t;: Ll
-I~
..::..
..
..
....
. . . . . . ... .. ~ .. . .... .
..:..
?.
??
??
????
..I~
??
??
--"' .
_ .:..
. . .. ...
. . ..
Figure 1: Top: Three scales of the multiresolution decomposition of J(t) . Bottom: Euclidean
distance VS. translation for each scale. Resolution decreases from left to right.
4 Affine-invariant classification
There are many linear transformations which can be used in equation 1. In this work, we
consider manifolds generated by affine transformations
1jJ(x,p)
=[
X
0
y
0
1000]
0 x yIP
= ~(x)p,
(8)
where P is the vector of parameters which characterize the transformation. Taking the
gradient of equation 8 with respect to p. V'p1jJ(x,p) = ~(x)T. using equation 4. and
substituting in equation 7.
p~+1
= pr
+"
[ ~ 4> (x) TV x N ' (x) viN' (x) 4> (xl ] -I
L D'(x)~(x)TV'xN'(x),
(9)
x
PI?'
where N'(x) = N(1jJ(x,
and D'(x) = M(x) - N'(x). For a given levell of the
multiresolution decomposition, the iterative process of equation 9 can be summarized as
follows.
1. Compute N'(x) by warping the pattern to classify N(x) according to the best
current estimate of p, and compute its spatial gradient V'xN'(x).
2. Update the estimate of PI according to equation 9.
3. Stop if convergence, otherwise go to 1.
Once the final PI is obtained, it is passed to the multiresolution level below (by doubling the
translation parameters), where it is used as initial estimate. Given the values of Pi which
minimize the MD between a pattern to classify and a set of prototypes in the database, a
K-nearest neighbor classifier is used to find the pattern's class.
5 Robust classifiers
One issue of importance for pattern recognition systems is that of robustness to outliers, i.e
errors which occur with low probability, but which can have large magnitude. Examples
are errors due to variation of facial features (e.g. faces shot with or without glasses) in
face recognition, errors due to undesired blobs of ink or uneven line thickness in character
recognition, or errors due to partial occlusions (such as a hand in front of a face) or partially
Multiresolution Tangent Distance/or Affine-invariant Classification
847
missing patterns (such as an undoted i). It is well known that a few (maybe even one)
outliers of high leverage are sufficient to throw mean squared error estimators completely
off-track [6] .
Several robust estimators have been proposed in the statistics literature to avoid this problem.
In this work we consider M-estimators [4] which can be very easily incorporated in the
MD classification framework. M-estimators are an extension of least squares estimators
where the square function is substituted by a functional p(x) which weighs large errors less
heavily. The robust-estimator version of the tangent distance then becomes to minimize the
cost function
T(M, N) = min
p(M(x) - Tp[N{x)]) ,
(10)
p
I:
x
and it is straightforward to show that the "robust" equivalent to equation 9 is
p~+' ~ pr +" [~P"[D(X))oI>(X)TI7XN'(X)I7;;:N'(X)oI>(X)T]-' x
[~P'[D(X))oI>(X)Tl7xN' (X)] ,
(11)
where D(x) = M(x) - N'(x) and p'(x) and p"(x) are, respectively, the first and second
derivatives of the function p( x) with respect to its argument.
6 Experimental results
In this section, we report on experiments carried out to evaluate the performance of the MD
classifier. The first set of experiments was designed to test the invariance of the TD to affine
transformations of the input. The second set was designed to evaluate the improvement
obtained under the multiresolution framework.
6.1
Affine invariance of the tangent distance
Starting from a single view of a reference face, we created an artificial dataset composed
by 441 affine transformations of it. These transformations consisted of combinations of
all rotations in the range from - 30 to 30 degrees with increments of 3 degrees, with all
scaling transformations in the range from 70% to 130% with increments of 3%. The faces
associated with the extremes of the scaling/rotation space are represented on the left portion
of figure 2.
On the right of figure 2 are the distance surfaces obtained by measuring the distance
associated with several metrics at each of the points in the scaling/rotation space. Five
metrics were considered in this experiment: the Euclidean distance (ED), the TD, the MD
computed through Newton's method, the MRMD, and the MRTD.
While the TD exhibits some invariance to rotation and scaling, this invariance is restricted
to a small range of the parameter space and performance only slightly better than the
obtained with the ED. The performance of the MD computed through Newton's method
is dramatically superior, but still inferior to those achieved with the MRTD (which is very
close to zero over the entire parameter space considered in this experiment), and the MRMD.
The performance of the MRTD is in fact impressive given that it involves a computational
increase of less than 50% with respect to the TD, while each iteration of Newton's method
requires an increase of 100%, and several iterations are typically necessary to attain the
minimum MD.
N. Vasconcelos and A. Lippman
848
-30
!i
~
-0
a:
0
1.3
0.7
Scaling
Figure 2: Invariance of the tangent distance. In the right, the surfaces shown correspond to ED, TO,
MO through Newton's method, MRTO, and MRMO. This ordering corresponds to that of the nesting
of the surfaces, i.e. the ED is the cup-shaped surface in the center, while the MRMO is the flat surface
which is approximately zero everywhere.
6.2
Face recognition
To evaluate the performance of the multiresolution tangent distance on a real classification
task, we conducted a series of face recognition experiments, using the Olivetti Research
Laboratories (ORL) face database. This database is composed by 400 images of 40 subjects,
10 images per subject, and contains variations in pose, light conditions, expressions and
facial features, but small variability in terms of scaling, rotation, or translation. To correct
this limitation we created three artificial datasets by applying to each image three random
affine transformations drawn from three multivariate normal distributions centered on the
identity transformation with different covariances. A small sample of the faces in the
database is presented in figure 3, together with its transformed version under the set of
transformations of higher variability.
Figure 3: Left: sample of the ORL face database. Right: transformed version.
We next designed three experiments with increasing degree of difficulty. In the first, we
selected the first view of each subject as the test set, using the remaining nine views as
training data. In the second, the first five faces were used as test data while the remaining
five were used for training. Finally, in the third experiment, we reverted the roles of the
datasets used in the first. The recognition accuracy for each of these experiments and each
of the datasets is reported on figure 4 for the ED, the TO, the MRTD, and a robust version
of this distance (RMRTO) with p(x) = 1x2 if x::; aT and p(x) = ~2 otherwise, where T
is a threshold (set to 2.0 in our experiments), and a a robust version of the error standard
deviation defined as a = median lei - median (ei )1 /0.6745.
Several conclusions can be taken from this figure. First, it can be seen that the MRTD
provides a significantly higher invariance to linear transformations than the ED or the TO,
MultiresolUlion Tangent Distance for Affine-invariant Classification
849
increasing the recognition accuracy by as much as 37.8% in the hardest datasets. In fact,
for the easier tasks of experiments one and two, the performance of the multiresolution
classifier is almost constant and always above the level of 90% accuracy. It is only for the
harder experiment that the invariance of the MRTO classifier starts to break down. But even
in this case, the degradation is graceful- the recognition accuracy only drops below 75%
for considerable values of rotation and scaling (dataset D3).
On the other hand, the ED and the single resolution TO break down even for the easier
tasks, and fail dramatically when the hardest task is performed on the more difficult datasets.
Furthermore, their performance does not degrade gracefully, they seem to be more invariant
when the training set has five views than when it is composed by nine faces of each subject
in the database.
I
'iJ
....
'~--------~,~-=-=~~ ~
,
'
,
___ _ ...l _ _ _ _ _ __ _ l...
,
- - --1- -
,
,
,
,
- - -- - - --r- --- - - - - -t- -
. .. _
__ _____ _ L _ _ _ _ ___
,
,,
_ _ _ _ ______ L- __ _ __ __
,,
,,
am .
----
- -- - -- - -r-- ---- - -t - - - - --- - t,
,-
,,
...1 ___ __ _ __ l-
I
,
__
....
-------;----...---.........
:0
__ ~
L . _
_ _
....1_ _
_
I
.OII IIL.
r
?
iii""
----- - - - j- ------ - -t - - - - - - - - r -TD""""
_
I .
_ _ ______ IL _____ __ ...1! _ _ _ ?_ _ _ _ .l-
11041
10>00_
OO~
_
JOQI _
..,CV _
,,
_ ____ __
,LI __ _ ____ ,,
,,
,,
_ __ ___ _ _
,,
,,
,L, _
,,
-10 _ _ ____ __ ... _
_ _ __ ____
j
l'W'> _ _ _ _ _ _ _
_
~
__ __ _ _
~ -- - ---- ~ -- --- -- - ~ -
_
~
_ _ _ _ _ _ _ ... _ _ _ _ _ _ _ _ ...
,
I
,
,
,
I
I
,
I
I
I
1Mb
IIRm
- ~?~?~:~j?~~~~~~~}~~~~~~~~~
:
:
'~
-- - - -- -- t-- -- -- -- -~ - -- ----t" _ _ _ _ _ _ _ _ 1-
!
I
_
,
- -- - -~ -
,
,
,,
,,
...1_ ______ , _
,,
,
,,
,
"".(1,1 ____ __ _ _ _ ~- - ---- --+ ---- --- -~,
,
,,
,
,
,
'!O.(U.
.,
.
---- - -I-- ------ - - ------ - r -m-
,GI;IIIIII_
110 m ?
? _
:1111,1,1 .
__ _ _ _
_
___ L ___ _ _ __ ...l ___ ___ ___ _
?>.011 _
_ _ __ ___
I
I
-
,
_
_
~
,
,
I____ _ ___
I
?
+____ _____ t _
,
,
_____ _ _ I _ _ _ _ _ ___ ~
I
__ __ ___ L _____ ____ _
,
,
,,
,
,;.
Figure 4: Recognition accuracy. From left to right: results from the first, second, and third
experiments. Oatasets are ordered by degree of variability: 00 is the ORL database 03 is subject to
the affine transfonnations of greater amplitude.
Acknowledgments
We would like to thank Federico Girosi for first bringing the tangent distance to our attention,
and for several stimulating discussions on the topic.
References
[1J P. Anandan, J. Bergen, K. Hanna, and R. Hingorani. Hierarchical Model-Based Motion Estimation. In M. Sezan and R. Lagendijk, editors, Motion Analysis and Image
Sequence Processing, chapter 1. Kluwer Academic Press, 1993.
[2J D. Bertsekas. Nonlinear Programming. Athena Scientific, 1995.
[3J P. Burt and E. Adelson. The Laplacian Pyramid as a Compact Image Code. IEEE
Trans. on Communications, Vol. 31:532-540,1983.
[4] P. Huber. Robust Statistics. John Wiley, 1981 .
[5] B. Lucas and T. Kanade. An Iterative Image Registration Technique with an Application
to Stereo Vision. In Proc. DARPA Image Understanding Workshop, 198 I.
[6J P. Rousseeuw and A. Leroy. Robust Regression and Outlier Detection. John Wiley,
1987.
[7] P. Simard, Y. Le Cun, and J. Denker. Efficient Pattern Recognition Using a New
Transformation Distance. In Proc. Neurallnfonnation Proc. Systems, Denver, USA,
1994.
[8] P. Simard, Y. Le Cun, and 1. Denker. Memory-based Character Recognition Using a
Transformation Invariant Metric. In Int. Conference on Pattern Recognition, Jerusalem,
Israel, 1994.
| 1474 |@word version:7 covariance:1 decomposition:5 harder:1 shot:1 initial:3 series:2 contains:2 current:1 john:2 girosi:1 shape:1 tlp:1 designed:3 update:1 drop:1 v:1 selected:1 guess:2 plane:1 coarse:1 provides:1 ames:1 five:4 consists:1 huber:1 themselves:2 multi:3 td:8 considering:1 increasing:2 becomes:2 medium:2 lowest:1 israel:1 transformation:23 transfonnations:1 classifier:6 bertsekas:1 local:4 approximately:1 range:7 acknowledgment:1 ond:1 lf:1 i____:1 lippman:4 procedure:3 significantly:3 attain:1 regular:2 get:1 close:1 applying:1 equivalent:2 missing:1 center:1 go:1 straightforward:1 starting:2 attention:1 jerusalem:1 resolution:13 simplicity:2 estimator:6 nesting:1 embedding:1 variation:7 coordinate:1 increment:2 heavily:1 programming:1 recognition:16 expensive:1 approximated:1 database:7 bottom:1 role:1 region:1 vxn:1 ordering:1 decrease:1 environment:1 solving:1 completely:1 easily:4 po:1 darpa:1 represented:1 chapter:1 describe:1 artificial:2 hyper:1 choosing:1 otherwise:2 federico:1 ability:1 statistic:2 gi:1 final:1 blob:1 sequence:1 mb:1 combining:1 multiresolution:18 deformable:1 achieve:2 convergence:6 extending:1 illustrate:1 andrew:1 oo:1 pose:1 ij:1 nearest:1 throw:1 involves:1 exhibiting:1 drawback:1 correct:1 centered:1 vx:1 stringent:1 require:2 extension:2 hold:1 around:1 considered:3 normal:1 iil:1 mo:1 substituting:3 major:1 purpose:2 estimation:3 proc:3 sensitive:2 minimization:7 mit:2 gaussian:1 always:1 pn:5 avoid:1 improvement:3 am:1 glass:1 bergen:1 typically:2 lj:1 entire:1 originating:1 transformed:2 issue:2 classification:9 among:1 oii:1 lucas:1 spatial:3 yip:1 once:1 vasconcelos:4 shaped:1 hardest:3 adelson:1 report:1 inherent:1 few:2 composed:3 occlusion:2 detection:1 screening:1 extreme:1 light:1 amenable:1 partial:1 necessary:2 facial:2 euclidean:4 taylor:2 weighs:1 classify:3 tp:13 measuring:1 applicability:1 cost:2 deviation:1 conducted:1 too:3 front:1 characterize:1 reported:1 thickness:1 combined:3 st:1 off:1 together:2 squared:1 simard:3 leading:3 derivative:2 li:1 sec:1 summarized:1 int:1 caused:1 performed:3 view:5 break:2 analyze:1 portion:1 start:1 maintains:1 vin:1 minimize:3 square:2 oi:3 accuracy:6 il:1 correspond:1 vp:2 classified:1 reach:1 suffers:1 ptp:3 ed:7 nuno:2 associated:4 hamming:1 stop:1 dataset:2 amplitude:1 higher:4 formulation:1 though:1 furthermore:2 until:1 hand:2 ei:1 nonlinear:1 ujj:1 scientific:1 lei:1 usa:1 consisted:1 true:1 laboratory:2 iteratively:1 undesired:1 sin:1 ll:1 inferior:1 motion:2 image:19 reverted:1 common:1 rotation:6 superior:1 functional:1 denver:1 slight:1 kluwer:1 significant:3 cambridge:1 cup:1 cv:1 similarity:1 impressive:2 surface:5 neurallnfonnation:1 multivariate:1 olivetti:1 moderate:1 seen:1 minimum:12 greater:1 anandan:1 employed:1 signal:1 isa:1 reduces:1 smooth:1 levell:1 match:1 academic:1 laplacian:1 regression:1 vision:2 metric:8 iteration:9 pyramid:1 achieved:1 median:2 bringing:1 wkt:1 subject:9 seem:1 leverage:1 iii:1 enough:1 idea:1 prototype:1 i7:1 expression:2 passed:1 effort:1 stereo:1 hessian:1 nine:2 jj:2 dramatically:2 maybe:1 rousseeuw:1 reduced:1 notice:2 trapped:1 track:1 per:1 vol:1 key:1 nevertheless:1 threshold:1 drawn:1 d3:1 registration:2 imaging:1 everywhere:1 almost:1 scaling:7 orl:3 layer:2 guaranteed:1 leroy:1 occur:1 x2:1 flat:1 argument:1 min:3 graceful:1 tv:2 according:2 combination:1 smaller:1 slightly:1 character:4 appealing:1 cun:2 outlier:3 invariant:10 pr:2 restricted:1 sided:1 taken:1 computationally:1 equation:14 fail:1 tractable:1 denker:2 hierarchical:1 hingorani:1 robustness:1 top:1 remaining:2 newton:9 approximating:1 ink:1 warping:1 lex:2 md:11 exhibit:3 gradient:5 distance:37 thank:1 athena:1 gracefully:1 degrade:1 topic:1 manifold:14 code:1 minimizing:1 difficult:2 datasets:5 discarded:1 extended:2 incorporated:1 variability:3 communication:1 burt:1 introduced:1 required:1 narrow:1 trans:1 below:2 pattern:14 memory:1 difficulty:1 rely:2 e15:1 improve:1 mrmd:4 carried:1 created:2 ljj:4 literature:2 understanding:1 tangent:18 limitation:4 degree:4 affine:11 sufficient:1 editor:1 pi:4 translation:6 drastically:1 iim:1 neighbor:1 template:1 face:14 taking:1 benefit:1 xn:3 valid:1 made:1 simplified:1 avoided:1 compact:1 global:4 iterative:3 lip:1 kanade:1 nature:1 robust:10 hanna:1 expansion:1 complex:1 necessarily:1 substituted:1 main:1 repeated:1 allowed:1 depicts:1 wiley:2 xl:1 third:2 down:2 embed:1 bad:1 specific:1 showing:1 disregarding:1 dl:1 workshop:1 importance:1 magnitude:1 disregarded:1 easier:2 led:1 likely:1 ordered:1 partially:2 doubling:1 corresponds:2 relies:1 extracted:1 ma:1 stimulating:1 identity:2 considerable:1 degradation:1 invariance:12 experimental:1 uneven:1 latter:1 evaluate:3 |
519 | 1,475 | Structural Risk Minimization for
Nonparametric Time Series Prediction
Ron Meir*
Department of Electrical Engineering
Technion, Haifa 32000, Israel
rmeir@dumbo.technion.ac.il
Abstract
The problem of time series prediction is studied within the uniform convergence framework of Vapnik and Chervonenkis. The dependence inherent in the temporal structure is incorporated into the analysis, thereby
generalizing the available theory for memoryless processes. Finite sample bounds are calculated in terms of covering numbers of the approximating class, and the tradeoff between approximation and estimation is
discussed. A complexity regularization approach is outlined, based on
Vapnik's method of Structural Risk Minimization, and shown to be applicable in the context of mixing stochastic processes.
1 Time Series Prediction and Mixing Processes
A great deal of effort has been expended in recent years on the problem of deriving robust
distribution-free error bounds for learning, mainly in the context of memory less processes
(e.g. [9]). On the other hand, an extensive amount of work has been devoted by statisticians
and econometricians to the study of parametric (often linear) models of time series, where
the dependence inherent in the sample, precludes straightforward application of many of
the standard results form the theory of memoryless processes. In this work we propose
an extension of the framework pioneered by Vapnik and Chervonenkis to the problem of
time series prediction. Some of the more elementary proofs are sketched, while the main
technical results will be proved in detail in the full version of the paper.
Consider a stationary stochastic process X = { ... ,X-1, X 0, X 1, ... }, where Xi is a random variable defined over a compact domain in R and such that IXil ::; B with probability
1, for some positive constant B. The problem of one-step prediction, in the mean square
sense, can then be phrased as that of finding a function f (.) of the infinite past, such that
E
IXo - f(X=~) 12 is minimal, where we use the notation xf
= (Xi, Xi ti, ... ,Xj ),
?This work was supported in part by the a grant from the Israel Science Foundation
Structural Risk Minimization/or Nonparametric Time Series Prediction
309
j ~ i. It is well known that the optimal predictor in this case is given by the conditional
mean, E[XoIX:!J While this solution, in principle, settles the issue of optimal prediction, it does not settle the issue of actually computing the optimal predictor. First of all,
note that ~o compute the conditional mean, the probabilistic law generating the stochastic
process X must be known. Furthermore, the requirement of knowing the full past, X=-~,
is of course rather stringent. In this work we consider the more practical situation, where
a finite sub-sequence Xi" = (Xl, X 2,??? ,XN ) is observed, and an optimal prediction is
needed, conditioned on this data. Moreover, for each finite sample size N we allow the
pre.dictors to be based only on a finite lag vector of size d. Ultimately, in order to achieve
full generality one may let d -+ 00 when N -+ 00 in order to obtain the optimal predictor.
We first consider the problem of selecting an empirical estimator from a class of functions
Fd,n : Rd -+ R, where n is a complexity index of the class (for example, the number
of computational nodes in a feedforward neural network with a single hidden layer), and
If I ::; B for f E Fd,n. Consider then an empirical predictor fd,n,N(Xi=~), i > N,
for Xi based on the finite data set Xi" and depending on the d-dimensional lag vector
Xi=~, where fd,n,N E Fd,n. It is possible to split the error incurred by this predictor into
three terms, each possessing a rather intuitive meaning. It is the competition between these
terms which determines the optimal solution, for a fixed amount of data. First, define the
IX
loss of a functional predictor f : Rd -+ R as L(f) = E i - f(xi=~) 12 , and let fd,n
be the optimal function in Fd,n minimizing this loss. Furthermore, denote the optimal lag
d predictor by fd' and its associated loss by L'd. We are then able to split the loss of the
empirical predictor fd,n,N into three basic components,
L(fd,n,N) = (Ld,n,N - L'd,n)
+ (L'd,n -
L'd)
+ L'd,
(I)
=
where Ld,n,N
L(fd,n,N). The third term, L'd, is related to the error incurred in using a finite memory model (of lag size d) to predict a process with potentially infinite memory. We
do not at present have any useful upper bounds for this term, which is related to the rate of
convergence in the martingale convergence theorem, which to the best of our knowledge is
unknown for the type of mixing processes we study in this work. The second term in (1) , is
related to the so-called approximation error, given by Elfei (X:=-~) - fel,n (Xf=~) 12 to which
it can be immediately related through the inequality IIal P- IblPI ::; pia - bll max( a, b) Ip-l .
This term measures the excess error incurred by selecting a function f from a class of limited complexity Fd,n, while the optimal lag d predictor fei may be arbitrarily complex. Of
course, in order to bound this term we will have to make some regularity assumptions about
the latter function. Finally, the first term in (1) r~resents the so called estimation error,
and is the only term which depends on the data Xl . Similarly to the problem of regression
for i.i.d. data, we expect that the approximation and estimation terms lead to conflicting
demands on the choice of the the complexity, n, of the functional class Fd,n. Clearly, in
order to minimize the approximation error the complexity should be made as large as possible. However, doing this will cause the estimation error to increase, because of the larger
freedom in choosing a specific function in Fd,n to fit the data. However, in the case of time
series there is an additional complication resulting from the fact that the misspecification
error L'd is minimized by choosing d to be as large as possible, while this has the effect
of increasing both the approximation as well as the estimation errors. We thus expect that
sOrhe optimal values of d and n exist for each sample size N.
Up to this point we have not specified how to select the empirical estimator f d,n,N. In this
work we follow the ideas of Vapnik [8], which have been studied extensively in the context of i.i.d observations, and restrict our selection to that hypothesis which minimizes the
IX
empirical error, given by LN(f) = N~d 2::~d+l
i - f(x:=~)12 . For this function it is
easy to establish (see for example [8]) that (Ld,n,N - L'd,n) ::; 2 sUP!E.rd,n IL(f) - LN(f)I?
The main distinction from the i.i.d case, of course, is that random variables appearing in
R. Meir
310
the empirical error, LN(f), are no longer independent. It is clear at this point that some
assumptions are needed regarding the stochastic process X, in order that a law of large
numbers may be established. In any event, it is obvious that the standard approach of using
randomization and symmetrization as in the i.i.d case [3] will not work here. To circumvent this problem, two approaches have been proposed. The first makes use of the so-called
method of sieves together with extensions of the Bernstein inequality to dependent data [6].
The second approach, to be pursued here, is based on mapping the problem onto one characterized by an i.i.d process [10], and the utilization of the standard results for the latter
case.
In order to have some control of the estimation error discussed above, we will restrict ourselves in this work to the class of so-called mixing processes. These are processes for which
the 'future' depends only weakly on the 'past', in a sense that will now be made precise.
Following the definitions and notation of Yu [10], which will be utilized in the sequel, let
(7t = (7(Xf) and (7:+m = (7(Xt~m)' be the sigma-algebras of events generated by the random variables Xf = (X 1 ,X2 , ??? ,Xt) and Xi1.m = (X1+m ,Xl+ m+1 , ?? . ), respectively.
We then define 13m, the coefficient of absolute regularity, as
13m = SUPt>l Esup {IP(BI(7I) - P(B)I : BE (7:+m} , where the expectation is taken
with respect-to (71 = (7(XD. A stochastic process is said to be 13-mixing if (3m -t 0 as
m -t 00. We note that there exist many other definitions of mixing (see [2] for details).
The motivation for using the 13-mixing coefficient is that it is the weakest form of mixing
for which uniform laws of large numbers can be established. In this work we consider two
type of processes for which this coefficient decays to zero, namely algebraically decaying
processes for which 13m ~ /3m- r , /3, r > 0, and exponentially mixing processes for which
13m ~ /3 exp{ -bm K } , jJ, b, I\, > O. Note that for Markov processes mixing implies exponential mixing, so that at least in this case, there is no loss of generality in assuming that
the process is exponentially mixing. Note also that the usual i.i.d process may be obtained
from either the exponentially or algebraically mixing process, by taking the limit I\, -t 00
or r -t 00, respectively.
In this section we follow the approach taken by Yu [10] in deriving uniform laws of large
numbers for mixing processes, extending her mainly asymptotic results to finite sample
behavior, and somewhat broadening the class of processes considered by her. The basic
idea in [10], as in many related approaches, involves the construction of an independentblock sequence, which is shown to be 'close' to the original process in a well-defined
probabilistic sense. We briefly recapitulate the construction, slightly modifying the notation in [10] to fit in with the present paper. Divide the sequence xi' into 2J-lN blocks,
each of size aN; we assume for simplicity that N = 2J-lNaN. The blocks are then numbered according to their order in the block-sequence. For 1 ~ j ~ J-lN define H j
{i : 2(j - l)aN + 1 ~ i ~ (2j - l)aN} and Tj = {i : (2j - l)aN + 1 ~ i ~ (2j)aN}.
Denote the random variables corresponding to the H j and Tj indices as X(j) = {Xi :
i E H j } and X' (j) = {Xi : i E T j }. The sequence of H-blocks is then denoted by
X aN = {X(j)}j:l. Now, construct a sequence of independent and identically distributed
(i.i.d.) blocks {3(j) )}j:l' where 3(j) = {~i : i E H j }, such that the sequence is independent of Xi" and each block has the same distribution as the block X(j) from the original
sequence. Because the process is stationary, the blocks 3(j) are not only independent but
also identically distributed. The basic idea in the construction of the independent block
sequence is that it is 'close', in a well-defined sense to the original blocked sequence X aN .
Moreover, by appropriately selecting the number of blocks, J-lN, depending on the mixing
nature of the sequence, one may relate properties of the original sequence X f", to those of
the independent block sequence 3 aN (see Lemma 4.1 in [10)).
=
Let F be a class of bounded functions, such that 0
~
f
~
B for any f
E
F. In order to
Structural Risk Minimizationfor Nonparametric Time Series Prediction
311
relate the uniform deviations (with respect to F) of the original sequence Xi' to those of
the independent-block sequence BaN' use is made of Lemma 4.1 from [10]. We also utilize
Lemma 4.2 from [10] and modify it so that it holds for finite sample size. Consider the
block-independent sequence BaN and define EJ.LN = J.L1N 'E~:1 f(B(j)) where f(=.(j?) =
'EiEH j f(~i)' j = 1,2, ... , J-lN, is a sequence of independent random variables such that
1
111 ~ aNB. In the remainder of the paper we use variables with a tilde above them to
denote quantities related to the transformed block sequence. Finally, we use the symbol
EN to denote the empirical average with respect to the original sequence, namely EN f =
(N - d)-1 'E~d+1 f(Xi). The following result can be proved by a simple extension of
Lemma 4.2 in [10] .
Lemma 1.1 Suppose F is a permissible class of boundedfunctions,
If I ~
B for
f E :F.
Then
p
{sup
lEN f - Efl >
~ 2P {sup
IEJ.LN 1 - Ell> aNt:} + 2J-lNf3aN'
fE';:
fE';:
t:}
(2)
The main merit of Lemma 1.1 is in the transformation of the problem from the domain of
dependent processes, implicit in the quantity lEN f - Efl, to one characterized by independent processes, implicit in the term EJ.LN
Ell corresponding to the independent blocks.
The price paid for this transformation is the extra term 2J-lN f3aN which appears on the r.h.s
of the inequality appearing in Lemma 1.1 .
1-
2 Error Bounds
The development in Section 1 was concerned with a scalar stochastic process X. In order
to use the results in the context of time series, we first define a new vector-valued pro........ . . . . . . . .
d+l . For
cess X' = { ... ,X- 1 ,XO ,X 1 , ... } where Xi = (Xi,X i - 1, . :.... ,Xi-d) E ~
this sequence the f3-mixing coefficients obey the inequality f3m(X ' ) ~ f3m-d(X). Let F
be a space of functions mapping Rd -7 R, and for each f E F let the loss function be
given by ff(Xf-d) = IXi - f(X:~~W? The loss space is given by L,;: = {ff: f E F} .
It is well known in the theory of empirical processes (see [7] for example), that in order to obtain upper bounds on uniform deviations of i.i.d sequences, use must be made
of the so-called covering number of the function class F, with respect to the empirical it,N norm, given by it ,N(f, g) = N- 1 'E~1 If(Xd - g(Xi)l? Similarly, we denote the empirical norm with respect to the independent block sequence by [1 ,J.LN' where
[1,J.LN(f,g) J-l,'/ 'E~:1 11(x(j)) - g(X(j) I, and where f(X(j?)
'EiEH j Xi and similarly for g. Following common practice we denote the t:-covering number of the functional
space F using the metric p by N(t:, F, p).
=
=
Definition 1 Let L';: be a class of real-valued functions from RD --t R, D = d + 1. For
eachff E L,;:andx = (Xl,X2, .. . ,XaN ), Xi E R D , let if (x) = 'E~:lff(Xi)' Then
define
?,;:
= {if: if E L';:} ,where if : RaND -7 R+.
In order to obtain results in terms of the covering numbers of the space L';: rather than ?,;:,
which corresponds to the transformed sequence, we need the following lemma, which is
not hard to prove.
Lemma 2.1 For any t: > 0
N
(t:, ?,;:, [1 ,J.LN)
~ N (t:jaN, L';:, h,N).
R. Meir
312
PROOF The result follows by sequence of simple inequalities, showing that ll.J1.N (j, g) ~
aNh,N(f, g).
I
We now present the main result of this section, namely an upper bound for the uniform
deviations of mixing processes, which in turn yield upper bounds on the error incurred by
the empirically optimal predictor fd ,n.N.
Theorem 2.1 Let X = { . .. ,Xl' X o, Xl, ... } be a bounded stationary (3-mixing stochastic process, with IXil ~ B, and let F be a class of bo unded functions, f : Rd ~ [0, B].
For each sample size N, let f~ be the function in :F which minimizes the empirical error,
and 1* is the function in F minimizing the true error L(f). Then,
where c' = c/128B.
PROOF The theorem is established by making use of Lemma 1.1, and the basic results from
the theory of uniform convergence for i.i.d. processes, together with Lemma 2.1 relating
the covering numbers of the spaces iF and LF. The covering numbers of LF and Fare
easily related using N(c, LF, Ll (P)) ~ N(c/2B, F, Ll (P)) .
I
Up to this point we have not specified J..tN and aN, and the result is therefore quite general.
In order to obtain weak consistency we require that that the r.h.s. of (3) converge to zero
for each c > O. This immediately yields the following conditions on J..tN (and thus also on
aN through the condition 2aNJ..tN = N).
Corollary 2.1 Under the conditions of Theorem 2.1, and the added requirements that d =
o(aN) and N(c, F, h,N) < 00, the following choices of J..tN are sufficient to guarantee the
weak consistency of the empirical predictor f N:
J..tN ,..", N/t/(1+/t)
(exponential mixing),
(4)
J..tN""" N s/{1+s), 0 < s < r
(algebraic mixing),
(5)
where the notation aN ,..", bN implies that O(bN)
~
aN
~
O(b N ).
PROOF Consider first the case of exponential mixing. In this case the r.h.s. of (3) clearly
converges to zero because of the finiteness of the covering number. The fastest rate of
convergence is achieved by balancing the two terms in the equation, leading to the choice
J..tN '" N/t/(1+/t). In the case of algebraic mixing, the second term on the r.h.s. of (3) is
of the order O(J..tNa"i/) where we have used d = o(aN). Since J..tNaN '" N, a sufficient
condition to guarantee that this term converge to zero is that J..tN ,..", Ns/(1+s), 0 < s < r,
as was claimed.
I
In order to derive bounds on the expected error, we need to make an assumption concerning
the covering number of the space F. In particular, we know from the work Haussler [4J
that the covering number is upper bounded as follows
N(c , F, L 1 (P))
~ e(Pdim(F) + 1)
2B) Pdim(F) '
(-7-
for any measure P. Thus, assuming the finiteness of the pseudo-dimension of F guarantees
a finite covering number.
Structural Risk Minimization/or Nonparametric Time Series Prediction
313
3 Structural Risk Minimization
The results in Section 2 provide error bounds for estimators formed by minimizing the empirical error over a fixed class of d-dimensional functions. It is clear that the complexity of
the class of functions plays a crucial role in the procedure. If the class is too rich, manifested by very large covering numbers, clearly the estimation error term will be very large.
On the other hand, biasing the class of functions by restricting its complexity, leads to poor
approximation rates. A well-known strategy for overcoming this dilemma is obtained by
considering a hierarchy of functional classes with increasing complexity. For any given
sample size, the optimal trade-off between estimation and approximation can then be determined by balancing the two terms. Such a procedure was developed in the late seventies
by Vapnik [8], and termed by him structural risk minimization (SRM). Other more recent
approaches, collectively termed complexity regularization, have been extensively studied
in recent years (e.g. [1]). It should be borne in mind, however, that in the context of time
series there is an added complexity, that does not exist in the case of regression. Recall
that the results derived in Section 2 assumed some fixed lag vector d. In general the optimal value of d is unknown, and could in fact be infinite. In order to achieve optimal
performance in a nonparametric setting, it is crucial that the size of the lag be chosen adaptively as well. This added complexity needs to be incorporated into the SRM framework,
if optimal performance in the face of unknown memory size is to be achieved.
Let Fd,n, d, n E
Fd ,n let
N be a sequence of functions, and define F
= U~l U~=l Fd,n ' For any
which from [4] is upper bounded by cc-Pdim(Fd.n). We observe in passing that Lugosi and
Nobel [5] have recently considered situations where the pseudo-dimension Pdim(Fd,n) is
unknown, and the covering number is estimated empirically from the data. Although this
line of thought is potentially very useful, we do not pursue it here, but rather assume that
upper bounds on the pseudo-dimensions of Fd,n are known, as is the case for many classes
of functions used in practice (see for example [9]).
In line with the standard approach in [8] we introduce a new empirical function, which
takes into account both the empirical error as well as the complexity costs penalizing overly
complex models (large complexity index n and lag size d). Let
(6)
where LN(f) is the empirical error of the predictor f and the complexity penalties
given by
IogN1 (c, Fd,n) + Cn
J-lN /64(2B)4
/-LN /64(2B)4
.
~
are
(7)
(8)
The specific form and constants in these definitions are chosen with hindsight, so as to
achieve the optimal rates of convergence in Theorem 3.1 below. The constants Cn and Cd
are positive constants obeying l:~=1 e- Cn :::; 1 and similarly for Cd . A possible choice is
Cn = 210g n + 1 and Cd = 210g d + 1. The value of J-lN can be chosen in accordance with
Corollary 2.1.
Let id,n,N minimize the empirical error LN(f) within the class of functions Fd ,n' ",!e
assume that the classes Fd,n are compact, so that such a minimizer exists. Further, let IN
R. Meir
314
be the function in F minimizing the complexity penalized loss (6), namely
Ld n
, ,
N(1~) = min
min Ld n N(1~ n N)
d2: 1 n2: 1
"
"
(9)
The following basic result establishes the consistency of the structural risk minimization
approach, and yields upper bounds on its performance.
Theorem 3.1 Let Fd,n, d, n E N be sequence offunctional classes, where 1 E Fd,n is
a mapping from Rd to R The expected loss of the function iN, selected according to the
SRM principle, is upper bounded by
EL(iN) ::; min {inf L(J)
d,n
d,n
+ Cl
The main merit of Theorem 3.1 is the demonstration that the SRM procedure achieves an
optimal balance between approximation and estimation, while retaining its non parametric
attributes. In particular, if the optimal lag d predictor 1J belongs to Fd,no for some no, the
SRM predictor would converge to it at the same rate as if no were known in advance. The
same type of adaptivity is obtained with respect to the lag size d. The non parametric rates
of convergence of the SRM predictor will be discussed in the full paper.
References
[1] A. Barron. Complexity Regularization with Application to Artificial Neural Networks. In G. Roussas, editor, Nonparametric Functional Estimation and Related
Topics, pages 561-576. Kluwer Academic Press, 1991.
[2] L. Gyorfi, W. HardIe, P. Sarda, and P. Vieu. Nonparametric Curve Estimation from
Time Series. Springer Verlag, New York, 1989.
[3] D. Haussler. Decision Theoretic Generalizations of the PAC Model for Neural Net
and Other Learning Applications. Information and Computation, 100:78-150, 1992.
[4] D. Haussler. Sphere Packing Numbers for Subsets of the Boolean n-Cube with
Bounded Vapnik-Chervonenkis Dimesnion. J. Combinatorial Theory, Series A
69:217-232,1995.
[5] G. Lugosi and A. Nobel. Adaptive Model Selection Using Empirical Complexities.
Submitted to Annals Statis., 1996.
[6] D. Modha and E. Masry. Memory Universal Prediction of Stationary Random Processes. IEEE Trans. Inj. Th., January, 1998.
[7] D. Pollard. Convergence of Empirical Processes. Springer Verlag, New York, 1984.
[8] V. N. Vapnik. Estimation of Dependences Based on Empirical Data. Springer Verlag,
New York, 1992.
[9] M. Vidyasagar. A Theory of Learning and Generalization. Springer Verlag, New
York,1996.
[10] B. Yu. Rates of convergence for empirical processes of stationary mixing sequences.
Annals of Probability, 22:94-116, 1984.
| 1475 |@word briefly:1 version:1 norm:2 d2:1 bn:2 recapitulate:1 paid:1 thereby:1 ld:5 series:13 selecting:3 chervonenkis:3 past:3 must:2 ixil:2 j1:1 statis:1 stationary:5 pursued:1 selected:1 node:1 ron:1 complication:1 prove:1 introduce:1 expected:2 behavior:1 considering:1 increasing:2 notation:4 moreover:2 bounded:6 anh:1 israel:2 anj:1 minimizes:2 pursue:1 developed:1 finding:1 transformation:2 hindsight:1 guarantee:3 temporal:1 pseudo:3 ti:1 xd:2 utilization:1 control:1 grant:1 f3m:2 masry:1 positive:2 engineering:1 accordance:1 modify:1 limit:1 id:1 modha:1 lugosi:2 studied:3 vieu:1 fastest:1 limited:1 bi:1 gyorfi:1 practical:1 practice:2 block:16 lf:3 procedure:3 jan:1 universal:1 empirical:21 thought:1 pre:1 numbered:1 onto:1 close:2 selection:2 risk:8 context:5 straightforward:1 simplicity:1 immediately:2 estimator:3 haussler:3 deriving:2 annals:2 construction:3 suppose:1 play:1 pioneered:1 hierarchy:1 ixi:1 hypothesis:1 utilized:1 observed:1 role:1 electrical:1 trade:1 complexity:17 ultimately:1 weakly:1 algebra:1 dilemma:1 packing:1 easily:1 artificial:1 choosing:2 quite:1 lag:10 larger:1 valued:2 precludes:1 ip:2 sequence:27 net:1 propose:1 remainder:1 mixing:23 achieve:3 intuitive:1 fel:1 competition:1 convergence:9 regularity:2 requirement:2 extending:1 generating:1 converges:1 depending:2 derive:1 ac:1 involves:1 implies:2 attribute:1 modifying:1 stochastic:7 stringent:1 settle:2 require:1 generalization:2 randomization:1 elementary:1 extension:3 hold:1 considered:2 exp:1 great:1 mapping:3 predict:1 achieves:1 estimation:12 applicable:1 combinatorial:1 symmetrization:1 him:1 establishes:1 minimization:7 clearly:3 supt:1 rather:4 ej:2 corollary:2 derived:1 mainly:2 sense:4 dependent:2 el:1 hidden:1 her:2 transformed:2 sketched:1 issue:2 denoted:1 retaining:1 development:1 ell:2 cube:1 construct:1 f3:1 seventy:1 yu:3 dumbo:1 minimized:1 future:1 inherent:2 ourselves:1 statistician:1 freedom:1 fd:27 tj:2 devoted:1 divide:1 haifa:1 minimal:1 boolean:1 pia:1 cost:1 deviation:3 subset:1 uniform:7 technion:2 predictor:15 srm:6 too:1 adaptively:1 sequel:1 probabilistic:2 xi1:1 off:1 together:2 borne:1 leading:1 expended:1 account:1 coefficient:4 depends:2 doing:1 sup:3 decaying:1 len:2 minimize:2 il:2 square:1 formed:1 yield:3 ant:1 weak:2 cc:1 submitted:1 definition:4 obvious:1 proof:4 associated:1 proved:2 xan:1 recall:1 knowledge:1 actually:1 appears:1 follow:2 rand:1 generality:2 furthermore:2 implicit:2 hand:2 hardie:1 roussas:1 effect:1 true:1 regularization:3 sieve:1 memoryless:2 deal:1 ll:3 covering:12 theoretic:1 tn:8 pro:1 meaning:1 possessing:1 recently:1 common:1 functional:5 empirically:2 exponentially:3 discussed:3 fare:1 relating:1 kluwer:1 blocked:1 rd:7 outlined:1 consistency:3 similarly:4 longer:1 recent:3 inf:1 belongs:1 termed:2 claimed:1 verlag:4 manifested:1 inequality:5 arbitrarily:1 additional:1 somewhat:1 algebraically:2 converge:3 full:4 technical:1 xf:5 academic:1 characterized:2 sphere:1 concerning:1 prediction:11 basic:5 regression:2 expectation:1 metric:1 achieved:2 finiteness:2 crucial:2 appropriately:1 permissible:1 extra:1 structural:8 feedforward:1 split:2 easy:1 concerned:1 bernstein:1 sarda:1 identically:2 xj:1 fit:2 restrict:2 idea:3 regarding:1 knowing:1 tradeoff:1 pdim:4 cn:4 effort:1 penalty:1 algebraic:2 pollard:1 passing:1 cause:1 jj:1 york:4 useful:2 clear:2 amount:2 nonparametric:7 extensively:2 meir:4 exist:3 estimated:1 overly:1 rmeir:1 penalizing:1 ce:1 utilize:1 efl:2 year:2 decision:1 bound:12 layer:1 fei:1 x2:2 phrased:1 min:3 department:1 according:2 poor:1 slightly:1 making:1 xo:1 taken:2 ln:19 equation:1 turn:1 needed:2 know:1 merit:2 mind:1 resents:1 available:1 obey:1 observe:1 barron:1 appearing:2 original:6 l1n:1 establish:1 approximating:1 added:3 quantity:2 parametric:3 strategy:1 dependence:3 usual:1 said:1 topic:1 nobel:2 assuming:2 index:3 minimizing:4 demonstration:1 balance:1 fe:2 potentially:2 relate:2 sigma:1 unknown:4 upper:9 observation:1 markov:1 finite:9 january:1 ixo:1 situation:2 tilde:1 incorporated:2 bll:1 misspecification:1 precise:1 anb:1 overcoming:1 namely:4 specified:2 extensive:1 distinction:1 conflicting:1 established:3 trans:1 able:1 below:1 biasing:1 max:1 memory:5 vidyasagar:1 event:2 circumvent:1 asymptotic:1 law:4 loss:9 expect:2 adaptivity:1 foundation:1 incurred:4 sufficient:2 principle:2 editor:1 balancing:2 cd:3 course:3 penalized:1 ban:2 supported:1 free:1 iej:1 allow:1 taking:1 face:1 absolute:1 distributed:2 curve:1 calculated:1 xn:1 dimension:3 rich:1 made:4 adaptive:1 bm:1 excess:1 compact:2 assumed:1 xi:22 nature:1 robust:1 broadening:1 complex:2 cl:1 domain:2 main:5 motivation:1 n2:1 x1:1 en:2 ff:2 martingale:1 n:1 sub:1 tna:1 obeying:1 exponential:3 xl:6 third:1 late:1 ix:2 theorem:7 specific:2 xt:2 showing:1 pac:1 symbol:1 lff:1 decay:1 weakest:1 exists:1 vapnik:7 restricting:1 conditioned:1 demand:1 generalizing:1 scalar:1 bo:1 collectively:1 springer:4 corresponds:1 minimizer:1 determines:1 conditional:2 price:1 hard:1 infinite:3 determined:1 lemma:11 called:5 inj:1 select:1 latter:2 |
520 | 1,476 | Nonparametric Model-Based
Reinforcement Learning
Christopher G. Atkeson
College of Computing, Georgia Institute of Technology,
Atlanta, GA 30332-0280, USA
ATR Human Information Processing,
2-2 Hikaridai, Seiko-cho, Soraku-gun, 619-02 Kyoto, Japan
cga@cc.gatech.edu
http://www .cc.gatech.edu/fac/Chris.Atkeson/
Abstract
This paper describes some of the interactions of model learning
algorithms and planning algorithms we have found in exploring
model-based reinforcement learning. The paper focuses on how local trajectory optimizers can be used effectively with learned nonparametric models. We find that trajectory planners that are fully
consistent with the learned model often have difficulty finding reasonable plans in the early stages of learning. Trajectory planners
that balance obeying the learned model with minimizing cost (or
maximizing reward) often do better, even if the plan is not fully
consistent with the learned model.
1
INTRODUCTION
We are exploring the use of nonparametric models in robot learning (Atkeson et al.,
1997b; Atkeson and Schaal , 1997). This paper describes the interaction of model
learning algorithms and planning algorithms, focusing on how local trajectory optimization can be used effectively with nonparametric models in reinforcement learning. We find that trajectory optimizers that are fully consistent with the learned
model often have difficulty finding reasonable plans in the early stages of learning .
The message of this paper is that a planner should not be entirely consistent with
the learned model during model-based reinforcement learning. Trajectory optimizers that balance obeying the learned model with minimizing cost (or maximizing
reward) often do better, even if the plan is not fully consistent with the learned
model:
1009
Nonparametric Model-Based Reinforcement Learning
'~~
A
-
1" -
-
.--/
-
~
~
---e L V 2\ \ '\ V V '\
If \
[iV II
'\\ l\
'\ \
./
()
\
\i\
\ \
\ \
I""
l""
/
+
)
. /V
\ \ !"" ~
.""
\~
\
/
I'---"
/
/
~ I" "
~ ~ ~ it
~
V
Figure 1: A: Planning in terms of trajectory segments. B: Planning in terms of
trajectories all the way to a goal point .
Two kinds of reinforcement learning algorithms are direct (non-model-based) and
indirect (model-based) . Direct reinforcement learning algorithms learn a policy or
value function without explicitly representing a model of the controlled system (Sutton et al. , 1992) . Model-based approaches learn an explicit model of the system simultaneously with a value function and policy (Sutton, 1990 , 1991a,b; Barto et al. ,
1995; Kaelbling et al. , 1996) . We will focus on model-based reinforcement learning ,
in which the learner uses a planner to derive a policy from a learned model and an
optimization criterion .
2
CONSISTENT LOCAL PLANNING
An efficient approach to dynamic programming, a form of global planning, is to use
local trajectory optimizers (Atkeson, 1994) . These local planners find a plan for
each starting point in a grid in the state space. Figure 1 compares the output of
a traditional cell based dynamic programming process with the output of a planner based on integrating local plans. Traditional dynamic programming generates
trajectory segments from each cell to neighboring cells, while the planner we use
generates entire trajectories. These locally optimal trajectories have local policies
and local models of the value function along the trajectories (Dyer and McReynolds,
1970; Jacobson and Mayne, 1970). The locally optimal trajectories are made consistent with their neighbors by using the local value function to predict the value
of a neighboring trajectory. If all the local value functions are consistent with their
neighbors the aggregate value function is a unique solution to the Bellman equation
and the corresponding trajectories and policy are globally optimal. We would like
any local planning algorithm to produce a local model of the value function so we
can perform this type of consistency checking. We would also like a local policy
from the local planner, so we can respond to disturbances and modeling errors.
Differential dynamic programming is a local planner that has these characteristics (Dyer and McReynolds. 1970; Jacobson and Mayne, 1970). Differential dynamic programming maintains a local quadratic model of the value function along
the current best trajectory x* (t):
V (x,t) = Vo(t)
+ Vx(t)(x -
x*(t))T
+ 0.5(x -
x*(t))TVxx(t)(x - x*(t))
(1)
C. G. Atkeson
1010
as well as a local linear model of the corresponding policy:
U(X,t) = u*(t)
+ K(t)(x -
x*(t))
(2)
u(x, t) is the local policy at time t, the control signal u as a function of state x.
u * (t) is the model's estimate of the control signal necessary to follow the current
best trajectory x*(t). K(t) are the feedback gains that alter the control signals in
response to deviations from the current best trajectory. These gains are also the
first derivative of the policy along the current best trajectory.
The first phase of each optimization iteration is to apply the current local policy
to the learned model, integrating the modeled dynamics forward in time and seeing
where the simulated trajectory goes. The second phase of the differential dynamic
programming approach is to calculate the components of the local quadratic model
of the value function at each point along the trajectory: the constant term Vo (t), the
gradient Vx (t), and the Hessian Vxx (t). These terms are constructed by integrating
backwards in time along the trajectory. The value function is used to produce a
new policy, which is represented using a new x*(t), u*(t), and K(t).
The availability of a local value function and policy is an attractive feature of
differential dynamic programming. However, we have found several problems when
applying this method to model-based reinforcement learning with nonparametric
models:
1. Methods that enforce consistency with the learned model need an initial
trajectory that obeys that model, which is often difficult to produce.
2. The integration of the learned model forward in time often blows up when
the learned model is inaccurate or when the plant is unstable and the current policy fails to stabilize it.
3. The backward integration to produce the value function and a corresponding policy uses derivatives of the learned model, which are often quite inaccurate in the early stages of learning, producing inaccurate value function
estimates and ineffective policies.
3
INCONSISTENT LOCAL PLANNING
To avoid the problems of consistent local planners, we developed a trajectory optimization approach that does not integrate the learned model and does not require
full consistency with the learned model. Unfortunately, the price of these modifications is that the method does not produce a value function or a policy, just a
trajectory (x(t), u(t)). To allow inconsistency with the learned model, we represent
the state history x(t) and the control history u(t) separately, rather than calculate
x(t) from the learned model and u(t). We also modify the original optimization
criterion C = Lk C(Xk, Uk) by changing the hard constraint that Xk+1 = f(Xk' Uk)
on each time step into a soft constraint:
Cnew =
L
[C(Xk' Uk) +~IXk+1-f(Xk,Uk)12]
(3)
k
C(Xk' Uk) is the one step cost in the original optimization criterion. ~ is the penalty
on the trajectory being inconsistent with the learned model Xk+1 = f(Xk' Uk).
IXk +1 - f (Xk' Uk) I is the magnitude of the mismatch of the trajectory and the model
prediction at time step k in the trajectory. ~ provides a way to control the amount
of inconsistency. A small ~ reflects lack of confidence in the model, and allows
Nonparametric Model-Based Reinforcement Learning
//If\<'' ,,,? "
~j
1011
Figure 2:
The
SARCOS robot
arm with a pendulum gripped in
the hand.
The
pendulum
aXIS
is aligned with
the fingers and
with the forearm in this arm
configuration.
the optimized trajectory to be inconsistent with the model in favor of reducing
A large). reflects confidence in the model, and forces the optimized trajectory to be more consistent with the model. ). can increase with time or with the
number of learning trials . If we use a model that estimates the confidence level of
a prediction, we can vary). for each lookup based on Xk and Uk. Locally weighted
learning techniques provide exactly this type of local confidence estimate (Atkeson
et al., 1997a) .
C(Xk, Uk)'
Now that we are not integrating the trajectory we can use more compact representations of the trajectory, such as splines (Cohen , 1992) or wavelets (Liu et al.,
1994). We no longer require that Xk+l = f(Xk, Uk), which is a condition difficult to
fulfill without having x and u represented as independent values on each time step.
We can now parameterize the trajectory using the spline knot points, for example.
In this work we used B splines (Cohen, 1992) to represent the trajectory. Other
choices for spline basis functions would probably work just as well. We can use any
nonlinear programming or function optimization method to minimize the criterion
in Eq. 3. In this work we used Powell's method (Press et al., 1988) to optimize the
knot points, a method which is convenient to use but not particularly efficient.
4
IMPLEMENTATION ON AN ACTUAL ROBOT
Both local planning methods work well with learned parametric models. However ,
differential dynamic programming did not work at all with learned nonparametric
models, for reasons already discussed. This section describes how the inconsistent
local planning method was used in an application of model-based reinforcement
learning: robot learning from demonstration using a pendulum swing up task (Atkeson and Schaal, 1997). The pendulum swing up task is a more complex version of
the pole or broom balancing task (Spong, 1995) . The hand holds the axis of the
pendulum, and the pendulum rotates about this hinge in an angular movement
(Figure 2). Instead of starting with the pendulum vertical and above its rotational
joint, the pendulum is hanging down from the hand, and the goal of the swing up
task is to move the hand so that the pendulum swings up and is then balanced
in the inverted position . The swing up task was chosen for study because it is a
difficult dynamic maneuver and requires practice for humans to learn, but it is easy
to tell if the task is successfully executed (at the end of the task the pendulum is
balanced upright and does not fall down) .
We implemented learning from demonstration on a hydraulic seven degree of free-
1012
.
C. G. Atkeson
c:
1.0
'tI
0.0
.S!
~
CD
human demonstration
1st trial (imitation)
2nd trial
3rd trial
-1.0
"S. -2.0
c:
III
?
-3.0
'S -4.0
b
i
~
.!
CD
.?.
c:
.2
?1::
-5.0
0.0
0.2
0.4
0.5
0.4
0.3
0.2
0.1
.........
-0.0
.
..................
~ -0.1
~
-0.2
] -0.3
III
0.0 0.2
~
--_-.
0.6
0.8
1.0
1.2
1.4
1.6
1.0
1.2
1.4
1.6
1.8
2.0
1.8
2.0
.~.~
'.,
/.
--,.,.,."..~
?-.----?
seconds
Figure 3: The hand and pendulum motion during robot learning from demonstration using a nonparametric model.
dom anthropomorphic robot arm (SARCOS Dextrous Arm located at ATR, Figure 2). The robot observed its own performance with the same stereo vision system
that was used to observe the human demonstrations.
The robot observed a human swinging up a pendulum using a horizontal hand
movement (dotted line in Figure 3) . The most obvious approach to learning from
demonstration is to have the robot imitate the human motion, by following the
human hand trajectory. The dashed lines in Figures 3 show the robot hand motion
as it attempts to follow the human demonstration of the swing up task, and the
corresponding pendulum angles. Because of differences in the task dynamics for
the human and for the robot, this direct imitation failed to swing the pendulum
up, as the pendulum did not get even halfway up to the vertical position, and then
oscillated about the hanging down position.
The approach we used was to apply a planner to finding a swing up trajectory
that worked for the robot, based on learning both a model and a reward function
and using the human demonstration to initialize the planning process. The data
collected during the initial imitation trial and subsequent trials was used to build
a model. Nonparametric models were constructed using locally weighted learning
as described in (Atkeson et al., 1997a) . These models did not use knowledge of the
model structure but instead assumed a general relationship:
(4)
where () is the pendulum angle and x is the hand position. Training data from
the demonstrations was stored in a database, and a local model was constructed
to answer each query. Meta-parameters such as distance metrics were tuned using
cross validation on the training set. For example, cross validation was able to
quickly establish that hand position and velocity (x and x) played an insignificant
role in predicting future pendulum angular velocities.
The planner used a cost function that penalizes deviations from the demonstration
trajectory sampled at 60H z:
C(Xk,
Uk) =
(Xk -
X~)T(Xk
-
X~) + uluk
(5)
Nonparametric Model-Based Reinforcement Learning
1013
where the state is x = ((J, il, x , x), x d is the demonstrated motion, k is the sample
index, and the control is u = (x). Equation 3 was optimized using B splines to
represent x and u. The knot points for x and u were initially separately optimized
to minimize
(6)
and
(7)
The tolerated inconsistency, ). was kept constant during a set of trials and set
at values ranging from 100 to 100000. The exact value of ). did not make much
difference. Learning failed when). was set to zero , as there was no way for the
learned model to affect the plan. The planning process failed when ). was set too
high , enforcing the learned model too strongly.
The next attempt got the pendulum up a little more. Adding this new data to the
database and replanning resul ted in a movement that succeeded (trial 3 in Figure 3).
The behavior shown in Figure 3 is quite repeatable. The balancing behavior at the
end of the trial is learned separately and continues for several minutes, at which
point the trial is automatically terminated (Schaal, 1997).
5
DISCUSSION AND CONCLUSION
We applied locally weighted regression (Atkeson et aI. , 1997a) in an attempt to avoid
the structural modeling errors of idealized parametric models during model-based
reinforcement learning, and also to see if a priori knowledge of the structure of the
task dynamics was necessary. In an exploration of the swingup task, we found that
these nonparametric models required a planner that ignored the learned model to
some extent. The fundamental reason for this is that planners amplify modeling
error. Mechanisms for this amplification include:
? The planners take advantage of any modeling error to reduce the cost of
the planned trajectory, so the planning process seeks out modeling error
that reduces apparent cost .
? Some planners use derivatives of the model, which amplifies any noise in
the model.
Models that support fast learning will have errors and noise. For example , in order
to learn a model of the complexity necessary to accurately model the full robot
dynamics between the commanded and actual hand accelerations a large amount
of data is required, independent of modeling technique. The input would be 21
dimensional (robot state and command) ignoring actuator dynamics. Because there
are few robot trials during learning, there is not enough data to make such a model
even just in the vicinity of a successful trajectory. If it was required that enough
data is collected during learning to make an accurate model. robot learning would
be greatly slowed down.
One solution to this error amplification is to bias the nonparametric modeling tools
to oversmooth the data. This reduces the benefit of nonparametric modeling, and
also ignores the true learned model to some degree. Our solution to this problem
is to introduce a controlled amount of inconsistency with the learned model into
the planning process. The control parameter). is explicit and can be changed as a
function of time, amount of data, or as a function of confidence in the model at the
query point.
1014
C. G. Atkeson
References
Atkeson, C. G. (1994). Using local trajectory optimizers to speed up global optimization in dynamic programming. In Cowan, J. D., Tesauro, G., and Alspector,
J., editors, Advances in Neural Information Processing Systems 6, pages 663-670.
Morgan Kaufmann, San Mateo, CA.
Atkeson, C . G., Moore, A. W., and Schaal, S. (1997a). Locally weighted learning.
Artificial Intelligence Review, 11:11-73.
Atkeson, C. G., Moore, A. W., and Schaal, S. (1997b). Locally weighted learning
for control. Artificial Intelligence Review, 11:75-113.
Atkeson, C. G. and Schaal, S. (1997) . Robot learning from demonstration. In
Proceedings of the 1997 International Conference on Machine Learning.
Barto, A. G ., Bradtke, S. J., and Singh, S. P. (1995). Learning to act using real-time
dynamic programming. Artificial Intelligence, 72(1):81-138.
Cohen, M. F . (1992). Interactive spacetime control for animation. Computer Graphics, 26(2):293-302.
Dyer, P. and McReynolds, S. (1970). The Computational Theory of Optimal Control.
Academic, NY.
Jacobson, D. and Mayne, D. (1970). Differential Dynamic Programming. Elsevier,
NY.
Kaelbling, L. P., Littman, M. L., and Moore, A. W. (1996). Reinforcement learning:
A survey. lournal of Artificial Intelligence Research, 4:237-285 .
Liu, Z., Gortler, S. J., and Cohen, M. F. (1994). Hierarchical spacetime control.
Computer Graphics (SIGGRAPH '94 Proceedings), pages 35-42.
Press, W. H., Teukolsky, S. A., Vetterling, W. T., and Flannery, B. P. (1988).
Numerical Recipes in C. Cambridge University Press, New York, NY.
Schaal, S. (1997). Learning from demonstration . In Mozer, M. C., Jordan, M., and
Petsche, T ., editors, Advances in Neural Information Processing Systems 9, pages
1040-1046. MIT Press, Cambridge, MA.
Spong, M. W. (1995). The swing up control problem for the acrobot. IEEE Control
Systems Magazine, 15(1):49-55.
Sutton, R. S. (1990). Integrated architectures for learning, planning, and reacting
based on approximating dynamic programming. In Seventh International Machine Learning Workshop, pages 216-224. Morgan Kaufmann, San Mateo, CA.
http://envy.cs.umass.edu/People/sutton/publications.html.
Sutton, R. S. (1991a). Dyna, an integrated architecture for learning, planning
and reacting. http://envy.cs.umass.edu/People/sutton/publications.html, Working Notes of the 1991 AAAI Spring Symposium on Integrated Intelligent Architectures pp. 151-155 and SIGART Bulletin 2, pp. 160-163.
Sutton, R . S. (1991b). Planning by incremental dynamic programming. In Eighth
International Machine Learning Workshop, pages 353-357. Morgan Kaufmann,
San Mateo, CA. http://envy.cs.umass.edu/People/sutton/publications.html.
Sutton, R. S., Barto, A. G., and Williams, R. J. (1992). Reinforcement learning is
direct adaptive optimal control. IEEE Control Systems Magazine, 12:19-22.
| 1476 |@word trial:11 version:1 nd:1 seek:1 initial:2 configuration:1 liu:2 uma:3 tuned:1 current:6 numerical:1 subsequent:1 intelligence:4 imitate:1 xk:16 sarcos:2 provides:1 along:5 constructed:3 direct:4 differential:6 symposium:1 introduce:1 alspector:1 behavior:2 planning:17 bellman:1 globally:1 automatically:1 actual:2 little:1 kind:1 developed:1 finding:3 ti:1 act:1 interactive:1 exactly:1 uk:11 control:15 producing:1 maneuver:1 gortler:1 local:28 modify:1 sutton:9 reacting:2 mateo:3 commanded:1 obeys:1 unique:1 practice:1 optimizers:5 powell:1 got:1 convenient:1 confidence:5 integrating:4 seeing:1 get:1 amplify:1 ga:1 applying:1 www:1 optimize:1 demonstrated:1 maximizing:2 go:1 williams:1 starting:2 survey:1 swinging:1 oscillated:1 magazine:2 exact:1 programming:14 us:2 velocity:2 particularly:1 located:1 continues:1 database:2 observed:2 role:1 parameterize:1 calculate:2 movement:3 balanced:2 mozer:1 complexity:1 reward:3 littman:1 dynamic:19 dom:1 singh:1 segment:2 learner:1 basis:1 joint:1 indirect:1 siggraph:1 represented:2 finger:1 hydraulic:1 vxx:1 fac:1 fast:1 query:2 artificial:4 tell:1 aggregate:1 quite:2 apparent:1 favor:1 advantage:1 interaction:2 neighboring:2 aligned:1 mayne:3 amplification:2 amplifies:1 recipe:1 produce:5 incremental:1 derive:1 eq:1 implemented:1 c:3 mcreynolds:3 exploration:1 human:10 vx:2 require:2 anthropomorphic:1 exploring:2 hold:1 predict:1 vary:1 early:3 replanning:1 successfully:1 tool:1 reflects:2 weighted:5 mit:1 rather:1 fulfill:1 avoid:2 barto:3 gatech:2 command:1 publication:3 focus:2 schaal:7 greatly:1 elsevier:1 inaccurate:3 entire:1 vetterling:1 integrated:3 initially:1 html:3 priori:1 plan:7 integration:2 initialize:1 having:1 ted:1 alter:1 future:1 seiko:1 spline:5 intelligent:1 few:1 simultaneously:1 phase:2 attempt:3 atlanta:1 message:1 jacobson:3 accurate:1 succeeded:1 necessary:3 iv:1 penalizes:1 modeling:8 soft:1 planned:1 cost:6 kaelbling:2 deviation:2 pole:1 successful:1 seventh:1 too:2 graphic:2 stored:1 answer:1 tolerated:1 cho:1 st:1 fundamental:1 international:3 quickly:1 aaai:1 derivative:3 japan:1 blow:1 lookup:1 stabilize:1 availability:1 explicitly:1 idealized:1 ixk:2 pendulum:18 maintains:1 minimize:2 il:1 kaufmann:3 characteristic:1 accurately:1 knot:3 trajectory:41 cc:2 history:2 pp:2 obvious:1 gain:2 sampled:1 knowledge:2 focusing:1 follow:2 response:1 strongly:1 just:3 stage:3 angular:2 hand:11 working:1 horizontal:1 christopher:1 nonlinear:1 lack:1 cnew:1 usa:1 true:1 swing:9 vicinity:1 moore:3 attractive:1 during:7 criterion:4 vo:2 motion:4 bradtke:1 ranging:1 cohen:4 discussed:1 cambridge:2 ai:1 rd:1 grid:1 consistency:3 robot:17 longer:1 own:1 tesauro:1 meta:1 inconsistency:4 inverted:1 morgan:3 signal:3 ii:1 dashed:1 full:2 kyoto:1 reduces:2 academic:1 cross:2 controlled:2 prediction:2 regression:1 spong:2 vision:1 metric:1 iteration:1 represent:3 cell:3 separately:3 resul:1 ineffective:1 probably:1 cowan:1 inconsistent:4 jordan:1 structural:1 backwards:1 iii:2 easy:1 enough:2 affect:1 architecture:3 reduce:1 penalty:1 stereo:1 soraku:1 hessian:1 york:1 ignored:1 cga:1 amount:4 nonparametric:14 locally:7 lournal:1 http:4 dotted:1 changing:1 kept:1 backward:1 halfway:1 angle:2 respond:1 planner:16 reasonable:2 entirely:1 played:1 spacetime:2 quadratic:2 constraint:2 worked:1 generates:2 speed:1 spring:1 hanging:2 describes:3 modification:1 slowed:1 equation:2 mechanism:1 dyna:1 dyer:3 end:2 apply:2 observe:1 actuator:1 hierarchical:1 enforce:1 petsche:1 original:2 include:1 hinge:1 build:1 hikaridai:1 establish:1 approximating:1 move:1 already:1 parametric:2 traditional:2 gradient:1 distance:1 rotates:1 atr:2 simulated:1 gun:1 chris:1 seven:1 collected:2 unstable:1 broom:1 reason:2 dextrous:1 enforcing:1 extent:1 modeled:1 relationship:1 index:1 rotational:1 balance:2 minimizing:2 demonstration:12 difficult:3 unfortunately:1 executed:1 sigart:1 implementation:1 policy:16 perform:1 vertical:2 forearm:1 required:3 optimized:4 learned:27 able:1 mismatch:1 eighth:1 difficulty:2 force:1 disturbance:1 predicting:1 arm:4 representing:1 technology:1 lk:1 axis:2 review:2 swingup:1 checking:1 fully:4 plant:1 validation:2 integrate:1 degree:2 consistent:10 editor:2 balancing:2 cd:2 oversmooth:1 changed:1 free:1 bias:1 allow:1 institute:1 neighbor:2 fall:1 bulletin:1 benefit:1 feedback:1 ignores:1 forward:2 made:1 reinforcement:15 san:3 adaptive:1 atkeson:16 compact:1 global:2 assumed:1 imitation:3 learn:4 ca:3 ignoring:1 complex:1 did:4 terminated:1 noise:2 animation:1 envy:3 georgia:1 ny:3 fails:1 position:5 explicit:2 obeying:2 wavelet:1 down:4 minute:1 repeatable:1 insignificant:1 workshop:2 adding:1 effectively:2 magnitude:1 acrobot:1 flannery:1 failed:3 teukolsky:1 ma:1 goal:2 acceleration:1 price:1 hard:1 upright:1 reducing:1 college:1 support:1 people:3 |
521 | 1,477 | Synaptic Transmission: An
Information-Theoretic Perspective
Amit Manwani and Christof Koch
Computation and Neural Systems Program
California Institute of Technology
Pasadena, CA 91125
email: quixote@klab.caltech.edu
koch@klab.caltech.edu
Abstract
Here we analyze synaptic transmission from an infonnation-theoretic
perspective. We derive c1osed-fonn expressions for the lower-bounds on
the capacity of a simple model of a cortical synapse under two explicit
coding paradigms. Under the "signal estimation" paradigm, we assume
the signal to be encoded in the mean firing rate of a Poisson neuron. The
perfonnance of an optimal linear estimator of the signal then provides
a lower bound on the capacity for signal estimation. Under the "signal
detection" paradigm, the presence or absence of the signal has to be detected. Perfonnance of the optimal spike detector allows us to compute
a lower bound on the capacity for signal detection. We find that single
synapses (for empirically measured parameter values) transmit infonnation poorly but significant improvement can be achieved with a small
amount of redundancy.
1 Introduction
Tools from estimation and infonnation theory have recently been applied by researchers
(Bialek et. ai, 1991) to quantify how well neurons transmit infonnation about their random
inputs in their spike outputs. In these approaches, the neuron is treated like a black-box,
characterized empirically by a set of input-output records. This ignores the specific nature
of neuronal processing in tenns of its known biophysical properties. However, a systematic
study of processing at various stages in a biophysically faithful model of a single neuron
should be able to identify the role of each stage in infonnation transfer in tenns of the parameters relating to the neuron's dendritic structure, its spiking mechanism, etc. Employing
this reductionist approach, we focus on a important component of neural processing, the
synapse, and analyze a simple model of a cortical synapse under two different representa?tional paradigms. Under the "signal estimation" paradigm, we assume that the input signal
202
A. Manwani and C. Koch
is linearly encoded in the mean firing rate of a Poisson neuron and the mean-square error
in the reconstruction of the signal from the post-synaptic voltage quantifies system performance. From the performance of the optimal linear estimator of the signal, a lower
bound on the capacity for signal estimation can be computed. Under the "signal detection"
paradigm, we assume that information is encoded in an all-or-none format and the error in
deciding whether or not a presynaptic spike occurred by observing the post-synaptic voltage
quantifies system performance. This is similar to the conventional absentipresent(Yes-No)
decision paradigm used in psychophysics. Performance of the optimal spike detector in
this case allows us to compute a lower bound on the capacity for signal detection.
0x- 0
NoSpI<o
Poisson
Encoding
!'-.
NoR
&,
Release
1
Spoke
1
Optimal
Estimator
1-
Stimulys
h{l)
Stochastic
Variable EPSC
Vesicle Release
Amplitude
EPSP
Shape
Optimal
Detector
Spike I
No Spike
Spike I
No Spike
Encodlna
Synaptic Channel
Decoding
Figure 1: Schematic block diagram for the signal detection and estimation tasks. The
synapse is modeled as a binary channel followed by a filter h(t) = at exp( -tits). where
a is a random variable with probability density, P(a) = a (aa)k- 1 exp( -aa)/(k - 1)!.
The binary channel, (inset, EO = Pr[spontaneous release], E1 = Pr [release failure]) models
probabilistic vesicle release and h(t) models the variable epsp size observed for cortical
synapses. n( t) denotes additive post-synaptic voltage noise and is assumed to be Gaussian
and white over a bandwidth En. Performance of the optimal linear estimator (Wiener
Filter) and the optimal spike detector (Matched Filter) quantify synaptic efficacy for signal
estimation and detection respectively.
2
The Synaptic Channel
Synaptic transmission in cortical neurons is known to be highly random though the role of
this variability in neural computation and coding is still unclear. In central synapses, each
synaptic bouton contains only a single active release zone, as opposed to the hundreds or
thousands found at the much more reliable neuromuscular junction. Thus, in response to an
action potential in the presynaptic terminal at most one vesicle is released (Kom and Faber,
1991). Moreover, the probability of vesicle release p is known to be generally low (0.1
to 0.4) from in vitro studies in some vertebrate and invertebrate systems (Stevens, 1994).
This unreliability is further compounded by the trial-to-trial variability in the amplitude of
the post-synaptic response to a vesicular release (Bekkers et. ai, 1990). In some cases, the
variance in the size of EPSP is as large as the mean. The empirically measured distribution
of amplitudes is usually skewed to the right (possibly biased due the inability of measuring
very small events) and can be modeled by a Gamma distribution.
In light of the above, we model the synapse as a binary channel cascaded by a random amplitude filter (Fig. 1). The binary channel accounts for the probabilistic vesicle release. EO
Synaptic Trarumission: An Information-Theoretic Perspective
203
and ?l denote the probabilities of spontaneous vesicle release and failure respectively. We
follow the binary channel convention used in digital communications (? 1 = 1-p), whereas,
p is more commonly used in neurobiology. The filter h(t) is chosen to correspond to the
epsp profile of a fast AMPA-like synapse. The amplitude of the filter a is modeled as random variable with density Pea), mean J.la and standard deviation aa. The CV (standard
deviation/mean) of the distribution is denoted by eVa. We also assume that additive Gaussian voltage noise net) at the post-synaptic site further corrupts the epsp response. net) is
assumed to white with variance a~ and a bandwidth En corresponding to the membrane
time constant T. One can define an effective signal-to-noise ratio, SN R = Ea/No? given
by the ratio of the energy in the epsp pulse, Eh = 00 h2 (t) dt to the noise power spectral
density, No = a;/ En. The performance of the synapse depends on the SN R and not on
the absolute values of Eh or an. In the above model, by regarding synaptic parameters as
constants, we have tacitly ignored history dependent effects like paired-pulse facilitation,
vesicle depletion, calcium buffering. etc, which endow the synapse with the nature of a
sophisticated nonlinear filter (Markram and Tsodyks, 1997).
10
a)
m(t)
'"
'~t ",!t)
b)
N.:" V..:~
Spike
n(t)
~ SpIke
X=l
Y=l
l?P.
Effective Continuous
Estimation Channel
3
.
Effective Elinm.
Detection Channel
Figure 2: (a) Effective channel model for signal estimation. met), met), net) denote
the stimulus, the best linear estimate, and the reconstruction
noise respectively. (b) Effective
channel model for signal detection. X and Y denote the binary variables corresponding to
the input and the decision respectively. Pi and Pm are the
effective error probabilities.
Signal Estimation
Let us assume that the spike train of the presynaptic neuron can be modeled as a doubly
stochastic Poisson process with a rate A(t) = k(t) * met) given as a convolution between
the stimulus met) and a filter k(t). The stimulus is drawn from a probability distribution
which we assume to be Gaussian. k(t) = exp( -tiT) is a low-pass filter which models
the phenomenological relationship between a neuron's firing rate and its input current. T
is chosen to correspond to the membrane time constant. The exact form of k(t) is not
crucial and the above form is assumed primarily for analytical tractability. The objective is
to find the optimal estimator ofm(t) from the post-synaptic voltage v(t), where optimality
is in a least-mean square sense. The optimal mean-square estimator is, in general, nonlinear and reduces to a linear filter only when all the signals and noises are Gaussian.
However, instead of making this assumption, we restrict ourselves to the analysis of the
optimal linear estimator, met) = get) * vet), i.e. the filter get) which minimizes the
mean-square error E = (m(t) - m(t))2) where (.) denotes an ensemble average. The
overall estimation system shown in Fig. 1 can be characterized by an effective continuous
channel (Fig. 2a) where net) = met) - met) denotes the effective reconstruction noise.
System performance can be quantified by E, the lower E, the better the synapse at signal
transmission. The expression for the optimal filter (Wiener filter) in the frequency domain is
g(w) = Smv( -w)/Svv(w) where Smv(w) is the cross-spectral density (Fourier transform
of the cross-correlation Rmv) ofm(t) and set) and Svv(w) is the power spectral density of
vet). The minimum mean-square error is given by, E = a~
I Smv(w) 12 / Svv(w) dw.
The set S = {w 1 Svv (w) =J. O} is called the support of Svv (w).
- Is
A. Manwani and C. Koch
204
Another measure of system performance is the mutual information rate I (m; v) between
m(t) and v(t), defined as the rate of information transmitted by v(t) about s(t). By the
Data Processing inequality (Cover 1991), l(m, v) ~ l(m, m). A lower bound of l(m, m)
and thus of l(m; v) is given by the simple expression lib = ~
log2[~::/w/l dw (units
of bits/sec). The lower bound is achieved when n(t) is Gaussian and is independent of
m(t). Since the spike train s(t) = L 6(t - ti) is a POiSSOl!process with rate k(t) * m(t),
its power spectrum is given by the expression, Sss(w) = >'+ 1 K(w) 12 Smm(w) where
). is the mean firing rate. We assume that the mean (J..Lm) and variance (CT~) of m(t) are
chosen such that the probability that >.(t) < 0 is negligible 1 The vesicle release process
is the spike train gated by the binary channel and so it is also a Poisson process with rate
(1 - E1 )>.(t). Since v(t) = L aih(t - ti) + n(t) is a filtered P~isson process, its power
spectral density is given by Svv (w) =1 H(w) 12 {(J..L~+CT~)(1-E1)>'+J..L~(1-E1)21 K(w) 12
Smm(w)} + Snn{w). The cross-spectral density is given by the expression Svm(w) =
(1 - Et)J..LaSmm(w)H(w)K(w). This allows us to write the mean-square error as,
Is
Thus, the power spectral density ofn(t) is given by Snn = >'eff(w) + Self(w). Notice
that if K (w) ---+ 00, E ---+ 0 i. e. perfect reconstruction takes place in the limit of high
firing rates. For the parameter values chosen, SefJ{w) ? >'e//(w), and can be ignored.
Consequently, signal estimation is shot noise limited and synaptic variability increases shot
noise by a factor N syn = (1 + eVa2 ) / (1 - E1)' For eVa = 0.6 and E1 = 0.6, N syn = 3.4,
and for eVa = 1 and E1 = 0.6, N syn = 5. If m(t) is chosen to be white, band-limited to
Bm Hz, closed-form expressions for E and lib can be obtained. The expression for lib is
tedious and provides little insight and so we present only the expression for E below.
2
,1
-1
BT
E(r,BT ) = CTm [1- ~-B tan (~)l
1+,
T
+,
E is a monotonic function of, (decreasing) and BT (increasing). ,can be considered as
the effective number of spikes available per unit signal bandwidth and BT is the ratio of
the signal bandwidth and the neuron bandwidth. Plots of normalized reconstruction error
Er = E/CT~ and llb versus mean firing rate ().) for different values of signal bandwidth Bm
are shown in Fig. 3a and Fig. 3b respectively. Observe that lib (bits/sec) is insensitive to Bm
for firing rates upto 200Hz because the decrease in quality of estimation (E increases with
Bm) is compensated by an increase in the number of independent samples (2Bm) available
per second. This phenomenon is characteristic of systems operating in the low SNR regime.
lib has the generic form, llb = B log(1 + S/(N B)), where B, S and N denote signal
bandwidth, signal power and noise power respectively. For low SNR, I ~ B S / (N B) =
S / N, is independent of B. So one can argue that, for our choice of parameters, a single
synapse is a low SNR system. The analysis generalizes very easily to the case of multiple
synapses where all are driven by the same signal s (t). (Manwani and Koch, in preparation).
However, instead of presenting the rigorous analysis, we appeal to the intuition gained from
the single synapse case. Since a single synapse can be regarded as a shot noise source,
n parallel synapses can be treated as n parallel noise sources. Let us make the plausible
lWe choose pm and O'm so that
X=
30').
(std of ,X) so that Prob['x(t) ~ 0]
< 0.01.
Synaptic Transmission: An Information-Theoretic Perspective
205
assumption that these noises are uncorrelated. If optimal estimation is carried out separately
for each synapse and the estimates are combined optimally, the effective noise variance
is given by the harmonic mean of the individual variances i.e. l/u~eff = Li l/u~i.
However, if the noises are added first and optimal estimation is carried out with respect
to the sum, the effective noise variance is given by the arithmetic mean of the individual
variances, i.e. u~ef f :::: Li u~dn2. If we assume that all synapses are similar so that
U~i = u 2, u~ef f = u 2In. Plots of Er and Jib for the case of 5 identical synapses are
shown in Fig. 3c and Fig. 3d respectively. Notice that Jib increases with Bm suggesting
that the system is no longer in the low SNR regime. Thus, though a single synapse has very
low capacity, a small amount of redundancy causes a considerable increase in performance.
This is consistent with the fact the in the low S N R regime, J increases linearly with S N R ,
consequently, linearly with n, the number of synapses.
a)
b)
x x
x
x
0
0
x
x
0
o.a
~
~ + ~
..
..
+
+
+ +
000
000
0
0
0
0
X
X
X
+
x x x
X
~
e
x
o
W 0.7
B
m
= 10Hz
-
0.8
E
-
m
m
-
-
B m=75Hz
-
Bm: 100Hz
Bm= 100 Hz
0.5
20
40
~
x
o.a
0
0
l1li
l1li
100
120
140
1l1li
200
180
l1li
80
100
l1li
80
~
120
140
180
180
200
~
~
~
-
.
..
+"-: .... __ ..
+
+
..
1.
+
+
+
g
..
------
..
+
o
0
+
+
..
'- 0.8
o
W
"0 0.7
:-
12
UQ)
0
o
I
IO
(/)
0
UiS
."t:::
.0
Q)
.~
(Q o.s
-
Q)
E
o
Z
B
B=25Hz
Bm= 50Hz
B m=75HZ
-
~
=10Hz
x
o
Bm= 50 Hz
.~
Z
12
X
Bm- 25Hz
Q)m
o
X
o.s
"0
1U
1.
x x x
S
~.
0.5
.E2
.5
0 .?
20
Firing Rate (Hz)
40
~
Firing Rate (Hz)
Figure 3: Er and!,b vs. mean firing rate (X) for n = I [(a) and (b)] and n =5 [(c) and (d)] identical
synapses respectively (different values of Em) for signal estimation. Parameter values are 101 = 0.6,
100 = 0, eVa = 0.6, ts = 0.5 msec, T = I Omsec, (7n = 0.1 mY, En = 100 Hz.
4 Signal Detection
The goal in signal detection is to decide which member from a finite set of signals was
generated by a source, on the basis of measurements related to the output only in a statistical
sense. Our example corresponds to its simplest case, that of binary detection. The objective
is to derive an optimal spike detector based on the post-synaptic voltage in a given time
interval. The criterion of optimality is minimum probability of error (Pe ). A false alarm
A. Manwani and C. Koch
206
(FA) error occurs when a spike is falsely detected even when no presynaptic spike occurs
and a miss error (M) occurs when a spike fails to be detected. The probabilities of the errors
are denoted by P, and Pm respectively. Thus, Pe = (1- Po) P, +Po Pm where Po denotes
the a priori probability of a spike occurrence. Let X and Y be binary variables denoting
spike occurrence and the decision respectively. Thus, X = 1 if a spike occurred else X =
O. Similarly, Y = 1 expresses the decision that a spike occurred. The posterior likelihood
ratio is defined as ?(v) = Pr(v I X = l)/Pr(v I X = 0) and the prior likelihood as
?0 = (1 - Po)/Po. The optimal spike detector employs the well-known likelihood ratio
test, "If?(v) ~ ?0 Y=lelseY=O". When X = 1,v(t) = ah(t)+n(t) elsev(t) = n(t).
Since a is a random variable, ?(v) = (f Pr(v I X = 1; a) P(a) da)/ Pr(v I X = 0). If
the noise n( t) is Gaussian and white, it can be shown that the optimal decision rule reduces
to a matchedfilte?, i.e. if the correlation, r between v(t) and h(t) exceeds a particular
threshold (denoted by TJ), Y = 1 else Y = O. The overall decision system shown in
Fig. 1 can be treated as effective binary channel (Fig. 2b). The system perfonnance can
be quantified either by Pe or J (X; Y), the mutual infonnation between the binary random
variables, X and Y. Note that even when n(t) = 0 (SN R = 00), Pe =j:. 0 due to the
unreliability of vesicular release. Let Pe* denote the probability of error when S N R = 00.
If EO = 0, Pe* = Po El is the minimum possible detection error. Let PJ and P~ denote FA
and M errors when the release is ideal (El = 0, EO = 0). It can be shown that
Pe = Pe*
+ P~[Po(1- Ed -
+ PJ[(l - Po)(l P~ + El (1 - P~ + PI)
(1 - Po)EO]
P, = PJ ' Pm =
EO) -
PoEl]
Both PJ and P~ depend on TJ. The optimal value ofT) is chosen such that Pe is minimized.
In general, PJ and P~ can not be expressed in closed-fonn and the optimal 'f} is found using
the graphical ROC analysis procedure. Ifwe normalize a such that /-La = 1, PJ and P~ can
be parametrically expressed in tenns ofa nonnalized threshold 'f}*, PJ = 0.5[1- Er f('f}*)],
=
P~
0.5[1+ Iooo Erf(TJ* - JSNRa) P(a) da]. J(X;Y) can be computed using the
fonnula for the mutual infonnation for a binary channel, J = 1i (Po (1 - Pm) + (1 Po) P,) - Po 1i(Pm ) - (1- Po)1i(P, ) where 1i(x) -x log2 (x) - (1- x) log2(1- x) is
the binary entropy function. The analysis can be generalized to the case of n syna!Jses but
the expressions involve n-dimensional integrals which need to be evaluated numerically.
The Central Limit Theorem can be used to simplify the case of very large n. Plots of
Pe and J(X; Y) versus n for different values of SNR (1,10,00) for the case of identical
synapses are shown in Fig. 4a and Fig. 4b respectively. Yet again, we observe the poor
perfonnance of a single synapse and the substantial improvement due to redundancy. The
linear increase of J with n is similar to the result obtained for signal estimation.
=
5 Conclusions
We find that a single synapse is rather ineffective as a communication device but with
a little redundancy neuronal communication can be made much more robust. Infact, a
single synapse can be considered as a low SNR device, while 5 independent synapses
in parallel approach a high SNR system. This is consistently echoed in the results for
signal estimation and signal detection. The values of infonnation rates we obtain are very
small compared to numbers obtained from some peripheral sensory neurons (Rieke et. ai,
1996). This could be due to an over-conservative choice of parameter values on our part
or could argue for the preponderance of redundancy in neural systems. What we have
presented above are preliminary results of work in progress and so the path ahead is much
2 For deterministic a, the result is well-known, but even if a is a one-sided random variable, the
matched filter can be shown to be optimal.
Synaptic Tranrmission: An lnformation-Theoretic Perspective
a)
b)
SNR = In!.
..... SNR=10
--SNR=1
-4--
~
e
w
207
i'
~
0 ??
:0
0.
."r;:====~"'-'-------::::::::~
0.7
~
...
-4-- SNR = In!.
..... SNR=10
--SNR= 1
*
ex:
0.4
.~
0.:1
c:
E
00.2
0.'
?
0 ...
o~
, --~~2~~--~3--~--~'--~~
Number of Synapses (n)
..,
~~~--~2~----~3------~'----~
Number of Synapses (n)
Pe (a) and l,b (b) vs. the number of synapses, n, (different values of SN R) for signal detection.
SNR = Inf. corresponds to no post-synaptic voltage noise. All the synapses are assumed to be
identical. Parameter values are po = 0.5, 101
0.6, 100
0, eVa = 0.6, ts =0.5 msec, T = 10 msec,
an =0.1 mY, Bn = 100 Hz.
=
=
longer than the distance we have covered so far. To the best of our knowledge, analysis
of distinct individual components of a neuron from an communications standpoint has not
been carried out before.
Acknowledgements
This research was supported by NSF, NIMH and the Sloan Center for Theoretical Neuroscience. We thank Fabrizio Gabbiani for illuminating discussions.
References
Bekkers, J.M., Richerson, G.B. and Stevens, C.F. (1990) "Origin of variability in quantal
size in cultured hippocampal neurons and hippocampal slices," Proc. Natl. Acad. Sci. USA
87: 5359-5362.
Bialek, W. Rieke, F. van Steveninck, R.D.R. and Warland, D. (1991) "Reading a neural
code," Science 252: 1854-1857.
Cover, T.M., and Thomas, lA. (1991) Elements ofInformation Theory. New York: Wiley.
Kom, H. and Faber, D.S. (1991) "Quantal analysis and synaptic efficacy in the CNS,"
Trends Neurosci. 14: 439-445.
Markram, H. and Tsodyks, T. (1996) "Redistibution of synaptic efficacy between neocortical pyramidal neurons," Nature 382: 807-810.
Rieke, F. Warland, D. van Steveninck, R.D.R. and Bialek, W. (1996) Spikes: Exploring the
Neural Code. Cambridge: MIT Press.
Stevens, C.F. (1994) "What form should a cortical theory take," In: Large-Scale Neuronal
Theories ofthe Brain, Koch, C. and Davis, J.L., eds., pp. 239-256. Cambridge: MIT Press.
| 1477 |@word trial:2 tedious:1 pulse:2 bn:1 fonn:2 shot:3 contains:1 efficacy:3 denoting:1 current:1 yet:1 additive:2 shape:1 plot:3 v:2 device:2 record:1 filtered:1 provides:2 doubly:1 falsely:1 nor:1 brain:1 terminal:1 decreasing:1 snn:2 little:2 vertebrate:1 lib:5 increasing:1 matched:2 moreover:1 what:2 minimizes:1 ti:2 ofa:1 unit:2 christof:1 unreliability:2 before:1 negligible:1 limit:2 io:1 acad:1 encoding:1 firing:10 path:1 black:1 quantified:2 limited:2 steveninck:2 faithful:1 ofinformation:1 block:1 procedure:1 faber:2 get:2 fonnula:1 conventional:1 deterministic:1 compensated:1 center:1 estimator:7 insight:1 rule:1 regarded:1 facilitation:1 dw:2 rieke:3 transmit:2 spontaneous:2 tan:1 cultured:1 exact:1 origin:1 element:1 trend:1 std:1 observed:1 role:2 epsc:1 thousand:1 tsodyks:2 eva:5 decrease:1 substantial:1 intuition:1 nimh:1 tacitly:1 depend:1 tit:2 vesicle:8 basis:1 easily:1 po:14 various:1 train:3 distinct:1 fast:1 effective:12 detected:3 svv:6 encoded:3 kom:2 plausible:1 erf:1 transform:1 biophysical:1 net:4 analytical:1 reconstruction:5 epsp:6 poorly:1 normalize:1 transmission:5 perfect:1 derive:2 measured:2 progress:1 quantify:2 convention:1 met:7 stevens:3 filter:14 stochastic:2 pea:1 reductionist:1 eff:2 preliminary:1 dendritic:1 exploring:1 koch:7 klab:2 considered:2 deciding:1 exp:3 echoed:1 lm:1 released:1 ctm:1 estimation:18 proc:1 infonnation:8 gabbiani:1 tool:1 smv:3 mit:2 gaussian:6 vesicular:2 representa:1 rather:1 voltage:7 endow:1 release:13 focus:1 improvement:2 consistently:1 likelihood:3 rigorous:1 sense:2 tional:1 dependent:1 el:3 bt:4 pasadena:1 corrupts:1 overall:2 denoted:3 priori:1 psychophysics:1 mutual:3 identical:4 buffering:1 minimized:1 stimulus:3 simplify:1 primarily:1 employ:1 gamma:1 individual:3 ourselves:1 cns:1 detection:14 highly:1 light:1 tj:3 natl:1 integral:1 perfonnance:4 theoretical:1 lwe:1 cover:2 jib:2 measuring:1 tractability:1 deviation:2 parametrically:1 snr:14 hundred:1 optimally:1 my:2 combined:1 density:8 systematic:1 probabilistic:2 decoding:1 again:1 central:2 opposed:1 choose:1 possibly:1 li:2 account:1 potential:1 suggesting:1 coding:2 sec:2 sloan:1 depends:1 closed:2 analyze:2 observing:1 parallel:3 square:6 wiener:2 variance:7 characteristic:1 ensemble:1 correspond:2 identify:1 ofthe:1 yes:1 biophysically:1 none:1 researcher:1 history:1 ah:1 detector:6 synapsis:15 synaptic:22 email:1 ed:2 failure:2 quixote:1 energy:1 frequency:1 pp:1 e2:1 knowledge:1 amplitude:5 syn:3 sophisticated:1 ea:1 dt:1 follow:1 response:3 synapse:17 evaluated:1 box:1 though:2 stage:2 correlation:2 aih:1 nonlinear:2 quality:1 usa:1 effect:1 iooo:1 normalized:1 preponderance:1 manwani:5 white:4 skewed:1 self:1 davis:1 criterion:1 generalized:1 hippocampal:2 presenting:1 theoretic:5 neocortical:1 harmonic:1 ef:2 recently:1 spiking:1 empirically:3 vitro:1 insensitive:1 occurred:3 relating:1 l1li:5 numerically:1 significant:1 measurement:1 cambridge:2 ai:3 cv:1 pm:7 similarly:1 bekkers:2 phenomenological:1 longer:2 operating:1 etc:2 j:1 posterior:1 perspective:5 inf:1 driven:1 inequality:1 binary:13 tenns:3 caltech:2 transmitted:1 minimum:3 eo:6 paradigm:7 signal:37 rmv:1 arithmetic:1 multiple:1 reduces:2 compounded:1 exceeds:1 characterized:2 cross:3 post:8 e1:7 paired:1 schematic:1 poisson:5 llb:2 achieved:2 whereas:1 separately:1 interval:1 diagram:1 else:2 neuromuscular:1 source:3 crucial:1 standpoint:1 biased:1 pyramidal:1 ineffective:1 hz:16 member:1 presence:1 ideal:1 bandwidth:7 restrict:1 regarding:1 whether:1 expression:9 york:1 cause:1 action:1 ignored:2 generally:1 covered:1 involve:1 amount:2 band:1 simplest:1 nsf:1 notice:2 neuroscience:1 fabrizio:1 per:2 write:1 express:1 redundancy:5 threshold:2 drawn:1 pj:7 uis:1 spoke:1 sum:1 prob:1 place:1 decide:1 ofm:2 decision:6 bit:2 bound:7 ct:3 followed:1 syna:1 ahead:1 invertebrate:1 fourier:1 optimality:2 format:1 peripheral:1 poor:1 membrane:2 em:1 making:1 pr:6 sided:1 depletion:1 mechanism:1 junction:1 available:2 generalizes:1 observe:2 spectral:6 upto:1 generic:1 occurrence:2 uq:1 thomas:1 denotes:4 graphical:1 log2:3 warland:2 amit:1 objective:2 added:1 spike:25 occurs:3 fa:2 bialek:3 unclear:1 distance:1 thank:1 sci:1 capacity:6 presynaptic:4 argue:2 code:2 modeled:4 relationship:1 quantal:2 ratio:5 calcium:1 gated:1 neuron:14 convolution:1 finite:1 t:2 neurobiology:1 variability:4 communication:4 nonnalized:1 california:1 able:1 usually:1 below:1 regime:3 oft:1 reading:1 program:1 reliable:1 power:7 event:1 treated:3 eh:2 cascaded:1 technology:1 carried:3 ss:1 sn:4 prior:1 acknowledgement:1 versus:2 digital:1 h2:1 illuminating:1 consistent:1 uncorrelated:1 pi:2 supported:1 institute:1 markram:2 absolute:1 van:2 slice:1 cortical:5 dn2:1 ignores:1 sensory:1 commonly:1 made:1 bm:11 employing:1 far:1 smm:2 active:1 assumed:4 spectrum:1 infact:1 continuous:2 vet:2 quantifies:2 nature:3 transfer:1 channel:15 ca:1 robust:1 ampa:1 domain:1 da:2 linearly:3 neurosci:1 noise:18 alarm:1 profile:1 neuronal:3 fig:11 site:1 en:4 roc:1 wiley:1 fails:1 explicit:1 msec:3 pe:11 theorem:1 ofn:1 specific:1 inset:1 er:4 appeal:1 svm:1 false:1 gained:1 entropy:1 expressed:2 monotonic:1 aa:3 corresponds:2 goal:1 consequently:2 absence:1 considerable:1 miss:1 conservative:1 called:1 pas:1 la:3 zone:1 support:1 inability:1 preparation:1 bouton:1 phenomenon:1 ex:1 |
522 | 1,478 | Relative Loss Bounds for Multidimensional
Regression Problems
Jyrki Kivinen
Department of Computer Science
P.O. Box 26 (Teollisuuskatu 23)
FIN-00014 University of Helsinki, Finland
Manfred K. Warmuth
Department of Computer Science
University of California, Santa Cruz
Santa Cruz, CA 95064, USA
Abstract
We study on-line generalized linear regression with multidimensional
outputs, i.e., neural networks with multiple output nodes but no hidden
nodes. We allow at the final layer transfer functions such as the softmax function that need to consider the linear activations to all the output
neurons. We use distance functions of a certain kind in two completely
independent roles in deriving and analyzing on-line learning algorithms
for such tasks. We use one distance function to define a matching loss
function for the (possibly multidimensional) transfer function, which allows us to generalize earlier results from one-dimensional to multidimensional outputs. We use another distance function as a tool for measuring
progress made by the on-line updates. This shows how previously studied algorithms such as gradient descent and exponentiated gradient fit
into a common framework. We evaluate the performance of the algorithms using relative loss bounds that compare the loss of the on-line
algoritm to the best off-line predictor from the relevant model class, thus
completely eliminating probabilistic assumptions about the data.
1 INTRODUCTION
In a regression problem, we have a sequence of n-dimensional real valued inputs Zt E R n ,
t 1, ... ,f, and for each input Zt a k-dimensional real-valued desired output Yt E R".
Our goal is to find a mapping that at least approximately models the dependency between
Zt and Yt. Here we consider the parametric case Yt
f (w; Zt) where the actual output Yt
corresponding to the input Zt is determined by a parameter vector w E Rm (e.g., weights
in a neural network) through a given fixed model f (e.g., a neural network architecture).
=
=
1. Kivinen and M. K Wannuth
288
Thus, we wish to obtain parameters w such that, in some sense, I(w;:z:t} ~ Yt for all
t. The most basic model 1 to consider is the linear one, which in the one-dimensional
case k
1 means that I(w;:z:t)
w . :Z:t for w E Rfl. In the multidimensional case
O:Z:t. The
we actually have a whole matrix 0 E Rkxfl of parameters and 1(0;:z:t}
goodness of the fit is quantitatively measured in terms of a loss function; the square loss
given by Lt,j (Yt,j - ilt,j)2 /2 is a popular choice.
=
=
=
In generalized linear regression [MN89] we fix a transfer function 4> and apply it on top of a
?(w?:z:t). Here
linear model. Thus, in the one-dimensional case we would have I(w;:z:t)
? is usually a continuous increasing function from R to R, such as the logistic function
that maps z to 1/(1 + e- Z ). It is still possible to use the square loss, but this can lead to
problems. In particular, when we apply the logistic transfer function and try to find a weight
vector w that minimizes the total square loss over f examples (:Z:t, Yt), we may have up to
?fl local minima [AHW95, Bud93]. Hence, some other choice of loss function might be
more convenient. In the one-dimensional case it can be shown that any continuous strictly
increasing transfer function ? has a specific matching loss function LtP such that, among
other useful properties, Lt LtP(Yt, ?(w . :z:t}) is always convex in w, so local minima are
not a problem [AHW95]. For example, the matching loss function for the logistic transfer
function is the relative entropy (a generalization of the logarithmic loss for continuousvalued outcomes). The square loss is the matching loss function for the identity transfer
function (i.e., linear regression).
=
The main theme of the present paper is the application of a particular kind of distance functions to analyzing learning algorithms in (possibly multidimensional) generalized linear
regression problems. We consider a particular manner in which a mapping 4>: Rk -+ Rk
can be used to define a distance function D.4> : Rk x Rk -+ R; the assumption we must
make here is that 4> has a convex potential function. The matching loss function LtP mentioned above for a transfer function ? in the one-dimensional case is given in terms of
the distance function D.tP as LtP(?(a), ?(ii)) = D.tP(ii, a). Here, as whenever we use the
matching loss LtP (y, iI), we assume that Y and iI are in the range of ?, so we can write
Y = ?(a) and iI ?(ii) for some a and ii. Notice that for k 1, any strictly increasing
continuous function has a convex potential (i.e., integral) function. In the more interesting
case k > 1, we can consider transfer functions such as the softmax function, which is commonly used to transfer arbitrary vectors a E Rk into probability vectors y (i.e., vectors
such that iii ~ 0 for all i and Li iii 1). The matching loss function for the softmax function defined analogously with the one-dimensional case turns out to be the relative entropy
(or Kul1back-Leibler divergence), which indeed is a commonly used measure of distance
between probability vectors. For the identity transfer function, the matching loss function
is the squared Euclidean distance.
=
=
=
The first result we get from this observation connecting matching losses to a general notion
of distance is that certain previous results on generalized linear regression with matching
loss on one-dimensional outputs [HKW95] directly generalize to multidimensional outputs. From a more general point of view, a much more interesting feature of these distance
functions is how they allow us to view certain previously known learning algorithms, and
introduce new ones, in a simple unified framework. To briefly explain this framework without unnecessary complications, we restrict the foUowing discussion to the case k = 1, i.e.,
f(w;:z:) = ?(w . :z:) E R with w E Rfl.
We consider on-line learning algorithms, by which we here mean an algorithm that processes the training examples one by one, the pair (:Z:t, Yt) being processed at time t. Based
Relative Loss Bounds for Multidimensional Regression Problems
289
on the training examples the algorithm produces a whole sequence of weight vectors Wt,
= 1, ... ,f.. At each time t the old weight vector Wt is updated into WtH based on Zt and
Yt. The best-known such algorithm is on-line gradient descent. To see some alternatives,
consider first a distance function ll.,p defined on R n by some function ,p: Rn ~ Rn.
(Thus, we assume that,p has a convex potential.) We represent the update somewhat indirectly by introducing a new parameter vector 6 t ERn from which the actual weights Wt
are obtained by the mapping Wt
,p{6t ). The new parameters are updated by
t
=
(1)
where TJ > 0 is a learning rate. We call this algorithm the general additive algorithm with
parameterization function ,p. Notice that here 6 is updated by the gradient with respect
to w, so this is not just a gradient descent with reparameterization [JW98]. However, we
obtain the usual on-line gradient descent when ,p is the identity function. When,p is
the softmax function, we get the so-called exponentiated gradient (EG) algorithm [KW97 ,
HKW95].
The connection of the distance function ll.,p to the update (1) is two-fold. First, (1) can
be motivated as an approximate solution to a minimization problem in which the distance
ll.,p (6 t , 6 tH ) is used as a kind of penalty term to prevent too drastic an update based on
a single example. Second, the distance function ll.,p can be used as a potential function
in analyzing the performance of the resulting algorithm. The same distance functions have
been used previously for exactly the same purposes [KW97, HKW95] in important special
cases (the gradient descent and EG algorithms) but without realizing the full generality of
the method.
It should be noted that the choice of the parameterization function ,p is left completely
free, as long as ,p has a convex potential function. (In contrast, the choice of the transfer
function ? depends on what kind of a regression problem we wish to solve.) Earlier work
suggests that the softmax parameterization function (Le., the EG algorithm) is particularly
suited for situations in which some sparse weight vector W gives a good match to the
data [HKW95, KW97]. (Because softmax normalizes the weight vector and makes the
components positive, a simple transformation of the input data is typically added to realize
positive and negative weights with arbitrary norm.)
In work parallel to this, the analogue of the general additive update (1) in the context of
linear classification, i.e., with a threshold transfer function, has recently been developed
and analyzed by Grove et al. [GLS97] with methods and results very similar to ours. CesaBianchi [CB97] has used somewhat different methods to obtain bounds also in cases in
which the loss function does not match the transfer function. Jagota and Warmuth [JW98]
view (1) as an Euler discretization of a system of partial differential equations and investigate the performance of the algorithm as the discretization parameter approaches zero.
The distance functions we use here have previously been applied in the context of exponential families by Amari [Ama85] and others. Here we only need some basic technical
properties of the distance functions that can easily be derived from the definitions. For a
discussion of our line of work in a statistical context see Azoury and Warmuth [AW97].
In Section 2 we review the definition of a matching loss function and give examples. Section 3 discusses the general additive algorithm in more detail. The actual relative on-line
loss bounds we have for the general additive algorithm are explained in Section 4.
290
J. Kivinen and M. K. Warmuth
2 DISTANCE FUNCTIONS AND MATCIllNG LOSSES
Let
4>: R k
-t
R k be a function that has a convex potential function P4> (i.e.,
4> =
V' P4>
for some convex P 4>: Rk -4 R). We first define a distance/unction A4> for 4> by
(2)
Thus, the distance A4>(a, a) is the error we make if we approximate P4>(a) by its firstorder Taylor polynomial around a. Convexity of P 4> implies that A4> is convex in its first
argument. Further, A4>(a, a) is nonnegative, and zero if and only if 4>(a) = 4>( a).
I:
~
We can alternatively write (2) as A4>(a, a) =
(4)(r) - 4>(a)) . dr where the integral is
a path integral the value of which must be independent of the actual path chosen between
a and a. In the one-dimensional case, the integral is a simple definite integral, and ?
has a convex potential (i.e., integral) function if it is strictly increasing and continuous
[AHW95, HKW95].
Let now 4> have range V4> ~ Rk and distance function
function L4>: V4> x V4> -4 R such that L4>(4)(a) , 4>(a?
a, we say that L4> is the matching loss function for 4>.
A4>'
=
Assuming that there is a
A4>(a, a) holds for all a and
Example 1 Let 4> be a linear function given by 4>(a) = Aa where A E R kxk is symmetrical and positive definite. Then 4> has the convex potential function P4> (a) = aT Aa /2, and
(2) gives A4>(a, a) = Ha - a)T A(a - a). Hence, L4>(Y' y) = t(y - y)T A-l(y - y)
forally,YERk.
0
exp(a;)/E7=1 exp(aj), be the softmax function.
It has a potential function given by PO'(a) = In E7=1 exp(aj). To see that PO' is convex,
notice that the Hessian n 2PO' is given by D2PO'(a);j = dijO'i(a) - O'da)O'j(a). Given
Example2 Let 0': Rk
-4
Rk, O'i(a) =
a vector z E Rk, let now X be a random variable that has probability O';{a) of taking
E7=1 O'i{a)xl- E7=1 E7=1 0'; (a)xiO'j(a)xj
the value Xi? We have zTDO'(a)z
2
EX - (EX)2
VarX ~ O. Straightforward algebra now gives the relative entropy
LO'(y, y) = E;=l Yj In(Yj/Yj) as the matching loss function. (To allow Yj 0 or Yj 0,
we adopt the standard convention that OlnO
Oln(O/O)
0 and yln(y/O)
00 for
y> 0.)
0
=
=
=
=
=
=
=
=
In the relative loss bound proofs we use the basic property [JW98, Ama85]
This shows that our distances do not satisfy the triangle inequality. Usually they are not
symmetrical, either.
3 THE GENERAL ADDITIVE ALGORITHM
We consider on-line learning algorithms that at time t -first receive an input Zt E R n ,
then produce an output Yt E R k, and finally receive as feedback the desired output Yt E
Rk. To define the general additive algorithm. assume we are given a transfer function
Relative Loss Bounds for Multidimensional Regression Problems
291
l/J: Rk ~ Rk that has a convex potential function. (We wi11later use the matching loss as
a performance measure.) We also require that all the desired outputs Y t are in the range
of l/J. The algorithm's predictions are now given by Yt = l/J(Ot:et) where Ot E Rkxn is
the algorithm's weight matrix at time t. To see how the weight matrix is updated, assume
further we have a parameterization function ..p: R n ~ R n with a distance D....p. The
algorithm maintains kn real-valued parameters. We denote by 8 t the k x n matrix of the
values ofthese parameters immediately before trial t. Futher, we denote by 8 t ,i the jth row
of 8t. and by ..p(8t} the matrix with ..p(8t ,i) as its jth row. Given initial parameter values
8 1 and a learning rate 1] > 0, we now define the general additive (GA) algorithm as the
algorithm that repeats at each trial t the following prediction and update steps.
Prediction: Upon recieving the instance :et, give the prediction Yt = l/J(..p(8t):et).
Update: For j
:=
1, ... , k, set 8t+l,i = 8t,i - fJ(yt,i - Yt ,i ):et .
Note that (2) implies \7aD..l/J(a, a)) = l/J(a) -l/J(a), so this update indeed turns out to be
the same as (1) when we recall that Ll/J(Yt, Yt)
D..l/J(Ot:et, at} where Yt l/J(at).
=
=
The update can be motivated by an optimization problem given in terms of t~e loss and
distance. Consider updating an old parameter matrix 8 into a new matrix 8 based on
a single input :e and desired output y. A natural goal would be to minimize the loss
L l/J (y, l/J( ,p (8):e ) ). However, the algorithm must avoid losing too much of the information
it has gained during the previous trials and stored in the form of the old parameter matrix 8 .
We thus set as the algorithm's goal to minimize the sum D.."p(8, 8) + fJLl/J(Y' l/J(,p(8):e))
where fJ > 0 is a parameter regulating how fast the algorithm is willing to move its parameters. Under certain regularity assumptions, the update rule of the GA algorithm can
be shown to approximately solve this minimization problem. For more discussion and examples in the special case of linear regression, see [KW97]. An interesting related idea is
using all the previous examples in the update instead of just the last one. For work along
these lines in the linear case see Vovk [Vov97] and Foster [Fos91].
4 RELATIVE LOSS BOUNDS
Consider a sequence S := ((:el,yd, . .. ,(:el,Yl)) of training examples, and let
Lossl/J(GA, S)
2:!=1 Ll/J(Yt, Yt) be the loss incurred by the general additive algorithm
on this sequence when it always uses its current weights Ot for making the tth prediction
Yt? Similarly, let Lossl/J(O, S) = 2:!=1 Ll/J(Yt, l/J(O:ed) be the loss of a fixed predictor
O. Basically, our goal is to show that if some 0 achieves a small loss, then the algorithm is
not doing much worse, regardless of how the sequence S was generated. Making additional
probabilistic assumptions allows such on-line loss bounds to be converted into more traditional results about generalization errors [KW97]. To give the bounds for Lossl/J(GA, S)
in terms of Lossl/J(O, S) we need some additional parameters. The first one is the distance
D....p(8 1 ,8) where 0 = "p(8) and 8 1 is the initial parameter matrix of the GA algorithm
(which can be arbitrary). The second one is defined by
=
bx,,,p = sup {:e T n..p(9):e 19 E Rn,:e EX}
where X := {:el, ... ,:el} is the set of input vectors and n..p(9) is the Jacobian with
(D,p(9))ij
81Pi(9)/80j . The value bx,,p can be interpreted as the maximum norm of
=
J. Kivinen and M. K. Wannuth
292
any input vector in a norm defined by the parameterization function ..p. In Example 3 below
we show how bx,..p can easily be evaluated when 1/J is a linear function or the softmax
function. The third parameter ctP ' defined as
relates the matching loss function for the transfer function tP to the square loss. In Example 4 we evaluate this constant for linear functions, the softmax function, and the onedimensional case.
Example 3 Consider bounding the value
:I: TDO'( 0):1:
where 0' is the softmax function. As we saw in Example 2, this value is a variance of a random variable with the
range {Xl, ... ,Xn }. Hence, we have bx,O' ~ maXzex(maxixi - minixd 2 /4 ~
maXzex 11:l:11~ where 11:1:1100 = maXi IXil?
If 1/J is a linear function with 1/J( 8) = A8 for a symmetrical positive definite A, we clearly
0
have bx,..p ~ Amax max:l:ex :1: 2 where Amax is the largest eigenvalue of A.
Example 4 For the softmax function 0' the matching loss function LO' is the relative entropy (see Example 2), for which it is well known that LO'(y, y) 2: 2(y - y)2. Hence, we
have ctP ~ 1/4.
If tP is a linear function given by a symmetrical positive semidefinite matrix A, we see from
Example 1 that CtP is the largest eigenvalue of A.
Finally, in the special case k = 1, with ?: R -7 R differentiable and strictly increasing, we
can show ctP :::; Z if Z is a bound such that 0 < ?'(z) :::; Z holds for all z.
0
Assume now we are given constants b 2: bx ,1/J and C
for any parameter matrix 8 we have
2: ctP . Our first loss bound states that
when the learning rate is chosen as '1 = 1/(2bc). (Proofs are omitted from this extended
abstract.) The advantage of this bound is that with a fixed learning rate it holds for any
8, so we need no advance knowledge about a good 8. The drawback is the factor 2 in
front of Loss tP (..p (8), S), which suggests that asymptotically the algorithm might not ever
achieve the performance of the best fixed predictor. A tighter bound can be achieved by
more careful tuning. Thus, iven constants K 2: 0 and R > 0, if we choose the learning
rate as '1
(bcR)2 + KbcR - bcR)/(Kbc) (with '1 1/(2bc) if K 0) we obtain
=(
=
=
for any 8 that satisfies Loss tP (..p (8) , S) :::; K and d..p (8 1 , 8) :::; R. This shows that if
we restrict our comparison to parameter matrices within a given distance R of the initial
matrix of the algorithm, and we have a reasonably good guess K as to the loss of the best
fixed predictor within this distance, this knowledge allows the algorithm to asymptotically
match the performance of this best fixed predictor.
Relative Loss Bounds for Multidimensional Regression Problems
293
Acknowledgments
The authors thank Katy Azoury, Chris Bishop, Nicolo Cesa-Bianchi, David Helmbold, and
Nick Littlestone for helpful discussions. Jyrki Kivinen was supported by the Academy of
Finland and the ESPRIT project NeuroCOLT. Manfred Warmuth was supported by the NSF
grant CCR 9700201.
References
[Ama85]
S. Amari. Differential Geometrical Methods in Statistics. Springer Verlag,
Berlin, 1985.
[AHW95] P. Auer, M. Herbster, and M. K. Warmuth. Exponentially many local minima
for single neurons. In Proc. 1995 Neural Information Processing Conference,
pages 316-317. MIT Press, Cambridge, MA, November 1995.
[AW97]
K. Azoury and M. K. Warmuth. Relative loss bounds and the exponential family
of distributions. Unpublished manuscript, 1997.
[Bud93]
M. Budinich. Some notes on perceptron learning. 1. Phys. A.: Math. Gen.,
26:4237-4247, 1993 .
[CB97]
N. Cesa-Bianchi. Analysis of two gradient-based algorithms for on-line regression. In Proc. 10th Annu. Con! on Comput. Learning Theory, pages 163-170.
ACM,1997.
[Fos91]
D. P. Foster. Prediction in the worst case. The Annals of Statistics, 19(2):10841090, 1991.
[GLS97]
A. J. Grove, N. Littlestone, and D. Schuurmans. General convergence results
for linear discriminant updates. In Proc. 10th Annu. Con! on Comput. Learning
Theory, pages 171-183. ACM, 1997.
[HKW95] D. P. Helmbold, J. Kivinen, and M. K. Warmuth. Worst-case loss bounds for
sigmoided linear neurons. In Proc. Neural Information Processing Systems
1995, pages 309-315. MIT Press, Cambridge, MA, November 1995.
[JW98]
A. K. Jagota and M. K. Warmuth. Continuous versus discrete-time nonlinear
gradient descent: Relative loss bounds and convergence. Presented at Fifth Symposium on Artificial Intelligence and Mathematics, Ft. Lauderdale, FL, 1998.
[KW97]
J. Kivinen and M. K. Warmuth. Additive versus exponentiated gradient updates for linear prediction. Information and Computation, 132( 1): 1-64, January
1997.
[MN89]
P. McCullagh and J. A. NeIder. Generalized Linear Models. Chapman & Hall,
New York, 1989.
[Vov97]
V. Vovk. Competitive on-line linear regression. In Proc. Neural Information
Processing Systems 1997. MIT Press, Cambridge, MA, 1998.
| 1478 |@word trial:3 briefly:1 eliminating:1 polynomial:1 norm:3 willing:1 initial:3 ours:1 bc:2 current:1 discretization:2 unction:1 varx:1 activation:1 must:3 realize:1 ixil:1 cruz:2 additive:9 update:13 intelligence:1 guess:1 warmuth:10 parameterization:5 realizing:1 manfred:2 wth:1 math:1 node:2 complication:1 along:1 differential:2 symposium:1 introduce:1 manner:1 indeed:2 continuousvalued:1 actual:4 increasing:5 project:1 what:1 kind:4 interpreted:1 minimizes:1 iven:1 developed:1 unified:1 transformation:1 multidimensional:10 firstorder:1 exactly:1 esprit:1 rm:1 gls97:2 grant:1 positive:5 before:1 local:3 analyzing:3 path:2 approximately:2 yd:1 might:2 studied:1 suggests:2 range:4 acknowledgment:1 yj:5 definite:3 matching:16 convenient:1 get:2 ga:5 context:3 map:1 yt:23 straightforward:1 regardless:1 convex:12 immediately:1 helmbold:2 rule:1 amax:2 deriving:1 reparameterization:1 notion:1 updated:4 annals:1 losing:1 us:1 particularly:1 updating:1 role:1 ft:1 worst:2 oln:1 mentioned:1 convexity:1 xio:1 algebra:1 upon:1 completely:3 triangle:1 easily:2 po:3 fast:1 artificial:1 outcome:1 valued:3 solve:2 say:1 amari:2 statistic:2 final:1 sequence:5 eigenvalue:2 differentiable:1 advantage:1 p4:4 relevant:1 gen:1 achieve:1 academy:1 convergence:2 regularity:1 produce:2 measured:1 ij:1 progress:1 implies:2 convention:1 drawback:1 mn89:2 require:1 fix:1 generalization:2 tighter:1 strictly:4 hold:3 around:1 hall:1 exp:3 mapping:3 finland:2 adopt:1 achieves:1 omitted:1 purpose:1 proc:5 saw:1 largest:2 tool:1 minimization:2 mit:3 clearly:1 always:2 e7:5 avoid:1 hkw95:6 derived:1 contrast:1 sense:1 helpful:1 el:4 typically:1 hidden:1 algoritm:1 among:1 classification:1 softmax:11 special:3 chapman:1 others:1 quantitatively:1 divergence:1 regulating:1 investigate:1 analyzed:1 semidefinite:1 tj:1 grove:2 integral:6 partial:1 euclidean:1 old:3 taylor:1 desired:4 littlestone:2 instance:1 earlier:2 tp:6 measuring:1 goodness:1 introducing:1 euler:1 predictor:5 too:2 front:1 stored:1 dependency:1 kn:1 herbster:1 probabilistic:2 off:1 v4:3 yl:1 lauderdale:1 analogously:1 connecting:1 squared:1 cesa:2 choose:1 possibly:2 dr:1 worse:1 bx:6 li:1 potential:10 converted:1 satisfy:1 depends:1 ad:1 try:1 view:3 neider:1 doing:1 sup:1 competitive:1 maintains:1 parallel:1 minimize:2 square:5 variance:1 generalize:2 basically:1 explain:1 phys:1 teollisuuskatu:1 whenever:1 ed:1 definition:2 proof:2 con:2 popular:1 recall:1 knowledge:2 actually:1 auer:1 manuscript:1 evaluated:1 box:1 generality:1 just:2 nonlinear:1 logistic:3 aj:2 usa:1 hence:4 leibler:1 eg:3 ll:7 during:1 noted:1 generalized:5 fj:2 geometrical:1 recently:1 common:1 sigmoided:1 exponentially:1 onedimensional:1 cambridge:3 tuning:1 mathematics:1 similarly:1 nicolo:1 certain:4 verlag:1 inequality:1 jagota:2 minimum:3 additional:2 somewhat:2 cesabianchi:1 ii:7 relates:1 multiple:1 full:1 technical:1 match:3 long:1 prediction:7 regression:14 basic:3 represent:1 achieved:1 receive:2 ot:4 ltp:5 call:1 iii:2 xj:1 fit:2 architecture:1 restrict:2 idea:1 motivated:2 penalty:1 hessian:1 york:1 useful:1 santa:2 bcr:2 processed:1 tth:1 nsf:1 notice:3 ccr:1 discrete:1 write:2 threshold:1 prevent:1 asymptotically:2 sum:1 ilt:1 family:2 wannuth:2 bound:18 layer:1 fl:2 fold:1 nonnegative:1 helsinki:1 argument:1 ern:1 department:2 kw97:6 ofthese:1 making:2 explained:1 equation:1 previously:4 turn:2 discus:1 drastic:1 apply:2 indirectly:1 alternative:1 top:1 a4:8 move:1 added:1 parametric:1 usual:1 traditional:1 gradient:11 distance:27 thank:1 neurocolt:1 berlin:1 chris:1 discriminant:1 assuming:1 negative:1 tdo:1 zt:7 bianchi:2 neuron:3 observation:1 fin:1 descent:6 yln:1 november:2 january:1 situation:1 rfl:2 extended:1 ever:1 rn:3 arbitrary:3 david:1 pair:1 unpublished:1 connection:1 nick:1 california:1 usually:2 below:1 max:1 analogue:1 natural:1 kivinen:7 review:1 relative:14 loss:48 interesting:3 versus:2 incurred:1 foster:2 pi:1 normalizes:1 lo:3 row:2 repeat:1 last:1 free:1 supported:2 jth:2 allow:3 exponentiated:3 perceptron:1 taking:1 fifth:1 sparse:1 feedback:1 xn:1 author:1 made:1 commonly:2 approximate:2 symmetrical:4 unnecessary:1 xi:1 alternatively:1 continuous:5 ctp:5 transfer:16 reasonably:1 ca:1 schuurmans:1 da:1 main:1 azoury:3 whole:2 bounding:1 theme:1 wish:2 exponential:2 xl:2 comput:2 kbc:1 jacobian:1 third:1 rk:13 annu:2 example2:1 specific:1 bishop:1 maxi:1 gained:1 suited:1 entropy:4 lt:2 logarithmic:1 kxk:1 katy:1 springer:1 aa:2 futher:1 a8:1 satisfies:1 acm:2 ma:3 goal:4 jyrki:2 identity:3 careful:1 mccullagh:1 determined:1 vovk:2 wt:4 total:1 called:1 l4:4 evaluate:2 ex:4 |
523 | 1,479 | Unsupervised On-Line Learning of
Decision Trees for Hierarchical Data
Analysis
Marcus Held and Joachim M. Buhmann
Rheinische Friedrich-Wilhelms-U niversitat
Institut fUr Informatik III, ROmerstraBe 164
D-53117 Bonn, Germany
email: {held.jb}.cs.uni-bonn.de
WWW: http://www-dbv.cs.uni-bonn.de
Abstract
An adaptive on-line algorithm is proposed to estimate hierarchical
data structures for non-stationary data sources. The approach
is based on the principle of minimum cross entropy to derive a
decision tree for data clustering and it employs a metalearning idea
(learning to learn) to adapt to changes in data characteristics. Its
efficiency is demonstrated by grouping non-stationary artifical data
and by hierarchical segmentation of LANDSAT images.
1
Introduction
Unsupervised learning addresses the problem to detect structure inherent in unlabeled and unclassified data. The simplest, but not necessarily the best approach for extracting a grouping structure is to represent a set of data samples
= {Xi E Rdli = 1, ...
by a set of prototypes y = {Ya E Rdlo = 1, .. .
K ? N. The encoding usually is represented by an assignment matrix M = (Mia),
where Mia = 1 if and only if Xi belongs to cluster 0, and Mia = 0 otherwise. According to this encoding scheme, the cost function 1i (M, Y) = ~ L:~1 MiaV (Xi, Ya)
measures the quality of a data partition, Le., optimal assignments and prototypes
(M,y)OPt = argminM,y1i (M,Y) minimize the inhomogeneity of clusters w.r.t. a
given distance measure V. For reasons of simplicity we restrict the presentation
to the ' sum-of-squared-error criterion V(x, y) = !Ix - YI12 in this paper. To facilitate this minimization a deterministic annealing approach was proposed in [5]
which maps the discrete optimization problem, i.e. how to determine the data assignments, via the Maximum Entropy Principle [2] to a continuous parameter es-
X
,N}
,K},
Unsupervised On-line Learning of Decision Trees for Data Analysis
515
timation problem. Deterministic annealing introduces a Lagrange multiplier {3 to
control the approximation of 11. (M, Y) in a probabilistic sense. Equivalently to
maximize the entropy at fixed expected K-means costs we minimize the free energy
:F = ~ 2:f::1ln (2::=1 exp (-{3V (Xi, Ya:))) w.r.t. the prototypes Ya:. The assignments Mia: are treated as random variables yielding a fuzzy centroid rule
N
N
Ya: = L i=l (Mia:)xdLi=l (Mia:) ,
where the expected assignments (Mia:) are given by Gibbs distributions
(Mia:)
=
:x
p (-{3V (Xi,Ya:))
(1)
(2)
.
2:/l=1 exp ( -{3V (Xi, Ya:))
For a more detailed discussion of the DA approach to data clustering cf. [1, 3, 5].
In addition to assigning data to clusters (1,2), hierarchical clustering provides the
partitioning of data space with a tree structure. Each data sample X is sequentially
assigned to a nested structure of partitions which hierarchically cover the data
space Rd. This sequence of special decisions is encoded by decision rules which are
attached to nodes along a path in the tree (see also fig. 1).
Therefore, learning a decision tree requires to determine a tree topology, the accompanying assignments, the inner node labels S and the prototypes y at the leaves.
The search of such a hierarchical partition of the data space should be guided by
an optimization criterion, i.e., minimal distortion costs.
This problem is solvable by a two-stage approach, which on the one hand minimizes
the distortion costs at the leaves given the tree structure and on the other hand
optimizes the tree structure given the leaf induced partition of Rd. This approach,
due to Miller & Rose [3], is summarized in section 2. The extensions for adaptive online learning and experimental results are described in sections 3 and 4, respectively.
x
S
~
S
S
/4\
j'\
a
/'\
S
b
c
d
e
partition
ofdata space
?
a
f
Figure 1: Right: Topology of a decision tree. Left: Induced partitioning of the
data space (positions of the letters also indicate the positions of the prototypes).
Decisions are made according to the nearest neighbor rule.
2
Unsupervised Learning of Decision Trees
Deterministic annealing of hierarchical clustering treats the assignments of data to
inner nodes of the tree in a probabilistic way analogous to the expected assignments
of data to leaf prototypes. Based on the maximum entropy principle, the probability
~~j that data point Xi reaches inner node Sj is recursively defined by (see [3]):
~~root:=
1,
~~j
=
~~parent(j)1ri,j,
1ri,j
=
exp (-,V(Xi,Sj))
2:
exp(-,V(Xi,Sk)) ,
kEsiblings(j)
(3)
M. Held and 1. M. Buhmann
516
where the Lagrange multiplier, controls the fuzziness of all the transitions 1fi,j'
On the other hand, given the tree topology and the prototypes at the leaves, the
maximum entropy principle naturally recommends an ideal probability cpLl< at leaf
Yet, resp. at an inner node sj>
cp~
=
exp(-j1V(Xi,Yet))
exp(-j1V(Xi'Y/L))
',et
and
cpl.
L:
:E
=
kEdescendants(j)
l,J
(4)
CP!k'
,
/LEY
We apply the principle of minimum cross entropy for the calculation of the prototypes at the leaves given a priori the probabilities for the parents of the leaves. Minimization of the cross entropy with fixed expected costs (HXi) = L:et (Miet)V (Xi, Yet)
for the data point Xi yields the expression
m~n I({(Miet)}II{Cp~parent(et)/K}) =
{(M.e?}
min Let (Miet) In
{(M.e>)}
cpJMiet)
,
i,parent(et)
(5)
where I denotes the Kullback-Leibler divergence and K defines the degree of the
inner nodes. The tilted distribution
Cp~parent(et) exp (-j1V (Xi, Yet))
(
(6)
Miet) =
H
.
L: /L cp i,parent(/L) exp ( - j1V (Xi, Y/L))
generalizes the probabilistic assignments (2). In the case of Euclidian distances
we again obtain the centroid formula (1) as the minimum of the free energy
h
F = - L::l ln [L:etEY cprparent(et) exp (-j1V (Xi, Yet))]. Constraints induced by
the tree structure are incorporated in the assignments (6). For the optimization
of the hierarchy, Miller and Rose in a second step propose the minimization of the
distance between the hierarchical probabilities CP~. and the ideal probabilities Cp~ .,
the distance being measured by the Kullback-Leibler divergence
'
~T
L
,
BjEparent(Y)
I ({ Cp~,j }II{ Cp~j}) == ~,W
L
cp~ .
N
:E
cpL
BjEparent(Y)i=l
In cp~J
.
(7)
t,J
Equation (7) describes the minimization of the sum of cross entropies between the
probability densities CP~. and CP~. over the parents of the leaves. Calculating the
gradients for the inner ;'odes S j ~d the Lagrange multiplier, we receive
N
-2, L (Xi - Sj)
N
{cpL - cp!,parent(j)1fi,j} :=
-2,:E ~1 (Xi, Sj), (8)
i=l
i=l
N
L:E V (Xi, Sj)
i=l jES
N
{cpL - cpLparent(j)1fi,j} :=
L:E ~2 (Xi, Sj).
(9)
i=l jES
The first gradient is a weighted average of the difference vectors (Xi - Sj), where the
weights measure the mismatch between the probability CPtj and the probability induced by the transition 1fi,j' The second gradient (9) measures the scale - V (Xi, Sj)
- on which the transition probabilities are defined, and weights them with the mismatch between the ideal probabilities. This procedure yields an algorithm which
starts at a small value j1 with a complete tree and identical test vectors attached
to all nodes. The prototypes at the leaves are optimized according to (6) and the
centroid rule (1), and the hierarchy is optimized by (8) and (9). After convergence
one increases j1 and optimizes the hierarchy and the prototypes at the leaves again.
The increment of j1leads to phase transitions where test vectors separate from each
other and the formerly completely degenerated tree evolves its structure. For a
detailed description of this algorithm see [3].
Unsupervised On-line Learning of Decision Trees for Data Analysis
3
517
On-Line Learning of Decision Trees
Learning of decision trees is refined in this paper to deal with unbalanced trees
and on-line learning of trees. Updating identical nodes according to the gradients
(9) with assignments (6) weighs parameters of unbalanced tree structures in an
unsatisfactory way. A detailed analysis reveals that degenerated test vectors, i.e.,
test vectors with identical components, still contribute to the assignments and to
the evolution of /. This artefact is overcome by using dynamic tree topologies
instead of a predefined topology with indistinguishable test vectors. On the other
hand, the development of an on-line algorithm makes it possible to process huge
data sets and non-stationary data. For this setting there exists the need of on-line
learning rules for the prototypes at the leaves, the test vectors at the inner nodes
and the parameters / and (3. Unbalanced trees also require rules for splitting and
merging nodes.
Following Buhmann and Kuhnel [1] we use an expansion of order O(I/n) of (1) to
estimate the prototypes for the Nth datapoint
N '" N-l
Ya '" Ya
+ 'TJa (M:;;I)
N-1M
POI
(
N-I)
XN - Ya
,
(10)
where P~ ~ p~-1 +1/M ((M:;;I) - p~-l) denotes the probability of the occurence
of class o. The parameters M and'TJa are introduced in order to take the possible
non-stationarity of the data source into account. M denotes the size of the data
window, and 'TJa is a node specific learning rate.
Adaptation of the inner nodes and of the parameter / is performed by stochastic
approximation using the gradients (8) and (9)
(11)
(12)
For an appropriate choice of the learning rates 'TJ, the learning to learn approach of
Murata et al. [4] suggests the learning algorithm
(13)
The flow 1 in parameter space determines the change of w N -1 given a new datapoint
XN. Murata et al. derive the following update scheme for the learning rate:
rN
'TJN
(1 - 8)r N - 1
_
'TJ N -
I
+ 81 (XN, W N - 1 )
+ Vl",N-l
,
(v2I1rNII- 'TJN-l) ,
(14)
(15)
where VI, v2 and 8 are control parameters to balance the tradeoff between accuracy
and convergence rate. r N denotes the leaky average of the flow at time N.
The adaptation of (3 has to observe the necessary condition for a phase transition
(3 > (3erit == 1/28rnax , 8rnax being the largest eigenvalue of the covariance matrix [3]
M
~a
=L
i=l
M
(Xi - Ya) (Xi - Ya)t (Mia)/L(Mia ).
(16)
i=l
Rules for splitting and merging nodes of the tree are introduced to deal with unbalanced trees and non-stationary data. Simple rules measure the distortion costs
at the prototypes of the leaves. According to these costs the leaf with highest
M Held and 1. M Buhmann
518
distortion costs is split. The merging criterion combines neighboring leaves with
minimal distance in a greedy fashion. The parameter M (10), the typical time
scale for changes in the data distribution is used to fix the time between splitting
resp. merging nodes and the update of (3. Therefore, M controls the time scale for
changes of the tree topology. The learning parameters for the learning to learn rules
(13)-(15) are chosen empirically and are kept constant for all experiments .
4
Experiments
The first experiment demonstrates how a drifting two dimensional data source can
be tracked. This data source is generated by a fixed tree augmented with transition
probabilities at the edges and with Gaussians at the leaves. By descending the tree
structure this ~enerates an Li.d. random variable X E R2, which is rotated around
the origin of R to obtain a random variable T(N) = R(w, N)X . R is an orthogonal
matrix, N denotes the number of the actual data point and w denotes the angular
velocity, M = 500. Figure 2 shows 45 degree snapshots of the learning of this nonstationary data source. We start to take these snapshots after the algorithm has
developed its final tree topology (after ~ 8000 datapoints). Apart from fluctuations
of the test vectors at the leaves, the whole tree structure is stable while tracking
the rotating data source.
Additional experiments with higher dimensional data sources confirm the robustness
of the algorithm w.r.t. the dimension of the data space, i. e. similiar tracking
performances for different dimensions are observed, where differences are explained
as differences in the data sources (figure 3) . This performance is measured by the
variance of the mean of the distances between the data source trajectory and the
trajectories of the test vectors at the nodes of the tree.
7)
8)
Figure 2: 45 degree snapshots of the learning of a data source which rotates with a
velocity w = 271"/30000 (360 degree per 30000 data samples:.
A second experiment demonstrates the learning of a switching data source. The
results confirm a good performance concerning the restructuring of the tree (see
figure 4). In this experiment the algorithm learns a given data source and after
10000 data points we switch to a different source.
As a real-world example of on-line learning of huge data sources the algorithm is
applied to the hierarchical clustering of 6- dimensional LANDSAT data. The heat
519
Unsupervised On-line Learning of Decision Trees for Data Analysis
-0.5
?1
' 1.5
2dim4dim
~ .--
12dim .....
18dim
I~
.
~
.
-3.5
-4
-4.5
?5
?5.5
~
10000
0
20000
30000
40000
50000
60000
70000
80000
90000
N
Figure 3: Tracking performance for different dimensions. As data sources we use
d-dimensional Gaussians which are attached to a unit sphere. To the components
of every random sample X we add sin(wN) in order to introduce non stationarity.
The first 8000 samples are used for the development of the tree topology.
~,
A)
\
a
\
I
k
1\ "'-,
b
B)
.
\
. \ _,_____. '" ._________
\ .
.>,\"
\L
_.-" -'" -".(----?/l
<'
I:'
e
C
J
J
i
/
,/
/
'
"\"
'\
\
\
c
\\
1-
1 ""., .
\
1
~,
,--
..____-
__}-;-''-.--~.~
Ig -
j
\
-----b~~<\ :'
'
\
/
/
.? ' (
? -\--<I!..~
\
\'
.
k
.9"
_ ",
h \
v'
.'6>
h\,m
\\Q\
\
..__
- >I--."/_-J--;
#-\01f;;<1
\ ;r" " n'/
,
l/
"
~""
\
1
1:1'"
tf '\'
~
f
____
e '-
I
\"
\
\
II
\1
\
~
a
b
m
J k 1
def g nohi
Figure 4: Learning a switching data source: top: a) the partition of the data space
after 10000 data samples given the first source, b) the restructured partition after
additional 2500 samples. Below: accompanying tree topologies.
channel has been discarded because of its reduced resolution. In a preprocessing step
all channels are rescaled to unit variance, which alternatively could be established
by using a Mahalanobis distance. Note that the decision tree which clusters this
data supplies us with a hierarchical segmentation of the corresponding LANDSAT
image. A tree of 16 leaves has been learned on a training set of 128 x 128 data
samples, and it has been applied to a test set of 128 x 128 LANDSAT pixels. The
training is established by 15 sequential runs through the test set, where after each
M = 16384 run a split of one node is carried out. The resulting empirical errors
(0.49 training distortion and 0.55 test distortion) differ only slightly from the errors
obtained by the LBG algorithm applied to the whole training set (0.42 training
distortion and 0.52 test distortion). This difference is due to the fact that not
every data point is assigned to the nearest leaf prototype by a decision tree induced
partition. The segmentation of the test image is depicted in figure 5.
5
Conclusion
This paper presents a method for unsupervised on-line learning of decision trees.
We overcome the shortcomings of the original decision tree approach and extend
520
M Held and J. M. Buhmann
Figure 5: Hierarchical segmentation of the test image. The root represents the
original image, i.e., the gray scale version of the three color channels.
it to the realm of on-line learning of huge data sets and of adaptive learning of
non-stationary data. Our experiments demonstrate that the approach is capable of
tracking gradually changing Or switching environments. Furthermore, the method
has been successfully applied to the hierarchical segmentation of LANDSAT images.
Future work will address active data selection issues to significantly reduce the
uncertainty of the most likely tree parameters and the learning questions related to
different tree topologies.
Acknowledgement: This work has been supported by the German Israel Foundation
for Science and Research Development (GIF) under grant #1-0403-001.06/95 and by the
Federal Ministry for Education, Science and Technology (BMBF #01 M 3021 A/4).
References
[1] J. M. Buhmann and H. Kuhnel. Vector quantization with complexity costs. IEEE
Transactions on Information Theory, 39(4):1133-1145, July 1993.
[2] T.M. Cover and J. Thomas. Elements of Information Theory. Wiley & Sons, 1991.
[3] D. Miller and K Rose. Hierarchical unsupervised learning with growing via phase
transitions. Neural Computation, 8:425-450, February 1996.
[4] N. Murata, K-R. Muller, A. Ziehe, and S. Amari . Adaptive on-line learning in changing
environments. In M.C. Mozer, M.I. Jordan, and T. Petsche, editors, Advances in Neural
Information Processing Systems, number 9, pages 599-605. MIT Press, 1997.
[5] K Rose, E. Gurewitz, and G.C. Fox. A deterministic annealing approach to clustering.
Pattern Recognition Letters, 11(9):589-594, September 1990.
| 1479 |@word version:1 covariance:1 euclidian:1 recursively:1 assigning:1 yet:5 tilted:1 partition:8 j1:2 update:2 stationary:5 greedy:1 leaf:19 provides:1 contribute:1 node:16 along:1 supply:1 combine:1 introduce:1 expected:4 growing:1 actual:1 window:1 israel:1 minimizes:1 gif:1 fuzzy:1 developed:1 every:2 demonstrates:2 control:4 partitioning:2 unit:2 grant:1 treat:1 switching:3 encoding:2 path:1 fluctuation:1 cpl:4 suggests:1 procedure:1 empirical:1 significantly:1 unlabeled:1 selection:1 descending:1 www:2 argminm:1 map:1 deterministic:4 demonstrated:1 resolution:1 simplicity:1 splitting:3 rule:9 datapoints:1 increment:1 analogous:1 resp:2 hierarchy:3 origin:1 velocity:2 element:1 recognition:1 updating:1 observed:1 highest:1 rescaled:1 rose:4 mozer:1 environment:2 complexity:1 dynamic:1 efficiency:1 completely:1 represented:1 heat:1 shortcoming:1 refined:1 encoded:1 distortion:8 otherwise:1 amari:1 inhomogeneity:1 final:1 online:1 sequence:1 eigenvalue:1 propose:1 adaptation:2 neighboring:1 description:1 parent:8 cluster:4 convergence:2 rotated:1 derive:2 measured:2 nearest:2 c:2 indicate:1 differ:1 artefact:1 guided:1 ley:1 stochastic:1 education:1 require:1 fix:1 opt:1 extension:1 accompanying:2 around:1 exp:9 label:1 largest:1 tf:1 successfully:1 weighted:1 minimization:4 federal:1 mit:1 poi:1 joachim:1 rheinische:1 fur:1 unsatisfactory:1 centroid:3 detect:1 sense:1 dim:2 landsat:5 vl:1 germany:1 pixel:1 issue:1 priori:1 development:3 special:1 identical:3 represents:1 unsupervised:8 future:1 jb:1 inherent:1 employ:1 divergence:2 phase:3 stationarity:2 huge:3 introduces:1 yielding:1 tj:2 held:5 predefined:1 edge:1 capable:1 niversitat:1 necessary:1 institut:1 orthogonal:1 tree:43 fox:1 rotating:1 weighs:1 minimal:2 cover:2 assignment:12 cost:9 density:1 probabilistic:3 squared:1 again:2 tjn:2 kuhnel:2 li:1 account:1 de:2 summarized:1 vi:1 performed:1 root:2 start:2 minimize:2 accuracy:1 variance:2 characteristic:1 wilhelms:1 miller:3 yield:2 murata:3 informatik:1 trajectory:2 mia:10 datapoint:2 metalearning:1 reach:1 email:1 energy:2 naturally:1 color:1 realm:1 segmentation:5 jes:2 higher:1 furthermore:1 angular:1 stage:1 hand:4 defines:1 quality:1 gray:1 facilitate:1 multiplier:3 evolution:1 assigned:2 leibler:2 deal:2 mahalanobis:1 indistinguishable:1 sin:1 criterion:3 complete:1 demonstrate:1 cp:14 image:6 fi:4 empirically:1 tracked:1 attached:3 extend:1 gibbs:1 rd:2 hxi:1 stable:1 add:1 belongs:1 optimizes:2 apart:1 muller:1 minimum:3 additional:2 ministry:1 determine:2 maximize:1 july:1 ii:3 adapt:1 calculation:1 cross:4 sphere:1 concerning:1 represent:1 receive:1 addition:1 ode:1 annealing:4 source:17 induced:5 flow:2 jordan:1 extracting:1 nonstationary:1 ideal:3 iii:1 recommends:1 split:2 wn:1 switch:1 topology:10 restrict:1 inner:8 idea:1 prototype:14 reduce:1 tradeoff:1 expression:1 detailed:3 simplest:1 reduced:1 http:1 per:1 discrete:1 changing:2 kept:1 sum:2 run:2 unclassified:1 letter:2 uncertainty:1 decision:17 def:1 constraint:1 ri:2 y1i:1 bonn:3 min:1 according:5 describes:1 slightly:1 son:1 evolves:1 explained:1 gradually:1 ln:2 equation:1 german:1 generalizes:1 gaussians:2 apply:1 observe:1 hierarchical:12 v2:1 appropriate:1 petsche:1 robustness:1 drifting:1 original:2 thomas:1 denotes:6 clustering:6 cf:1 top:1 calculating:1 february:1 question:1 september:1 gradient:5 distance:7 separate:1 rotates:1 reason:1 marcus:1 degenerated:2 balance:1 equivalently:1 snapshot:3 discarded:1 similiar:1 incorporated:1 rn:1 introduced:2 friedrich:1 optimized:2 learned:1 established:2 address:2 usually:1 etey:1 mismatch:2 below:1 pattern:1 rnax:2 yi12:1 treated:1 buhmann:6 solvable:1 nth:1 scheme:2 technology:1 carried:1 occurence:1 gurewitz:1 formerly:1 acknowledgement:1 foundation:1 degree:4 principle:5 editor:1 supported:1 free:2 neighbor:1 leaky:1 overcome:2 dimension:3 xn:3 transition:7 world:1 made:1 adaptive:4 preprocessing:1 ig:1 transaction:1 sj:9 uni:2 kullback:2 confirm:2 sequentially:1 reveals:1 active:1 xi:24 alternatively:1 continuous:1 search:1 timation:1 sk:1 tja:3 learn:3 channel:3 expansion:1 necessarily:1 da:1 hierarchically:1 whole:2 augmented:1 fig:1 fashion:1 wiley:1 bmbf:1 position:2 lbg:1 ix:1 learns:1 formula:1 specific:1 r2:1 grouping:2 exists:1 quantization:1 merging:4 sequential:1 entropy:8 depicted:1 likely:1 lagrange:3 restructuring:1 tracking:4 nested:1 determines:1 presentation:1 fuzziness:1 change:4 typical:1 e:1 ya:12 experimental:1 ziehe:1 romerstrabe:1 unbalanced:4 artifical:1 |
524 | 148 | 124
ADAPTIVE NEURAL NET PREPROCESSING
FOR SIGNAL DETECTION
IN NON-GAUSSIAN NOISE1
Richard P. Lippmann and Paul Beckman
MIT Lincoln Laboratory
Lexington, MA 02173
ABSTRACT
A nonlinearity is required before matched filtering in mInimum error
receivers when additive noise is present which is impulsive and highly
non-Gaussian. Experiments were performed to determine whether the
correct clipping nonlinearity could be provided by a single-input singleoutput multi-layer perceptron trained with back propagation. It was
found that a multi-layer perceptron with one input and output node, 20
nodes in the first hidden layer, and 5 nodes in the second hidden layer
could be trained to provide a clipping nonlinearity with fewer than 5,000
presentations of noiseless and corrupted waveform samples. A network
trained at a relatively high signal-to-noise (SIN) ratio and then used
as a front end for a linear matched filter detector greatly reduced the
probability of error. The clipping nonlinearity formed by this network
was similar to that used in current receivers designed for impulsive noise
and provided similar substantial improvements in performance.
INTRODUCTION
The most widely used neural net, the adaptive linear combiner (ALe). is a singlelayer perceptron with linear input and output nodes. It is typically trained using the
LMS algorithm and forms one of the most common components of adaptive filters.
ALes are used in high-speed modems to construct equalization filters, in telephone
links as echo cancelers, and in many other signal processing applications where linear
filtering is required [9]. The purpose of this study was to determine whether multilayer perceptrons with linear input and output nodes but with sigmoidal hidden
nodes could be as effective for adaptive nonlinear filtering as ALes are for linear
filtering.
1 This work wa.s sponsored by the Defense Advanced Research Projects Agency and the Department of the Air Force . The views expressed are those of the authors and do not reflect the policy
or position of the U . S. Government.
Adaptive Neural Net Preprocessing for Signal Detection
The task explored in this paper is signal detection with impulsive noise where an
adaptive nonlinearity is required for optimal performance. Impulsive noise occurs
in underwater acoustics and in extremely low frequency communications channels
where impulses caused by lightning strikes propagate many thousands of miles [2].
This task was selected because a nonlinearity is required in the optimal receiver, the
structure of the optimal receiver is known, and the resulting signal detection error
rate provides an objective measure of performance. The only other previous studies
of the use of multi-layer perceptrons for adaptive nonlinear filtering that we are
aware of [6,8] appear promising but provide no objective performance comparisons.
In the following we first present examples which illustrate that multi-layer perceptrons trained with back-propagation can rapidly form clipping and other nonlinearities useful for signal processing with deterministic training. The signal detection
task is then described and theory is presented which illustrates the need for nOlllinear processing with non-Gaussian noise. Nonlinearities formed when the input
to a net is a corrupted signal and the desired output is the uncorrupted signal are
then presented for no noise, impulsive noise, and Gaussian noise. Finally, signal
detection performance results are presented that demonstrate large improvements
in performance with an adaptive nonlinearity and impulsive noise.
FORMING DETERMINISTIC NONLINEARITIES
A theorem proven by Kohnogorov and described in [5] demonstrates that singleinput single-output continuous nonlinearities can be formed by a multi-layer perceptron with two layers of hidden nodes. This proof, however, requires complex nonlinear functions in the hidden nodes that are very sensitive to the desired input/output
function and may be difficult to realize. "More recently, Lapedes [4] presented an
intuitive description of how multi-layer perceptrons with sigmoidal nonlinearities
could produce continuous nonlinear mappings. A careful mathematical proof was
recently developed by Cybenko [1] which demonstrated that continuous nonlinear
mappings can be formed using sigmoidal nonlinearities and a multi-layer perceptron
with one layer of hidden nodes. This proof, however, is not constructive and does
not indicate how many nodes are required in the hidden layer. The purpose of our
study was to determine whether multi-layer perceptrons with sigmoidal nonlinearities and trained using back-propagation could adaptively and rapidly form clipping
nonlinearities.
Initial experiments were performed to determine the difficulty of learning complex
mappings using multi-layer perceptrons trained using back-propagation. Networks
with 1 and 2 hidden layers and from 1 to 50 hidden nodes per layer were evaluated. Input and output nodes were linear and all other nodes included sigmoidal
nonlinearities. Best overall performance was provjded by the three-layer perceptron
shown in Fig. 1. It has 20 nodes in the first and 5 nodes in the second hidden layer.
This network could form a wide variety of mappings and required only slightly more
training than other networks. It was used in all experiments.
125
126
Lippmann and Beckman
y - OUTPUT
(LInear Sum)
(20 Nodes)
x - INPUT
Figure 1: The multi-layer perceptron with linear input and output nodes that was
used in all experiments.
The three-layer network shown in Fig. 1 was used to form clipping and other
deterministic nonlinearities. Results in Fig. 2 demonstrate that a clipping nonlinearity ('auld be formed with fewer than 1,000 input samples. Input/output point pairs
were determined by selecting the input at random over the range plotted and using
tlte deterministic clipping function shown as a solid line in Fig. 2. Back-propagation
training [7] was used with the gain term (11) equal to 0.1 and the momentum term
(0') equal to 0.5. These values provide good convergence rates for the clipping function and all other functions tested. Initial connection weights were set to small
random values.
The multi-layer percept ron from Fig. 1 was also used to form the four nonlinear
functions shown in Fig. 3. The "Hole Punch" is useful in nonlinear signal processing. It performs much the same function as the clipper but completely eliminates
amplitudes above a certain threshold le\'el. Accurate approximation of this function
required more than 50,000 input samples. The "Step" has one sharp edge and could
be roughly approximated after 2,000 input samples. The "Double Pulse" requires
approximation of two close "pulses" and is the nonlinear function analogy of the
disjoint region problem studied in [3]. In this examplf>, back-propagation training
approximated the rightmost pulse first after 5,000 input samples. Both pulses were
then approximated fairly well after 50,000 input samples. The "Gaussian Pulse"
is a smooth curve that could be approximated well after only 2,000 input samples.
These results demonstrate that back-propagation training with sigmoidal 1I0nlinearities can form many different nonlinear functions. Qualitative results on training
times are similar to those reported in [.1]. In this previous study it was de mOll-
Adaptive Neural Net Preprocessing for Signal Detection
....
.!:
1
1000 TRIALS
40 TRIALS
BEFORE TRAINING
DESIRED
ACTUAL
????????.
~ 0 ......
~
0- 1
-2_ 2
-1
2
0
-2
t
-1
0
INPUT (I()
2
-2
-1
0
2
RMS ERROR
O.
II:
0
II:
II:
u.I
(f)
:IE
ex:
200
400
606
TRIALS
800
1000
Figure 2: Clipping nonlinearities formed using back-propagation training and the
multi-layer perceptron from Fig. 1 (top) and the rms error produced by these Ilonlinearities versus training time (bottom).
strated that simple half-plane decision regions could be formed for classification
problems with little training while complex disjoint decision regions required long
training times. These new results suggest that complex nonlinearities with many
sharp discontinuities require much more training time than simple smooth curves.
THE SIGNAL DETECTION TASK
The signal detection task was to discriminate between two equally likely input signals as shown in Fig. 4. One signal (so(t)) corresponds to no input and the other
signal (Sl(t)) was a sinewa\'c pulse with fixed duration and known amplitude, frequency, and phase. Noise was added to these inputs, the resultant signal was passed
through a memoryless nonlinearity, and a matched filter was then used to select hypothesis Ho corresponding to no input or HI corresponding to the sinewave pulse.
The matched filter multiplied the output of the nonlinearity by the known timealigned signal waveform, integrated this product over time, and decided HI if the
result was greater than a threshold and Ho otherwise. The threshold was selected
to provide a minimum overall error rate. The optimum nonlinearity Ilsed in the detector depends on the noise distribu tion. If the signal levels are small relati\'e to the
noise levels, then the optimum nonlinearity is approximated by f (J') = t;~ In{ (In (J')).
where r .. (x) is the instantaneous probability density function of the noise (2]- This
function is linear for Gaussian noise but has a clipping shape for impulsi\'e noise.
127
128
Lippmann and Beckman
HOLE PUNCH
STEP
2r---~---r--~---.
I
-,
f-
.,
~
?2
I:::l
no
I::l
0
N.5.ooo
----Nco 50.000
?2
-,
Nco 2.000
---_.
0
.-
l?/
f/ I ..........
I
????~?:?~5
............... i.~j ..............
N.5oo
.
----- -
"-.
0
I
.,
1
2
?2
DOUBLE PULSE
1
-1--'
o
2
GAUSSIAN PULSE
2
?2
?1
o
2
INPUT (xl
Figure 3: Four deterministic nonlinearities formed using the multi-layer perceptron
from Fig. 1. Desired functions are plotted as solid lines while functions formed
using back-propagation with different numbers of input samples are plotted using
dots and dashes.
Examples of the signal, impulsive noise and Gaussian noise are presented in Fig. 5.
The signal had a fixed duration of 250 samples and peak amplitude of 1.0. The
impulsive noise was defined by its amplitude distribution and inter-arrival time.
Amplit udes had a zero mean, Laplacian distribution with a standard de\'iation (IJ)
of 14.1 in all experiments. The standard deviation was reduced to 2.8 in Fig. 5
for illustrative purposes. Inter-arrival times (L\T) between noise impulses had a
Poisson distribution. The mean inter-arrival time was varied in experiments to
obtain different SIN ratios after adding noise. For example varying inter-arrival
times from 500 to 2 samples results in SIN ratios that vary from roughly 1 dB to
- 24 dB. Additive Gaussian noise had zero mean and a standard oeviation (IJ) of
0.1 in all experiments.
ADAPTIVE TRAINING WITH NOISE
The three-layer perceptron was traineq as shown in Fig. 6 using the signal plus Iloist>
as the input and the uncorrupted signal as the desired output. Network weights
were adapted after every sample input using back-propagation training. Adaptive
nonlinearitics formed during training are shown in Fig. 7. These are similar to those
Adaptive Neural Net Preprocessing for Signal Detection
MEMORYlESS
NONLINEARITY
SO(II---
S,II)~
MATCHED
FILTER
DETECTOR
~-...o{
NOISE
'I ? I(x)
N(tl
Figure 4: The signal detection task was to discriminate between a sinewa\?e pulse
and a no-input condition with additive impulsive noise.
UNCORRUPTED SIGNAL
IMPULSE NOISE
ct = 2.8
.H~
o
50 100 150 200 250 0
12
50 100 150 200 2500
GAUSSIAN NOISE
maO
(J ? O. t
50 100 150 200 250
SAMPLES
Figure 5: The input to the nonlinearity with no noise, additive impulsive noise, and
additive Gaussian noise.
required by theory. No noise results in nonlinearity that is linear over the range
of the input sinewave (-1 to + 1) after fewer than 3,000 input samples. Impulsive
noise at a high SIN ratio (6.T
125 or SIN
-5 dB) results in a nonlinearity
that clips above the signal level after roughly 5,000 input samples and then slowly
forms a "Hole Punch" nonlinearity as the number of training samples increases.
Gaussian noise noise results in a nonlinearity that is roughly linear over the range
of the input sinewave after fewer than 5,000 input samples.
=
=
SIGNAL DETECTION PERFORMANCE
Signal detection performance was measured using a matched filter detector and the
nonlinearity shown in the center of Fig. 7 for 10,000 input training samples. The
error rate with a minimum-error matched filter is plotted in Fig. 8 for impulsive
lIoise at SIN ratios ranging from roughly 5 dB to -24 dB. This error rate was
estimated from 2,000 signal detection trials. Signal detection performance always
improved with the nonlinearity and sometimes the improvement was dramatic. The
error rate provided with the adaptively-formed nonlinearity is essentially identical
129
130
Lippmann and Beckman
DESIRED
OUTPUT
5(1)
X
MULTI-LAYER
) - -....--1 PERCEPTRON
+
Y
- I.
E
NOISE
BACK-PROPAGATION 1 - _.....
ALGORITHM
Figure 6: The procedure used for adaptive training.
NO NOISE
IMPULSE NOISE
GAUSSIAN NOISE
N. '00,000
N.5,000
2
~
...
::;)
...
Q.
::;)
0
0
N.1.000
-,
?2
-2
N . 2.ooo
N.3.ooo
-,
0
2 ?2
.,
0
2 ?2
.,
0
2
INPUT (x)
Figure 7: Nonlinearities formed with adaptive training with no additive noise, with
additive impulsive noise at a SIN level of -5 dB, and with additive Gaussian noise.
to that provided by a clipping nonlinearity that clips above the signal level. This
error rate is roughly zero down to - 24 dB and then rises rapidly with higher levels
of impulsive noise. This rapid increase in error rate below -24 dB is not shown in
Fig. 8. The error rate with linear processing rises slowly as the SIN ratio drops and
reaches roughly 36% when the SIN ratio is -24 dB.
Further exploratory experiments demonstrated that the nonlinearity formed by
back-propagation was not robust to the SIN ratio used during training. A clipping
nonlinearity is only formed when the number of samples of uncorrupted sinewave
input is high enough to form the linear function of the curve and the number of
samples of noise pulses is low , but sufficient to form the non~ill('ar clipping section
of the nonlinearity. At high noise levels the resulting nonlinearity is not linear Over
the range of the input signal. It instead resembles a curve that interpolates between
a flat horizontal input-output curve and the desired clipping curve.
SUMMARY AND DISCUSSION
In summary, it was first demonstrated that multi-layer perccptrons with linear
Adaptive Neural Net Preprocessing for Signal Detection
40
w
~
<
a:
a:
oa:
20
~
a:
w
LINEAR
PROCESSING
10
ADAPTIVE
NONLINEAR
PROCESSING
o
??? ...,?????? .I......I???.? ~.~____-~.
-10
-5
o
5
-20
-15
-Z5 '
SIN RATIO
Figure 8: The signal detection error rate with impulsive noise when the SIN ratio
after adding the noise ranges from 5 dB to - 24 dB.
input and output nodes could approximate prespecified clipping nonlinearities required for signal detection with impulsive noise with fewer than 1,000 trials of
back-propagation training. More complex nonlinearities could also be formed but
required longer training times. Clipping nonunearities were also formed adaptively
using a multi-layer perceptron with the corrupted signal as the input and the noisefree signal as the desired output. Nonlinearities learned using this approach at high
S / N ratios were similar to those required by theory and improved signal detection
performance dramatically at low SIN ratios. Further work is necessary to further
explore the utility of this technique for forming adaptive nonlinearities. This work
should explore the robustness of the nonlinearity formed to variations in the input
S / N ratio. It should also explore the use of multi-layer perccptrons and backpropagation training for other adaptive nonlinear signal processing tasks such as
system identification, noise removal, and channel modeling.
131
132
Lippmann and Beckman
References
[1] G. Cybenko. Approximation by superpositions of a sigmoidal function. Research note, Department of Computer Science, Tufts University, October 1988.
[2] J. E. Evans and A. S Griffiths. Design of a sanguine noise processor based
upon world-wide extremely low frequency (elf) recordings. IEEE Transactions
on Communications, COM-22:528-539, April 1974.
[3] W. M. Huang and R. P. Lippmann. Neural net and traditional classifiers. In
D. Anderson, editor, Neural Information Processing Systems, pages 387-396,
New York, 1988. American Institute of Physics.
[4] A. Lapedes and R. Farber. How neural nets work. In D. Anderson, editor, Neural Information Processing Systems, pages 442-456, New York, 1988. American
Institute of Physics.
[5] G. G. Lorentz. The 13th problem of Hilbert. In F. E. Browder, editor, Afathematical Developments Arising from Hilbert Problems. American Mathematical
Society, Providence, R.I., 1976.
[6] D. Palmer and D. DeSieno. Removing random noise from ekg signals using a
back propagation network, 1987. HNC Inc., San Diego, CA.
[7] D. E. Rumelhart, G. E. Hinton, and R. J. Williams. Learning internal representations by error propagation. In D. E. Rumelhart and J. L. McClelland, editors,
Parallel Distributed Processing, volume 1: Foundations, chapter 8. MIT Press,
Cambridge, MA, 1986.
[8] S. Tamura and A. Wailbel. Noise reduction using connectionist models. In Proceedings IEEE International Conference on Acoustics, Speech and Signal Processing, volume 1: Speech Processing, pages 553-556, April 1988.
[9] B. Widrow and S. D. Stearns. Adaptive Signal Processing. Prentice-Hall, NJ,
1985.
| 148 |@word trial:5 pulse:11 propagate:1 dramatic:1 solid:2 reduction:1 initial:2 selecting:1 lapedes:2 rightmost:1 current:1 com:1 lorentz:1 realize:1 evans:1 additive:8 shape:1 designed:1 sponsored:1 drop:1 half:1 fewer:5 selected:2 plane:1 prespecified:1 provides:1 node:18 ron:1 sigmoidal:7 mathematical:2 qualitative:1 inter:4 rapid:1 roughly:7 udes:1 multi:17 actual:1 little:1 provided:4 project:1 matched:7 developed:1 lexington:1 nj:1 every:1 demonstrates:1 classifier:1 appear:1 before:2 plus:1 studied:1 resembles:1 palmer:1 range:5 decided:1 backpropagation:1 procedure:1 griffith:1 suggest:1 close:1 prentice:1 equalization:1 cancelers:1 deterministic:5 demonstrated:3 center:1 williams:1 duration:2 exploratory:1 underwater:1 variation:1 diego:1 hypothesis:1 rumelhart:2 approximated:5 bottom:1 thousand:1 region:3 substantial:1 agency:1 trained:7 upon:1 completely:1 chapter:1 effective:1 widely:1 otherwise:1 echo:1 net:9 product:1 rapidly:3 lincoln:1 intuitive:1 description:1 convergence:1 double:2 optimum:2 produce:1 illustrate:1 oo:1 widrow:1 measured:1 ij:2 indicate:1 waveform:2 clipper:1 correct:1 farber:1 filter:8 require:1 government:1 cybenko:2 hall:1 mapping:4 lm:1 vary:1 purpose:3 beckman:5 superposition:1 sensitive:1 mit:2 gaussian:14 always:1 varying:1 combiner:1 improvement:3 greatly:1 el:1 nco:2 typically:1 integrated:1 hidden:10 overall:2 classification:1 ill:1 development:1 fairly:1 equal:2 construct:1 aware:1 noisefree:1 identical:1 elf:1 connectionist:1 richard:1 phase:1 detection:19 highly:1 accurate:1 edge:1 necessary:1 desired:8 plotted:4 modeling:1 impulsive:16 ar:1 clipping:17 deviation:1 front:1 reported:1 providence:1 corrupted:3 adaptively:3 density:1 peak:1 international:1 ie:1 physic:2 reflect:1 huang:1 slowly:2 american:3 nonlinearities:18 de:2 inc:1 caused:1 depends:1 performed:2 view:1 tion:1 parallel:1 desieno:1 formed:17 air:1 percept:1 identification:1 produced:1 processor:1 detector:4 reach:1 frequency:3 resultant:1 proof:3 gain:1 hilbert:2 amplitude:4 back:14 higher:1 improved:2 april:2 ooo:3 singleoutput:1 evaluated:1 anderson:2 horizontal:1 nonlinear:11 propagation:15 impulse:4 memoryless:2 laboratory:1 mile:1 sin:13 during:2 illustrative:1 demonstrate:3 performs:1 ranging:1 instantaneous:1 recently:2 common:1 volume:2 cambridge:1 nonlinearity:27 lightning:1 dot:1 had:4 longer:1 certain:1 uncorrupted:4 minimum:3 greater:1 determine:4 strike:1 signal:45 ale:3 ii:5 smooth:2 long:1 equally:1 laplacian:1 z5:1 multilayer:1 noiseless:1 essentially:1 poisson:1 sometimes:1 tamura:1 eliminates:1 recording:1 db:11 enough:1 variety:1 moll:1 whether:3 rms:2 defense:1 utility:1 passed:1 interpolates:1 york:2 speech:2 dramatically:1 useful:2 clip:2 mcclelland:1 reduced:2 sl:1 stearns:1 punch:3 estimated:1 disjoint:2 per:1 arising:1 four:2 threshold:3 sum:1 decision:2 layer:28 noise1:1 hi:2 dash:1 ct:1 adapted:1 flat:1 speed:1 extremely:2 relatively:1 department:2 slightly:1 end:1 multiplied:1 tuft:1 robustness:1 ho:2 top:1 relati:1 society:1 objective:2 added:1 occurs:1 traditional:1 link:1 oa:1 ratio:13 difficult:1 october:1 rise:2 design:1 policy:1 modem:1 ekg:1 hinton:1 communication:2 varied:1 sharp:2 pair:1 required:12 connection:1 acoustic:2 learned:1 discontinuity:1 below:1 difficulty:1 force:1 advanced:1 hnc:1 removal:1 singlelayer:1 filtering:5 proven:1 analogy:1 versus:1 foundation:1 sufficient:1 editor:4 summary:2 distribu:1 perceptron:12 institute:2 wide:2 distributed:1 curve:6 world:1 author:1 adaptive:19 preprocessing:5 san:1 transaction:1 approximate:1 lippmann:6 receiver:4 continuous:3 promising:1 channel:2 robust:1 ca:1 complex:5 noise:51 paul:1 arrival:4 fig:16 tl:1 position:1 momentum:1 mao:1 xl:1 theorem:1 down:1 removing:1 tlte:1 sinewave:4 explored:1 adding:2 illustrates:1 hole:3 likely:1 explore:3 forming:2 expressed:1 corresponds:1 ma:2 presentation:1 careful:1 included:1 telephone:1 determined:1 discriminate:2 perceptrons:6 select:1 internal:1 constructive:1 tested:1 ex:1 |
525 | 1,480 | Ensemble Learning
for Multi-Layer Networks
David Barber?
Christopher M. Bishopt
Neural Computing Research Group
Department of Applied Mathematics and Computer Science
Aston University, Birmingham B4 7ET, U.K.
http://www.ncrg.aston.ac.uk/
Abstract
Bayesian treatments of learning in neural networks are typically
based either on local Gaussian approximations to a mode of the
posterior weight distribution, or on Markov chain Monte Carlo
simulations. A third approach, called ensemble learning, was introduced by Hinton and van Camp (1993). It aims to approximate
the posterior distribution by minimizing the Kullback-Leibler divergence between the true posterior and a parametric approximating distribution. However, the derivation of a deterministic algorithm relied on the use of a Gaussian approximating distribution
with a diagonal covariance matrix and so was unable to capture
the posterior correlations between parameters. In this paper, we
show how the ensemble learning approach can be extended to fullcovariance Gaussian distributions while remaining computationally
tractable. We also extend the framework to deal with hyperparameters, leading to a simple re-estimation procedure. Initial results
from a standard benchmark problem are encouraging.
1
Introduction
Bayesian techniques have been successfully applied to neural networks in the context of both regression and classification problems (MacKay 1992; Neal 1996). In
contrast to the maximum likelihood approach which finds only a single estimate
for the regression parameters, the Bayesian approach yields a distribution of weight
parameters, p(wID), conditional on the training data D, and predictions are ex?Present address: SNN, University of Nijmegen, Geert Grooteplein 21, Nijmegen, The
Netherlands. http://wvw.mbfys.kun . n1/snn/ email: davidbbbfys.kun.n1
tpresent address: Microsoft Research Limited, St George House, Cambridge CB2 3NH,
UK. http://vvv.research.microsoft . com email: cmbishopbicrosoft.com
D. Barber and C. M. Bishop
396
pressed in terms of expectations with respect to the posterior distribution (Bishop
1995). However, the corresponding integrals over weight space are analytically intractable. One well-established procedure for approximating these integrals, known
as Laplace's method, is to approximate the posterior distribution by a Gaussian,
centred at a mode of p(wID), in which the covariance of the Gaussian is determined by the local curvature of the posterior distribution (MacKay 1995). The
required integrations can then be performed analytically. More recent approaches
involve Markov chain Monte Carlo simulations to generate samples from the posterior (Neal 1996}. However, such techniques can be computationally expensive, and
they also suffer from the lack of a suitable convergence criterion.
A third approach, called ensemble learning, was introduced by Hinton and van
Camp (1993) and again involves finding a simple, analytically tractable, approximation to the true posterior distribution. Unlike Laplace's method, however, the
approximating distribution is fitted globally, rather than locally, by minimizing a
Kullback-Leibler divergence. Hinton and van Camp (1993) showed that, in the case
of a Gaussian approximating distribution with a diagonal covariance, a deterministic learning algorithm could be derived. Although the approximating distribution
is no longer constrained to coincide with a mode of the posterior, the assumption
of a diagonal covariance prevents the model from capturing the (often very strong)
posterior correlations between the parameters. MacKay (1995) suggested a modification to the algorithm by including linear preprocessing of the inputs to achieve a
somewhat richer class of approximating distributions, although this was not implemented. In this paper we show that the ensemble learning approach can be extended
to allow a Gaussian approximating distribution with an general covariance matrix,
while still leading to a tractable algorithm.
1.1
The Network Model
We consider a two-layer feed-forward network having a single output whose value
is given by
H
(1)
/(x,w) = LViU(Ui'X)
i=1
where w is a k-dimensional vector representing all of the adaptive parameters in the
model, x is the input vector, {ud, i = 1, ... , H are the input-to-hidden weights,
and {Vi}, i = 1, ... ,H are the hidden-to-output weights. The extension to multiple outputs is straightforward. For reasons of analytic tractability, we choose the
sigmoidal hidden-unit activation function u(a) to be given by the error function
u(a} =
f! loa
exp (-8 2/2) d8
(2)
which (when appropriately scaled) is quantitatively very similar to the standard
logistic sigmoid. Hidden unit biases are accounted for by appending the input
vector with a node that is always unity. In the current implementation there are
no output biases (and the output data is shifted to give zero mean), although the
formalism is easily extended to include adaptive output biases (Barber and Bishop
1997) . . The data set consists of N pairs of input vectors and corresponding target
output values D = {x~, t~} ,It = 1, ... , N. We make the standard assumption
of Gaussian noise on the target values, with variance (3-1. The likelihood of the
training data is then proportional to exp(-(3ED ), where the training error ED is
ED{w) =
",
21~
(J{x~, w) ~
t~)
2
.
(3)
Ensemble Leamingfor Multi-Layer Networks
397
The prior distribution over weights is chosen to be a Gaussian of the form
p(w)
exp (-Ew(w))
(X
(4)
where Ew(w) = !wT Aw, and A is a matrix of hyper parameters. The treatment of
(3 and A is dealt with in Section 2.1. From Bayes' theorem, the posterior distribution
over weights can then be written
p(wID)
= z1 exp (-(3ED(w) -
(5)
Ew(w))
where Z is a normalizing constant. Network predictions on a novel example are
given by the posterior average of the network output
(f(x)) =
J
(6)
f(x, w)p(wID) dw.
This represents an integration over a high-dimensional space, weighted by a posterior distribution p(wID) which is exponentially small except in narrow regions
whose locations are unknown a-priori. The accurate evaluation of such integrals is
thus very difficult.
2 .Ensemble Learning
Integrals of the form (6) may be tackled by approximating p(wID) by a simpler
distribution Q(w). In this paper we choose this approximating distribution to be
a Gaussian with mean wand covariance C. We determine the values of w and C
by minimizing the Kullback-Leibler divergence between the network posterior and
approximating Gaussian, given by
F [Q]
{
J
J
=
Q(w) }
Q(w) In p(wID) dw
(7)
J
(8)
Q(w) In Q(w)dw -
Q(w) Inp(wID) dw.
The first term in (8) is the negative entropy of a Gaussian distribution, and is easily
evaluated to give ! In det (C) + const.
From (5) we see that the posterior dependent term in (8) contains two parts that
depend on the prior and likelihood
J
Q(w)Ew(w)dw
+
J
(9)
Q(w)ED(w)dw.
Note that the normalization coefficient Z-l in (5) gives rise to a constant additive
term in the KL divergence and so can be neglected. The prior term Ew (w) is
quadratic in w, and integrates to give Tr(CA) + ~wT Aw. This leaves the data
dependent term in (9) which we write as
=
L
J
(3N
Q(W)ED(W)dw =
"2 I: l(xl!, tl!)
(10)
I!=l
where
l(x, t) =
J
Q(w) (J(x, W))2 dw - 2t
J
Q(w)f(x, w) dw
+ t2?
(11)
D. Barber and C. M. BisJwp
398
For clarity, we concentrate only on the first term in (11), as the calculation of the
term linear in I(x, w) is similar, though simpler. Writing the Gaussian integral
over Q as an average, ( ), the first term of (11) becomes
H
((I(x,
w?2) = L
(vivju(uTx)u(uJx?).
(12)
i,j=I
To simplify the notation, we denote the set of input-to-hidden weights (Ul' ... , UH)
by u and the set of hidden-to-output weights, (VI' ... ' V H) by v. Similarly, we
partition the covariance matrix C into blocks, C uu , C vu , C vv , and C vu = C~v. As
the components of v do not enter the non-linear sigmoid functions, we can directly
integrate over v, so that each term in the summation (12) gives
((Oij
+ (u - IT)T \Il ij (u - IT) + n~ (u - IT?) u (uTxi) u (uTxj))
(13)
where
(C vv -
Oij
\Il ij
nij
-
CvuC uu -lCuV)ij
+ "hvj
(14)
C uu -IC u,v=:i C lI=:j,uC uu -1,
(15)
2C uu -ICu,lI=:jVi.
(16)
Although the remaining integration in (13) over u is not analytically tractable, we
can make use of the following result to reduce it to a one-dimensional integration
(u (z?a + ao) u (z ?b
+ bo?)z =
zaTb + bolal
))
T
2
Vl a l (1 + Ib 1 ) - (a b)2 z
(17)
where a and b are vectors and 0.0, bo are scalar offsets. The avera~e on the left of
(17) is over an isotropic multi-dimensional Gaussian, p(z) ex: exp( -z z/2), while the
average on the right is over the one-dimensional Gaussian p(z) ex: exp( -z2 /2). This
result follows from the fact that the vector z only occurs through the scalar product
with a and b, and so we can choose a coordinate system in which the first two
components of z lie in the plane spanned by a and b. All orthogonal components
do not appear elsewhere in the integrand, and therefore integrate to unity.
(u (zlal
+ 0.0) u
(
2
The integral we desire, (13) is only a little more complicated than (17) and can be
evaluated by first transforming the coordinate system to an isotopic basis z, and
then differentiating with respect to elements of the covariance matrix to 'pull down'
the required linear and quadratic terms in the u-independent pre-factor of (13).
These derivatives can then be reduced to a form which requires only the numerical
evaluation of (17). We have therefore succeeded in reducing the calculation of the
KL divergence to analytic terms together with a single one-dimensional numerical
integration of the form (17), which we compute using Gaussian quadrature 1 .
Similar techniques can be used to evaluate the derivatives of the KL divergence with
respect to the mean and covariance matrix (Barber and Bishop 1997). Together with
the KL divergence, these derivatives are then used in a scaled conjugate gradient
optimizer to find the parameters w and C that represent the best Gaussian fit.
The number of parameters in the covariance matrix scales quadratically with the
number of weight parameters. We therefore have also implemented a version with
1 Although (17) appears to depend on 4 parameters, it can be expressed in terms of 3
independent parameters. An alternative to performing quadrature during training would
therefore be to compute a 3-dimensionallook-up table in advance.
Ensemble Learning for Multi-Layer Networks
Posterior
laplace fit
399
Minimum KLD fit
Minimum KL fit
Figure 1: Laplace and minimum Kullback-Leibler Gaussian fits to the posterior.
The Laplace method underestimates the local posterior mass by basing the covariance matrix on the mode alone, and has KL value 41. The minimum KullbackLeibler Gaussian fit with a diagonal covariance matrix (KLD) gives a KL value
of 4.6, while the minimum Kullback-Leibler Gaussian with full covariance matrix
achieves a value of 3.9.
a constrained covariance matrix
s
C = diag(di,? ? ?, d~) +
L sisT
(18)
i=l
which is the form of covariance used in factor analysis (Bishop 1997). This reduces
the number offree parameters in the covariance matrix from k(k + 1)/2 to k(s + 1)
(representing k(s + 1) - s(s - 1)/2 independent degrees of freedom) which is now
linear in k. Thus, the number of parameters can be controlled by changing sand,
unlike a diagonal covariance matrix, this model can still capture the strongest of the
posterior correlations. The value of s should be as large as possible, subject only
to computational cost limitations. There is no 'over-fitting' as s is increased since
more flexible distributions Q(w) simply better approximate the true posterior.
We illustrate the optimization of the KL divergence using a toy problem involving
the posterior distribution for a two-parameter regression problem. Figure 1 shows
the true posterior together with approximations obtained from Laplace's method,
ensemble learning with a diagonal covariance Gaussian, and ensemble learning using
an unconstrained Gaussian.
2.1
Hyperparameter Adaptation
So far, we have treated the hyperparameters as fixed. We now extend the ensemble
learning formalism to include hyperparameters within the Bayesian framework. For
simplicity, we consider a standard isotropic prior covariance matrix of the form
A = aI, and introduce hyperpriors given by Gamma distributions
lnp (a)
lnp (f3)
=
In
{aa-l exp ( -~) } + const
(19)
In
{f3 C- 1 exp ( -~) } + const
(20)
D. Barber and C. M. BisJwp
400
where a, b, c, d are constants. The joint posterior distribution of the weights and
hyperparameters is given by
p (w, a, ,BID)
<X
P (Dlw, j3) p (wla) p (a) p (,B)
(21)
in which
lnp (Dlw,,B)
- ,BED
N
+ "2 In,B + const
k
-alwl 2 + '2 In a
lnp (wla)
(22)
+ const
(23)
We follow MacKay (1995) by modelling the joint posterior p (w, a, ,BID) by a factorized approximating distribution of the form
(24)
Q(w)R(a)S(,B)
where Q(w) is a Gaussian distribution as before, and the functional forms of Rand
S are left unspecified. We then minimize the KL divergence
F[Q,R,S]
=
J
Q(w)R(a)S(,B) In
{
Q(W)R(a)S(,B) }
p(w,a,,BID)
dwdad,B.
(25)
Consider first the dependence of (25) on Q(w)
F [QJ
J
- -J
Q(w)R(a)S(j3) { -,BED(W) -
-
Q(w) { -73ED(W) -
~lwl2 -In Q(w) } + const
~lwl2 -In Q(W)} + const
(26)
(27)
where a = J R(a)ada and 73 = J S(,B)j3d,B. We see that (27) has the form of (8),
except that the fixed hyperparameters are now replaced with their average values.
To calculate these averages, consider the dependence of the functional F on R(a)
F[R]
JQ(W)R(a)S(j3){-~lwI2+~lna+(a-1)lna-i}
-J +
+
-
R(a) { ;
(r - 1) Ina -In R(a)} da
const
dwdad,B
(28)
where r = ~ +a and lis = ~lwl2 + ~TrC+ lib. We recognise (28) as the KullbackLeibler divergence between R(a) and a Gamma distribution. Thus the optimum
R(a) is also Gamma distributed
R(a)
We therefore obtain
<X
a r - 1 exp (-;) .
(29)
a = rs.
A similar procedure for S(,B) gives 73 = uv, where u = ~ + c and 11v = (ED) + lid,
in which (ED) has already been calculated during the optimization of Q(w) .
This defines an iterative procedure in which we start by initializing the hyperparameters (using the mean of the hyperprior distributions) and then alternately optimize
the KL divergence over Q(w) and re-estimate a and 73.
3
Results and Discussion
As a preliminary test of our method on a standard benchmark problem, we applied the minimum KL procedure to the Boston Housing dataset. This is a one
Ensemble Learning for Multi-Layer Networks
I Method
Ensemble (s == 1)
Ensemble (diagonal)
Laplace
401
Test Error
0.22
0.28
0.33
Table 1: Comparison of ensemble learning with Laplace's method. The test error
is defined to be the mean squared error over the test set of 378 examples.
dimensional regression problem, with 13 inputs, in which the data for 128 training
examples was obtained from the DELVE archive 2 ? We trained a network of four
hidden units, with covariance matrix given by (18) with s = 1, and specified broad
hyperpriors on a and (3 (a = 0.25, b = 400, c = 0.05, and d = 2000). Predictions are
made by evaluating the integral in (6). This integration can be done analytically
as a consequence of the form of the sigmoid function given in (2).
We compared the performance of the KL method against the Laplace framework of
MacKay (1995) which also treats hyperparameters through a re-estimation procedure. In addition we also evaluated the performance of the ensemble method using
a diagonal covariance matrix. Our results are summarized in Table 1.
Acknowledgements
We would like to thank Chris Williams for helpful discussions. Supported by EPSRC
grant GR/J75425: Novel Developments in Learning Theory for Neural Networks.
References
Barber, D. and C. M. Bishop (1997). On computing the KL divergence for
Bayesian neural networks. Technical report, Neural Computing Research
Group, Aston University, Birmingham, {;.K.
Bishop, C. M. (1995). Neural Networks for Pattern Recognition. Oxford University Press.
Bishop, C. M. (1997). Latent variables, mixture distributions and topographic
mappings. Technical report, Aston University. To appear in Proceedings of
the NATO Advanced Study Institute on Learning in Graphical Models, Erice.
Hinton, G. E. and D. van Camp (1993). Keeping neural networks simple by
minimizing the description length of the weights. In Proceedings of the Sixth
Annual Conference on Computational Learning Theory, pp. 5-13.
MacKay, D. J. C. (1992). A practical Bayesian framework for back-propagation
networks. Neural Computation 4 (3) , 448-472.
MacKay, D. J. C. (1995). Developments in probabilistic modelling with neural
networks--ensemble learning. In Neural Networks: Artificial Intelligence and
Industrial Applications. Proceedings of the 3rd Annual Symposium on Neural
Networks, Nijmegen, Netherlands, 14-15 September 1995, Berlin, pp. 191-198.
Springer.
MacKay, D. J. C. (1995). Probable networks and plausible predictions - a review of practical Bayesian methods for supervised neural networks. Network:
Computation in Neural Systems 6(3), 469-505.
Neal, R. M. (1996). Bayesian Learning for Neural Networks. Springer. Lecture
Notes in Statistics 118.
2See http://wvv . cs. utoronto. cal "-'delve I
| 1480 |@word version:1 wla:2 grooteplein:1 simulation:2 r:1 covariance:21 pressed:1 tr:1 initial:1 contains:1 current:1 com:2 z2:1 activation:1 trc:1 written:1 numerical:2 additive:1 partition:1 analytic:2 alone:1 intelligence:1 leaf:1 plane:1 isotropic:2 tpresent:1 node:1 location:1 sigmoidal:1 simpler:2 symposium:1 consists:1 fitting:1 introduce:1 mbfys:1 multi:5 globally:1 encouraging:1 snn:2 little:1 lib:1 becomes:1 notation:1 mass:1 factorized:1 sist:1 unspecified:1 finding:1 scaled:2 uk:2 unit:3 grant:1 appear:2 before:1 local:3 treat:1 consequence:1 oxford:1 delve:2 limited:1 practical:2 vu:2 block:1 cb2:1 procedure:6 pre:1 inp:1 cal:1 context:1 writing:1 kld:2 www:1 optimize:1 deterministic:2 straightforward:1 williams:1 simplicity:1 wvv:1 spanned:1 pull:1 dw:9 geert:1 coordinate:2 laplace:9 target:2 element:1 expensive:1 recognition:1 epsrc:1 initializing:1 capture:2 calculate:1 region:1 transforming:1 ui:1 neglected:1 trained:1 depend:2 basis:1 uh:1 easily:2 joint:2 derivation:1 monte:2 artificial:1 hyper:1 whose:2 richer:1 plausible:1 statistic:1 topographic:1 housing:1 product:1 adaptation:1 achieve:1 bed:2 description:1 wvw:1 convergence:1 optimum:1 illustrate:1 ac:1 ij:3 strong:1 lna:2 implemented:2 c:1 involves:1 uu:5 concentrate:1 j75425:1 wid:8 sand:1 ao:1 preliminary:1 probable:1 summation:1 extension:1 ic:1 exp:9 mapping:1 optimizer:1 achieves:1 estimation:2 birmingham:2 integrates:1 basing:1 successfully:1 weighted:1 gaussian:23 always:1 aim:1 rather:1 derived:1 modelling:2 likelihood:3 contrast:1 industrial:1 camp:4 helpful:1 dependent:2 vl:1 typically:1 hidden:7 jq:1 classification:1 flexible:1 priori:1 development:2 constrained:2 integration:6 mackay:8 uc:1 f3:2 having:1 represents:1 broad:1 t2:1 report:2 quantitatively:1 simplify:1 gamma:3 divergence:12 hvj:1 replaced:1 n1:2 microsoft:2 freedom:1 evaluation:2 mixture:1 chain:2 accurate:1 integral:7 succeeded:1 orthogonal:1 hyperprior:1 re:3 nij:1 fitted:1 increased:1 formalism:2 ada:1 tractability:1 cost:1 gr:1 kullbackleibler:2 aw:2 st:1 probabilistic:1 together:3 again:1 squared:1 choose:3 d8:1 derivative:3 leading:2 li:3 toy:1 centred:1 summarized:1 coefficient:1 vi:2 performed:1 start:1 bayes:1 relied:1 complicated:1 utx:1 minimize:1 il:2 variance:1 ensemble:17 yield:1 dealt:1 bayesian:8 carlo:2 strongest:1 ed:9 email:2 sixth:1 against:1 underestimate:1 pp:2 di:1 dataset:1 treatment:2 dlw:2 back:1 appears:1 feed:1 jvi:1 supervised:1 follow:1 rand:1 evaluated:3 though:1 done:1 correlation:3 christopher:1 lack:1 propagation:1 defines:1 mode:4 logistic:1 true:4 analytically:5 leibler:5 neal:3 avera:1 deal:1 offree:1 during:2 criterion:1 isotopic:1 novel:2 sigmoid:3 functional:2 b4:1 exponentially:1 nh:1 ncrg:1 extend:2 cambridge:1 enter:1 ai:1 rd:1 unconstrained:1 uv:1 mathematics:1 similarly:1 alwl:1 longer:1 curvature:1 posterior:25 recent:1 showed:1 lnp:4 minimum:6 george:1 somewhat:1 determine:1 ud:1 multiple:1 full:1 reduces:1 technical:2 calculation:2 controlled:1 prediction:4 involving:1 regression:4 j3:3 expectation:1 represent:1 normalization:1 addition:1 appropriately:1 unlike:2 archive:1 subject:1 vvv:1 bid:3 fit:6 reduce:1 det:1 qj:1 ul:1 suffer:1 involve:1 netherlands:2 locally:1 reduced:1 http:4 generate:1 shifted:1 write:1 hyperparameter:1 group:2 four:1 clarity:1 changing:1 wand:1 recognise:1 capturing:1 layer:5 tackled:1 quadratic:2 annual:2 integrand:1 performing:1 department:1 conjugate:1 unity:2 lid:1 modification:1 computationally:2 tractable:4 hyperpriors:2 appending:1 alternative:1 remaining:2 include:2 graphical:1 const:8 approximating:12 already:1 occurs:1 parametric:1 dependence:2 diagonal:8 september:1 gradient:1 unable:1 thank:1 berlin:1 chris:1 barber:7 reason:1 length:1 minimizing:4 kun:2 difficult:1 nijmegen:3 negative:1 rise:1 implementation:1 unknown:1 markov:2 benchmark:2 hinton:4 extended:3 david:1 introduced:2 pair:1 required:2 kl:13 specified:1 z1:1 quadratically:1 narrow:1 established:1 alternately:1 address:2 suggested:1 pattern:1 including:1 suitable:1 treated:1 oij:2 advanced:1 representing:2 aston:4 prior:4 review:1 acknowledgement:1 ina:1 lecture:1 limitation:1 proportional:1 integrate:2 degree:1 elsewhere:1 accounted:1 loa:1 supported:1 keeping:1 bias:3 allow:1 vv:2 institute:1 differentiating:1 van:4 distributed:1 calculated:1 evaluating:1 forward:1 made:1 adaptive:2 coincide:1 preprocessing:1 far:1 approximate:3 nato:1 kullback:5 iterative:1 latent:1 table:3 ca:1 icu:1 diag:1 da:1 noise:1 hyperparameters:7 quadrature:2 tl:1 xl:1 lie:1 house:1 ib:1 third:2 theorem:1 down:1 bishop:8 utoronto:1 offset:1 normalizing:1 intractable:1 boston:1 entropy:1 simply:1 prevents:1 desire:1 expressed:1 bo:2 scalar:2 springer:2 aa:1 conditional:1 determined:1 except:2 reducing:1 wt:2 called:2 ew:5 evaluate:1 ex:3 |
526 | 1,481 | The Error Coding and Substitution PaCTs
GARETH JAMES
and
TREVOR HASTIE
Department of Statistics, Stanford University
Abstract
A new class of plug in classification techniques have recently been developed in the statistics and machine learning literature. A plug in classification technique (PaCT) is a method that takes a standard classifier
(such as LDA or TREES) and plugs it into an algorithm to produce a
new classifier. The standard classifier is known as the Plug in Classifier (PiC). These methods often produce large improvements over using
a single classifier. In this paper we investigate one of these methods and
give some motivation for its success.
1 Introduction
Dietterich and Bakiri (1995) suggested the following method, motivated by Error Correcting Coding Theory, for solving k class classification problems using binary classifiers.
? Produce a k by B (B large) binary coding matrix, ie a matrix of zeros and ones.
We will denote this matrix by Z, its i, jth component by Zij, its ith row by Zi and
its j th column by zj.
? Use the first column of the coding matrix (Zl) to create two super groups by
assigning all groups with a one in the corresponding element of Zl to super group
one and all other groups to super group zero.
? Train your plug in classifier (PiC) on the new two class problem.
? Repeat the process for each of the B columns (Zl, Z2, ... ,ZB) to produce B
trained classifiers.
? For a new test point apply each of the B classifiers to it. Each classifier will
produce a 'Pi which is the estimated probability the test point comes from the
jth super group one. This will produce a vector of probability estimates, f> =
(PI , ih., . .. ,PB) T .
The Error Coding and Substitution PaCTs
543
? To classify the point calculate Li = 2::=1 IPi - Zij I for each of the k groups (ie
for i from 1 to k). This is the LI distance between p and Zi (the ith row of Z).
Classify to the group with lowest L 1 distance or equivalently argi min Li
We call this the ECOC PaCT. Each row in the coding matrix corresponds to a unique (nonminimal) coding for the appropriate class. Dietterich's motivation was that this allowed
errors in individual classifiers to be corrected so if a small number of classifiers gave a
bad fit they did not unduly influence the final classification. Several PiC's have been tested.
The best results were obtained by using tree's, so all the experiments in this paper are stated
using a standard CART PiC. Note however, that the theorems are general to any Pic.
In the past it has been assumed that the improvements shown by this method were attributable to the error coding structure and much effort has been devoted to choosing an
optimal coding matrix. In this paper we develop results which suggest that a randomized
coding matrix should match (or exceed) the performance of a designed matrix.
2 The Coding Matrix
Empirical results (see Dietterich and Bakiri (1995? suggest that the ECOC PaCT can produce large improvements over a standard k class tree classifier. However, they do not shed
any light on why this should be the case. To answer this question we need to explore its
probability structure. The coding matrix, Z, is central to the PaCT. In the past the usual
approach has been to choose one with as large a separation between rows (Zi) as possible
(in terms of hamming distance) on the basis that this allows the largest number of errors
to be corrected. In the next two sections we will examine the tradeoffs between a designed
(deterministic) and a completely randomized matrix.
Some of the results that follow will make use of the following assumption.
k
E[fji
I Z,X] = LZiiqi = ZjT q
= 1, ... ,B
j
(1)
i=l
where qi = P (Gil X) is the posterior probability that the test observation is from group
i given that our predictor variable is X. This is an unbiasedness assumption. It states that
on average our classifier will estimate the probability of being in super group one correctly.
The assumption is probably not too bad given that trees are considered to have low bias.
2.1
Deterministic Coding Matrix
Let Di = 1 - 2Ld B for i = 1 ... k. Notice that ar& min Li = arg i max Di so using Di
to classify is identical to the ECOC PaCT. Theorem 3 in section 2.2 explains why this is an
intuitive transformation to use.
Obviously no PaCT can outperform the Bayes Classifier. However we would hope that it
would achieve the Bayes Error Rate when we use the Bayes Classifier as our PiC for each
2 class problem. We have defined this property as Bayes Optimality. Bayes Optimality is
essentially a consistency result It states, if our PiC converges to the Bayes Classifier, as
the training sample size increases, then so will the PaCT.
Definition 1 A PaCT is said to be Bayes Optimal if, for any test set, it always classifies to
the bayes group when the Bayes Classifier is our PiC.
For the ECOC PaCT this means that argi max qi = argi max D i , for all points in the
predictor space, when we use the Bayes Classifier as our Pic. However it can be shown
that in this case
i
= 1, ... ,k
544
G. James and T. Hastie
It is not clear from this expression why there should be any guarantee that argi max Vi =
argi max qi . In fact the following theorem tells us that only in very restricted circumstances
will the ECOC PaCT be Bayes Optimal.
Theorem 1 The Error Coding method is Bayes Optimal iff the Hamming distance between
every pair of rows of the coding matrix is equal.
The hamming distance between two binary vectors is the number of points where they
differ. For general B and k there is no known way to generate a matrix with this property
so the ECOC PaCT will not be Bayes Optimal.
2.2
Random Coding Matrix
We have seen in the previous section that there are potential problems with using a deterministic matrix. Now suppose we randomly generate a coding matrix by choosing a zero
or one with equal probability for every coordinate. Let Pi = E(I- 21ih - Zilll T) where
T is the training set. Then Pi is the conditional expectation of Di and we can prove the
following theorem.
Theorem 2 For a random coding matrix, conditional on T, argi max Vi --+ argi max Pi
a.s. as B --+ 00. Or in other words the classification from the ECOC PaCT approaches the
classification from just using argi max Pi a.s.
This leads to corollary 1 which indicates we have eliminated the main concern of a deterministic matrix.
Corollary 1 When the coding matrix is randomly chosen the ECOC PaCT is asymptotically Bayes Optimal ie argi max Di --+ argi max qi a.s. as B --+ 00
This theorem is a consequence of the strong law. Theorems 2 and 3 provide motivation for
the ECOC procedure.
Theorem 3 Under assumption 1 for a randomly generated coding matrix
E j) i = E Pi = qi
i = 1 ... k
This tells us that Vi is an unbiased estimate of the conditional probability so classifying to
the maximum is in a sense an unbiased estimate of the Bayes classification.
Now theorem 2 tells us that for large B the ECOC PaCT will be similar to classifying using
argi max ILi only. However what we mean by large depends on the rate of convergence.
Theorem 4 tells us that this rate is in fact exponential.
Theorem 4 lfwe randomly choose Z then, conditional on T , for any fixed X
Pr(argi max Vi
i
argi max ILi) ::; (k - 1) . e- mB
for some constant m.
Note that theorem 4 does not depend on assumption 1. This tells us that the error rate for
the ECOC PaCT is equal to the error rate using argi max Pi plus a tenn which decreases
exponentially in the limit. This result can be proved using Hoeffding's inequality (Hoeffding (1963?.
Of course this only gives an upper bound on the error rate and does not necessarily indicate
the behavior for smaller values of B. Under certain conditions a Taylor expansion indicates
that Pr(argi maxDi i argi maxPi) :::::: 0.5 - mVE for small values of mVE. So we
The Error Coding and Substitution PaCTs
545
o
'"o
1/sqrt(B) convergence
lIB convergence
o
CO>
o
50
100
150
200
B
Figure 1: Best fit curves for rates 1/VB and 1/ B
might expect that for smaller values of B the error rate decreases as some power of B but
that as B increases the change looks more and more exponential.
To test this hypothesis we calculated the error rates for 6 different values of B
(15,26,40,70,100,200) on the LEITER data set (available from the Irvine Repository
of machine learning). Each value of B contains 5 points corresponding to 5 random matrices. Each point is the average over 20 random training sets. Figure 1 illustrates the results.
Here we have two curves. The lower curve is the best fit of 1/VB to the first four groups.
It fits those groups well but under predicts errors for the last two groups. The upper curve
is the best fit of 1/ B to the last four groups. It fits those groups well but over predicts errors
for the first two groups. This supports our hypothesis that the error rate is moving through
the powers of B towards an exponential fit.
We can see from the figure that even for relatively low values of B the reduction in error
rate has slowed substantially. This indicates that almost all the remaining errors are a result
of the error rate of argi max J-li which we can not reduce by changing the coding matrix.
The coding matrix can be viewed as a method for sampling from the distribution of
1- 21pj - Zij I. If we sample randomly we will estimate J-li (its mean). It is well known that
the optimal way to estimate such a parameter is by random sampling so it is not possible to
improve on this by designing the coding matrix. Of course it may be possible to improve
on argi max J-li by using the training data to influence the sampling procedure and hence
estimating a different quantity. However a designed coding matrix does not use the training
data. It should not be possible to improve on random sampling by using such a procedure
(as has been attempted in the past).
3
Why does the ECOC PaCT work?
The easiest way to motivate why the ECOC PaCT works, in the case of tree classifiers, is
to consider a very similar method which we call the Substitution PaCT. We will show that
under certain conditions the ECOC PaC!' is very similar to the Substitution PaCT and then
motivate the success of the later.
546
3.1
G. James and T. Hastie
Substitution PaCT
The Substitution PaCT uses a coding matrix to fonn many different trees just as the ECOC
PaCT does. However, instead of using the transformed training data to fonn a probability
estimate for each two class problem, we now plug the original (ie k-class) training data back
into the new tree. We use this training data to fonn probability estimates and classifications
just as we would with a regular tree. The only difference is in how the tree is fonned.
Therefore, unlike the ECOC PaCT, each tree will produce a probability estimate for each
of the k classes. For each class we simply average the probability estimate for that class
is the probability estimate for the Substitution PaCT, then
over our B trees. So if
pf
1
pf =
B
(2)
B LPij
j=l
where Pij is the probability estimate for the ith group for the tree fonned from the jth
column of the coding matrix.
Theorem 5 shows that under certain conditions the ECOC PaCT can be thought of as an
approximation to the Substitution PaCT.
Theorem 5 Suppose that Pij is independentfrom the jth column of the coding matrix, for
all i and j. Then as B approaches infinity the ECOC PaCT and Substitution PaCT will
converge ie they will give identical classification rules.
The theorem depends on an unrealistic assumption. However, empirically it is well known
that trees are unstable and a small change in the data set can cause a large change in the
structure of the tree so it may be reasonable to suppose that there is a low correlation.
To test this empirically we ran the ECOC and Substitution PaCT's on a simulated data
set. The data set was composed of 26 classes. Each class was distributed as a bivariate
normal with identity covariance matrix and uniformly distributed means. The training data
consisted of 10 observations from each group. Figure 2 shows a plot of the estimated
probabilities for each of the 26 classes and 1040 test data points averaged over 10 training
data sets. Only points where the true posterior probability is greater than 0.01 have been
plotted since groups with insignificant probabilities are unlikely to affect the classification.
If the two groups were producing identical estimates we would expect the data points to
lie on the dotted 45 degree.line. Clearly this is not the case. The Substitution PaCT is
systematically shrinking the probability estimates. However there is a very clear linear
relationship (R2 :::::: 95%) and since we are only interested in the arg max for each test
point we might expect similar classifications. In fact this is the. case with fewer than 4% of
points correctly classified by one group but not the other.
3.2
Why does the Substitution PaCT work?
pf
The fact that
is an average of probability estimates suggests that a reduction in variability may be an explanation for the success of the Substitution PaCT. Unfortunately it
has been well shown (see for example Friedman (1996? that a reduction in variance of
the probability estimates does not necessarily correspond to a reduction in the error rate.
However theorem 6 provides simplifying assumptions under which a relationship between
the two quantities exists.
Theorem 6 Suppose that
pT
and
pf
(a[ > 0)
(af > 0)
(3)
(4)
The Error Coding and Substitution PaCTs
547
~
II)
d
co
~
i
~Q.
co
d
C
.g
:0
]
....
d
:0
CI)
'"
0
0
d
0.0
0.4
0.2
0.6
08
1.0
Eeoc probabilitJea
Figure 2: Probability estimates from both the ECOC and Substitution PaCT's
where eS and eT have identical joint distributions with variance 1. pT is the probability
estimate of the ith group for a k class tree method, ao and al are constants and qi is the
true posterior probability. Let
'V _
,-
T
Var(p'!'ja
~
1 )
S
V ar(pil / a 1 )
and p = corr(pil ,Pi2) (assumed constantfor all i). Then
Pr(argmaxP7 = argmaxqi) ~ Pr(argmaxp; = argmaxqi)
(5)
if
(6)
and
I-p
- 'Y - P
B>--
(7)
The theorem states that under fairly general conditions, the probability that the Substitution
PaCT gives the same classification as the Bayes classifier is at least as great as that for the
tree method provided that the standardized variability is low enough. It should be noted
that only in the case of two groups is there a direct correspondence between the error rate
and 5. The inequality in 5 is strict for most common distributions (e.g. normal, uniform,
exponential and gamma) of e.
Now there is reason to believe that in general p will be small. This is a result of the
empirical variability of tree classifiers. A small change in the training set can cause a large
change in the structure of the tree and also the final probability estimates. So by changing
the super group coding we might expect a probability estimate that is fairly unrelated to
previous estimates and hence a low correlation.
To test the accuracy of this theory we examined the results from the simulation performed in
section 3.1. We wished to estimate 'Y and p. The following table summarizes our estimates
for the variance and standardizing (al) terms from the simulated data set.
I Classifier
Substitution PaCT
Tree Method
548
G. James and T. Hastie
Tree
'"o
ECOC
s..t>stilUtion
?o
------------------------5
10
50
100
B (log scale)
Figure 3: Error rates on the simulated data set for tree method, Substitution PaCf and
ECOC PaCT plotted against B (on log scale)
=
These quantities give us an estimate for, of l' 0.227 We also derived an estimate for p
of p = 0.125
We see that p is less than, so provided B ~ ~=~ ~ 8.6 we should see an improvement
in the Substitution PaCT over a k class tree classifier. Figure 3 shows that the Substitution
error rate drops below that of the tree classifier at almost exactly this point.
4
Conclusion
The ECOC PaCT was originally envisioned as an adaption of error coding ideas to classification problems. Our results indicate that the error coding matrix is simply a method
for randomly sampling from a fixed distribution. This idea is very similar to the Bootstrap where we randomly sample from the empirical distribution for a fixed data set. There
you are trying to estimate the variability of some parameter. Your estimate will have two
sources of error, randomness caused by sampling from the empirical distribution and the
randomness from the data set itself. In our case we have the same two sources of error,
error caused by sampling from 1 - 2!ftj - Zij! to estimate J-ti and error's caused by J-t itself.
In both cases the first sort of error will reduce rapidly and it is the second type we are really
interested in. It is possible to motivate the reduction in error rate of using argi max J-ti in
terms of a decrease in variability, provided B is large enough and our correlation (p) is
small enough.
References
Dietterich, T.G. and Bakiri G. (1995) Solving Multiclass Learning Problems via ErrorCorrecting Output Codes, Journal of Artificial Intelligence Research 2 (1995) 263-286
Diet~rich, T. G. and Kong, E. B. (1995) Error-Correcting Output Coding Corrects Bias
and Variance, Proceedings of the 12th International Conference on Machine Learning pp.
313-321 Morgan Kaufmann
Friedman, 1.H. (1996) On Bias, Variance, Oil-loss, and the Curse of Dimensionality, Dept
of Statistics, Stanford University, Technical Report
Hoeffding, W. (1963) Probability Inequalities for Sums of Bounded Random Variables.
"Journal of the American Statistical Association", March, 1963
| 1481 |@word kong:1 repository:1 simulation:1 covariance:1 simplifying:1 fonn:3 ld:1 reduction:5 substitution:21 contains:1 zij:4 past:3 z2:1 assigning:1 designed:3 plot:1 drop:1 tenn:1 fewer:1 intelligence:1 ith:4 provides:1 ipi:1 direct:1 prove:1 behavior:1 examine:1 ecoc:23 curse:1 pf:4 lib:1 provided:3 classifies:1 estimating:1 unrelated:1 bounded:1 lowest:1 what:1 easiest:1 substantially:1 developed:1 transformation:1 guarantee:1 every:2 ti:2 shed:1 exactly:1 classifier:25 zl:3 producing:1 limit:1 consequence:1 might:3 plus:1 examined:1 suggests:1 co:3 averaged:1 unique:1 bootstrap:1 procedure:3 empirical:4 thought:1 word:1 regular:1 suggest:2 influence:2 deterministic:4 correcting:2 rule:1 coordinate:1 pt:2 suppose:4 us:1 designing:1 hypothesis:2 element:1 predicts:2 calculate:1 decrease:3 ran:1 envisioned:1 trained:1 depend:1 solving:2 motivate:3 basis:1 completely:1 joint:1 train:1 artificial:1 tell:5 choosing:2 stanford:2 statistic:3 itself:2 final:2 obviously:1 lpij:1 eeoc:1 mb:1 rapidly:1 iff:1 achieve:1 intuitive:1 convergence:3 produce:8 argmaxp:1 converges:1 pi2:1 develop:1 wished:1 strong:1 come:1 indicate:2 differ:1 explains:1 ja:1 ao:1 really:1 considered:1 normal:2 great:1 largest:1 create:1 hope:1 clearly:1 always:1 super:6 corollary:2 derived:1 improvement:4 indicates:3 sense:1 unlikely:1 transformed:1 interested:2 arg:2 classification:13 fairly:2 equal:3 eliminated:1 sampling:7 identical:4 look:1 report:1 randomly:7 composed:1 gamma:1 individual:1 friedman:2 investigate:1 light:1 devoted:1 tree:23 taylor:1 plotted:2 column:5 classify:3 ar:2 predictor:2 uniform:1 argi:19 too:1 answer:1 unbiasedness:1 international:1 randomized:2 ie:5 corrects:1 central:1 choose:2 hoeffding:3 american:1 li:7 potential:1 standardizing:1 coding:33 caused:3 vi:4 depends:2 later:1 performed:1 bayes:16 sort:1 pil:2 accuracy:1 variance:5 kaufmann:1 correspond:1 randomness:2 classified:1 sqrt:1 trevor:1 definition:1 against:1 pp:1 james:4 di:5 hamming:3 irvine:1 proved:1 dimensionality:1 back:1 originally:1 follow:1 diet:1 just:3 correlation:3 lda:1 believe:1 oil:1 dietterich:4 consisted:1 unbiased:2 true:2 hence:2 noted:1 trying:1 recently:1 common:1 empirically:2 exponentially:1 association:1 consistency:1 moving:1 posterior:3 certain:3 inequality:3 binary:3 success:3 seen:1 morgan:1 greater:1 converge:1 ii:1 technical:1 match:1 plug:6 af:1 qi:6 essentially:1 circumstance:1 expectation:1 source:2 unlike:1 probably:1 strict:1 cart:1 call:2 exceed:1 enough:3 affect:1 fit:7 zi:3 gave:1 hastie:4 reduce:2 idea:2 tradeoff:1 multiclass:1 motivated:1 expression:1 effort:1 cause:2 clear:2 generate:2 maxpi:1 outperform:1 zj:1 notice:1 dotted:1 gil:1 estimated:2 correctly:2 group:25 four:2 pb:1 changing:2 pj:1 asymptotically:1 sum:1 you:1 almost:2 reasonable:1 separation:1 summarizes:1 vb:2 bound:1 correspondence:1 infinity:1 your:2 min:2 optimality:2 relatively:1 department:1 march:1 smaller:2 slowed:1 restricted:1 pr:4 errorcorrecting:1 fji:1 available:1 apply:1 appropriate:1 original:1 standardized:1 remaining:1 bakiri:3 question:1 quantity:3 usual:1 said:1 distance:5 simulated:3 unstable:1 reason:1 code:1 relationship:2 equivalently:1 unfortunately:1 stated:1 upper:2 observation:2 variability:5 pic:9 pair:1 unduly:1 suggested:1 below:1 ftj:1 max:18 explanation:1 power:2 unrealistic:1 pact:42 improve:3 literature:1 law:1 loss:1 expect:4 var:1 degree:1 pij:2 systematically:1 classifying:2 pi:8 row:5 course:2 repeat:1 mve:2 last:2 jth:4 bias:3 distributed:2 curve:4 calculated:1 rich:1 fonned:2 assumed:2 why:6 table:1 expansion:1 necessarily:2 did:1 main:1 motivation:3 allowed:1 attributable:1 shrinking:1 exponential:4 lie:1 theorem:19 bad:2 pac:1 r2:1 insignificant:1 concern:1 bivariate:1 exists:1 ih:2 corr:1 ci:1 illustrates:1 zjt:1 simply:2 explore:1 corresponds:1 adaption:1 gareth:1 conditional:4 viewed:1 identity:1 towards:1 change:5 corrected:2 uniformly:1 zb:1 ili:2 e:1 attempted:1 support:1 dept:1 tested:1 |
527 | 1,482 | A Principle for Unsupervised
Hierarchical Decomposition of Visual Scenes
Michael C. Mozer
Dept. of Computer Science
University of Colorado
Boulder, CO 80309-0430
ABSTRACT
Structure in a visual scene can be described at many levels of granularity. At a coarse level, the scene is composed of objects; at a finer level,
each object is made up of parts, and the parts of subparts. In this work, I
propose a simple principle by which such hierarchical structure can be
extracted from visual scenes: Regularity in the relations among different
parts of an object is weaker than in the internal structure of a part. This
principle can be applied recursively to define part-whole relationships
among elements in a scene. The principle does not make use of object
models, categories, or other sorts of higher-level knowledge; rather,
part-whole relationships can be established based on the statistics of a
set of sample visual scenes. I illustrate with a model that performs unsupervised decomposition of simple scenes. The model can account for
the results from a human learning experiment on the ontogeny of partwhole relationships.
1 INTRODUCTION
The structure in a visual scene can be described at many levels of granularity. Consider the scene in Figure I a. At a coarse level, the scene might be said to consist of stick
man and stick dog. However, stick man and stick dog themselves can be decomposed further. One might describe stick man as having two components, a head and a body. The
head in turn can be described in terms of its parts: the eyes, nose, and mouth. This sort of
scene decomposition can continue recursively down to the level of the primitive visual features. Figure I b shows a partial decomposition of the scene in Figure I a.
A scene decomposition establishes part-whole relationships among objects. For
example, the mouth (a whole) consists of two parts, the teeth and the lips. If we assume
that any part can belong to only one whole, the decomposition imposes a hierarchical
structure over the elements in the scene.
Where does this structure come from? What makes an object an object, a part a part?
I propose a simple principle by which such hierarchical structure can be extracted from
visual scenes and incorporate the principle in a simulation model. The principle is based
on the statistics of the visual environment, not on object models or other sorts of higherlevel knowledge, or on a teacher to classify objects or their parts.
53
Hierarchical Decomposition of Visual Scenes
2 WHAT MAKES A PART A PART?
Parts combine to form objects. Parts are combined in different ways to form different
objects and different instances of an object. Consequently, the structural relations among
different parts of an object are less regular than is the internal structure of a part. To illustrate, consider Figure 2, which depicts four instances of a box shell and lid. The components of the lid-the top and the handle-appear in a regular configuration, as do the
components of the shell-the sides and base-but the relation of the lid to the shell is variable. Thus, configural regularity is an indication that components should be grouped
together to form a unit. I call this the regularity principle. Other variants of the regularity
principle have been suggested by Becker (1995) and Tenenbaum (1994).
The regularity depicted in Figure 2 is quite rigid: one component of a part always
occurs in a fixed spatial position relative to another. The regularity principle can also be
cast in terms of abstract relationships such as containment and encirclement. The only difference is the featural representation that subserves the regularity discovery process. In
this paper, however, I address primarily regularities that are based on physical features and
fixed spatial relationships. Another generalization of the regularity principle is that it can
be applied recursively to suggest not only parts of wholes, but subparts of parts .
According to the regularity principle, information is implicit in the environment that
can be used to establish part-whole relationships. This information comes in the form of
statistical regularities among features in a visual scene. The regularity principle does not
depend on explicit labeling of parts or objects.
In contrast, Schyns and Murphy (1992, 1993) have suggested a theory of part ontogeny that presupposes explicit categorization of objects. They propose a homogeneity principle which states that "if a fragment of a stimulus plays a consistent role in
categorization, the perceptual parts composing the fragment are instantiated as a single
unit in the stimulus representation in memory." Their empirical studies with human subjects find support for the homogeneity principle.
Superficially, the homogeneity and regularity principles seem quite different: while
the homogeneity principle applies to supervised category learning (i.e., with a teacher to
classify instances), the regularity principle applies to unsupervised discovery. But it is possible to transform one learning paradigm into the other. For example, in a category learning task, if only one category is to be learned and if the training examples are all positive
instances of the category, then inducing the defining characteristics of the category is
equivalent to extracting regularities in the stimulus environment. Thus, category learning
in a diverse stimulus environment can be conceptualized as unsupervised regularity
extraction in multiple, narrow stimulus environments (each environment being formed by
taking all positive instances of a given class).
(a)
scene
(b)
~
stick dog
stick man
~
body
head
~
nose mouth
eyes
~
torso leg
arm
~teeth
lips
FIGURE 1. (a) A graphical depiction of stick man and his faithful companion, stick dog; (b) a
partial decomposition of the scene into its parts.
FIGURE 2. Four different instances of a box with a lid
M. C. Mozer
54
There are several other differences between the regularity principle proposed here
and the homogeneity principle of Schyns and Murphy, but they are minor. Schyns and
Murphy seem to interpret "fragment" more narrowly as spatially contiguous perceptual
features. They also don't address the hierarchical nature of part-whole relationships.
Nonetheless, the two principles share the notion of using the statistical structure of the
visual environment to establish part-whole relations.
3 A FLAT REPRESENTATION OF STRUCTURE
I have incorporated the regularity principle into a neural net that discovers partwhole relations in its environment. Neural nets, having powerful learning paradigms for
unsupervised discovery, are well suited for this task. However, they have a fundamental
difficulty representing complex, articulated data structures of the sort necessary to encode
hierarchies (but see Pollack, 1988, and Smolensky, 1990, for promising advances). I thus
begin by describing a novel representation scheme for hierarchical structures that can
readily be integrated into a neural net.
The tree structure in Figure I b depicts one representation of a hierarchical decomposition. The complete tree has as its leaf nodes the primitive visual features of the scene.
The tree specifies the relationships among the visual features. There is another way of capturing these relationships, more connectionist in spirit than the tree structure. The idea is
to assign to each primitive feature a tag-a scalar in [0, I)-such that features within a
subtree have similar values. For the features of stick man, possible tags might be: eyes .1,
nose .2, lips .28, teeth .32, arm .6, torso .7, leg .8.
Denoting the set of all features having tags in [a, ~] by Sea, ~), one can specify any
subtree of the stick man representation. For example, S(O, 1) includes all features of stick
man; S(0,.5) includes all features in the subtree whose root is stick man's head, S(.5,I) his
body; S(.25,.35) indicates the parts of the mouth. By a simple algorithm, tags can be
assigned to the leaf nodes of any tree such that any subtree can be selected by specifying
an appropriate tag range. The only requirement for this algorithm is knowledge of the
maximum branching factor. There is no fixed limit to the depth of the tree that can be thus
represented; however, the deeper the tree, the finer the tag resolution that wiII be needed.
The tags provide a "flat" way of representing hierarchical structure. Although the
tree is implicit in the representation, the tags convey all information in the tree, and thus
can capture complex, articulated structures. The tags in fact convey additional information. For example in the above feature list, note that lips is closer to nose than teeth is to
nose. This information can easily be ignored, but it is still worth observing that the tags
carry extra baggage not present in the symbolic tree structure.
It is convenient to represent the tags on a range [0, 21t) rather than [0, I]. This allows
the tag to be identified with a directional-or angular-value. Viewed as part of a cyclic
continuum, the directional tags are homogeneous, in contrast to the linear tags where tags
near and 1 have special status by virtue of being at endpoints of the continuum. Homogeneity results in a more elegant model, as described below.
The directional tags also permit a neurophysiological interpretation, albeit speculative. It has been suggested that synchronized oscillatory activities in the nervous system
can be used to convey information above and beyond that contained in the average firing
rate of individual neurons (e.g., Eckhorn et aI., 1988; Gray et aI., 1989; von der Malsburg,
1981). These osciIIations vary in their phase, the relative offset of the bursts. The directional tags could map directly to phases of oscillations, providing a means of implementing the tagging in neocortex.
?
4 REGULARITY DISCOVERY
Many learning paradigms allow for the djscovery of regUlarity. I have used an
autoencoder architecture (Plaut, Nowlan, & Hinton, 1986) that maps an input pattern-a
55
Hierarchical Decomposition of Visual Scenes
representation of visual features in a scene-to an output pattern via a small layer of hidden units. The goal of this type of architecture is for the network to reproduce the input
pattern over the output units. The task requires discovery of regularities because the hidden layer serves as an encoding bottleneck that limits the representational capacity of the
system. Consequently, stronger regularities (the most common patterns) will be encoded
over the weaker.
5 MAGIC
We now need to combine the autoencoder architecture with the notion of tags such
that regularity of feature configurations in the input will increase the likelihood that the
features will be assigned the same tags.
This goal can be achieved using a model we developed for segmenting an image into
different objects using supervised learning. The model, MAGIC (Mozer, Zemel, Behrmann, & Williams, 1992), was trained on images containing several visual objects and its
task was to tag features according to which object they belonged. A teacher provided the
target tags. Each unit in MAGIC conveys two distinct values: a probability that a feature is
present, which I will call the feature activity, and a tag associated with the feature. The tag
is a directional (angular) value, of the sort suggested earlier. (The tag representation is in
reality a complex number whose direction corresponds to the directional value and whose
magnitude is related to the unit's confidence in the direction. As this latter aspect of the
representation is not central to the present work, I discuss it no further.)
The architecture is a two layer recurrent net. The input or feature layer is set of spatiotopic arrays-in most simulations having dimensions 25x25-each array containing
detectors for features of a given type: oriented line segments at 0 ,45 ,90 and 135 In
addition, there is a layer of hidden units. Each hidden unit is reciprocally connected to
input from a local spatial patch of the input array; in the current simulations, the patch has
dimensions 4x4. For each patch there is a corresponding fixed-size pool of hidden units.
To achieve a translation invariant response across the image, the pools are arranged in a
spatiotopic array in which neighboring pools respond to neighboring patches and the
patch-to-pool weights are constrained to be the same at all locations in the array. There are
interlayer connections, but no intralayer connections.
The images presented to MAGIC consist of an arrangement of features over the input
array. The feature activity is clamped on (i.e., the feature is present), and the initial directional tag of the feature is set at random. Feature unit activities and tags feed to the hidden
units, which in turn feed back to the feature units. Through a relaxation process, the system settles on an assignment of tags to the feature units (as well as to the hidden units,
although read out from the model concerns only the feature units). MAGIC is a mean-field
approximation to a stochastic network of directional units with binary-gated outputs
(Zemel, Williams, & Mozer, 1995). This means that a mean-field energy functional can be
written that expresses the network state and controls the dynamics; consequently, MAGIC
is guaranteed to converge to a stable pattern of tags.
Each hidden unit detects a spatially local configuration offeatures, and it acts to reinstate a pattern of tags over the configuration. By adjusting its incoming and outgoing
weights during training, the hidden unit is made to respond to configurations that are consistently tagged in the training set. For example, if the training set contains many corner
junctions where horizontal and vertical lines come to a point and if the teacher tags all features composing these lines as belonging to the same object, then a hidden unit might
learn to detect this configuration, and when it does so, to force the tags of the component
features to be the same.
In our earlier work, MAGIC was trained to map the feature activity pattern to a target
pattern of feature tags, where there was a distinct tag for each object in the image. In the
present work, the training objective is rather to impose uniform tags over the features.
Additionally, the training objective encourages MAGIC to reinstate the feature activity
0
0
0
,
0
?
M C. Mozer
56
Iteration 1
,, - , ---~-
i
I
Iteration 2
Iteration 4
Iteration 6
Iteration II
----- -;,- ~
I
I
Directional Tag Spectrum
_
AI!II.?,
1 1
1.'I!t~
FIGURE 3. The state of MAGIC as processing proceeds for an image composed of a pair of
lines made out of horizontal and vertical line segments. The coloring of a segment represents
the directional tag. The segments belonging to a line are randomly tagged initially; over
processing iterations, these tags are brought into alignment.
pattern over the feature units; that is, the hidden units must encode and propagate information back to the feature units that is sufficient to specify the feature activities (if the feature
activities weren't clamped). With this training criterion, MAGIC becomes a type of
autoencoder. The key property of MAGIC is that it can assign a feature configuration the
same tag only if it learns to encode the configuration. If an arrangement is not encoded,
there will be no force to align the feature tags. Further, fixed weak inhibitory connections
between every pair of feature units serve to spread the tags apart if the force to align them
is not strong enough.
Note that this training paradigm does not require a teacher to tag features as belonging to one part or another. MAGIC will try to tag all features as belonging to the same part,
but it is able to do so only for configurations of features that it is able to encode. Consequently, highly regular and recurring configurations will be grouped together, and irregular
configurations will be pulled apart. The strength of grouping will be proportional to the
degree of regularity.
6 SIMULATION EXPERIMENTS
To illustrate the behavior of the model, I show a simple simulation in which MAGIC
is trained on pairs of lines, one vertical and one horizontal. Each line is made up of 6
colinear line segments. The segments are primitive input features of the model. The two
lines may appear in different positions relative to one another. Hence, the strongest regularity is in the segments that make up a line, not the junction between the lines. When
trained with two hidden units, MAGIC has sufficient resources to encode the structure
within each line, but not the relationships among the lines; because this structure is not
encoded, the features of the two lines are not assigned the same tags (Figure 3).
Although each "part" is made up of features having a uniform orientation and in a
colinear arrangement, the composition and structure of the parts is immaterial; MAGIC's
performance depends only on the regularity of the configurations. In the next set of simulations, MAGIC discovers regularities of a more arbitrary nature.
6.1 MODELING HUMAN LEARNING OF PART-WHOLE RELATIONS
Schyns and Murphy (1992) studied the ontogeny of part-whole relationships by
training human subjects on a novel class of objects and then examining how the subjects
decomposed the objects into their parts. I briefly describe their experiment, followed by a
simulation that accounts for their results.
In the first phase of the experiment, subjects were shown 3-D gray level "martian
rocks" on a CRT screen. The rocks were constructed by deforming a sphere, resulting in
various bumps or protrusions. Subjects watched the rocks rotating on the screen, allowing
them to view the rock from all sides. Subjects were shown six instances, all of which were
labeled "M 1 rocks" and were then tested to determine whether they could distinguish M 1
57
Hierarchical Decomposition of Visual Scenes
rocks from other rocks. Subjects continued training until they performed correctly on this
task. Every Ml rock was divided into octants; the protrusions on seven of the octants were
generated randomly, and the protrusions on the last octant were the same for all Ml rocks.
Two groups of subjects were studied. The A group saw M I rocks all having part A, the B
group saw M 1 rocks all having part B. Following training, subjects were asked to delineate the parts they thought were important on various exemplars. Subjects selected the target part from the category on which they were trained 93% of the time, and the alternative
target-the target from the other category-only 8% of the time, indicating that the learning task made a part dramatically more salient.
To model this phase of the experiment, I generated two dimensional contours of the
same flavor as Schyns and Murphy's martian rocks (Figure 4). Each rock-can it a "venusian rock" for distinction-can be divided into four quadrants or parts. Two groups of
venusian rocks were generated. Rocks of category A an contained part A (left panel, Figure 4), rocks of category B contained part B (center panel, Figure 4). One network was
trained on six exemplars of category A rocks, another network was trained on six exemplars of category B rocks. Then, with learning turned off, both networks were tested on
five presentations each of twelve new exemplars, six each of categories A and B.
Just as the human subjects were instructed to delineate parts, we must ask MAGIC to
do the same. One approach would be to run the model with a test stimulus and, once it settles, select an features having directional tags clustered tightly together as belonging to the
same part. However, this requires specifying and tuning a clustering procedure. To avoid
this additional step, I simply compared how tightly clustered were the tags of the target
part relative to those of the alternative target. I used a directional variance measure that
yields a value of 0 if all tags are identical and I if the tags are distributed uniformly over
the directional spectrum. By this measure, the variance was .30 for the target part and .68
for the alternative target (F(l, 118) = 322.0, P < .001), indicating that the grouping of features of the target part was significantly stronger. This replicates, at least qualitatively, the
finding of Schyns and Murphy.
In a second phase of Schyns and Murphy's experiment, subjects were trained on category C rocks, which were formed by adjoining parts A and B and generating the remaining six octants at random. Following training, subjects were again asked to delineate parts.
All subjects delineated A and B as distinct parts. In contrast, a naive group of subjects who
were trained on category C alone always grouped A and B together as a single part.
To model this phase, I generated six category C venusian rocks that had both parts A
and B (right panel, Figure 4). The versions of MAGIC that had been trained on category A
and B rocks alone were now trained on category C rocks. As a control condition, a third
version of MAGIC was trained from scratch on category C rocks alone. I compared the
tightness of clustering of the combined A-B part for the first two nets to the third. Using
the same variance measure as above, the nets that first received training on parts A and B
alone yielded a variance of .57, and the net that was only trained on the combined A-B
part yielded a variance of .47 (F(1,88) = 7.02, P < .02). One cannot directly compare the
variance of the A-B part to that of the A and B parts alone, because the measure is structured such that parts with more features always yield larger variances. However, one can
compare the two conditions using the relative variance of the combined A-B part to the A
'S'~7
t.r
~
~
-,~
(-"to.
~~"
I~~
,,~
~~""-........
,
~~
~
-,-./
,~~
FIGURE 4. Three examples of the martian rock stimuli used to train MAGIC. From left to
right, the rocks are of categories A, B, and C. The lighter regions are the contours that define
rocks of a given category.
58
M C. Mozer
and B parts alone. This yielded the same outcome as before (.21 for the first two nets, .12
for the third net, F(l,88) = 5.80, p < .02). Thus, MAGIC is also able to account for the
effects of prior learning on part ontogeny.
7 CONCLUSIONS
The regularity principle proposed in this work seems consistent with the homogeneity principle proposed earlier by Schyns and Murphy (1991, 1992). Indeed, MAGIC is
able to model Schyns and Murphy's data using an unsupervised training paradigm,
although Schyns and Murphy framed their experiment as a classification task.
This work is but a start at modeling the development of part-whole hierarchies based
on perceptual experience. MAGIC requires further elaboration, and I am somewhat skeptical that it is sufficiently powerful in its present form to be pushed much further. The main
issue restricting it is the representation of input features. The oriented-line-segment features are certainly too primitive and inflexible a representation. For example, MAGIC
could not be trained to recognize the lid and shell of Figure 2 because it encodes the orientation of the features with respect to the image plane, not with respect to one another. Minimally, the representation requires some version of scale and rotation invariance.
Perhaps the most interesting computational'issue raised by MAGIC is how the pattern of feature tags is mapped into an explicit part-whole decomposition. This involves
clustering together the similar tags as a unit, or possibly selecting all tags in a given range.
To do so requires specification of additional parameters that are external to the model
(e.g., how tight the cluster should be, how broad the range should be, around what tag
direction it should be centered). These parameters are deeply related to attentional issues,
and a current direction of research is to explore this relationship.
8 ACKNOWLEDGEMENTS
This research was supported by NSF PYI award IRI-9058450 and grant 97-18 from the McDonnell-Pew Program in Cognitive Neuroscience.
9 REFERENCES
Becker, S. (1995). JPMAX: Learning to recognize moving objects as a model-fitting problem. In G. Tesauro, D.
S. Touretzky, & T. K. Leen (Eds), Advances in Neural Informatio/l ProcessinK Systems 7 (pp. 933-940).
Cambridge, MA: MIT Press.
Eckhorn, Roo Bauer, R., Jordan, w., Brosch, M., Kruse, w., Munk, M., & Reitboek, H. J. (1988). Coherent oscillations: A mechanism of feature linking in the visual cortex? Biological Cybernetics, 60, 121-130.
Gray, C. M., Koenig, P., Engel, A. K., & Singer, W. (1989). Oscillatory responses in cat visual cortex exhibit
intercolumnar synchronization which reflects global stimulus properties. Nature (London), 338, 334-337.
Mozer, M. c., Zemel, R. S., Behrmann, M .. & Williams, C. K. I. (1992). Learning to segment images using
dynamic feature binding. Neural Computation, 4, 650-666.
Plaut, D. C. , Nowlan, S., & Hinton, G. E . (1986). Experiments 011 leaminK by back propagation (Technical
report CMU-CS-86- 126). Pittsburgh, PA: Carnegie-Mellon University, Department of Computer Science.
Pollack, J. B. (1988). Recursive auto-associative memory: Devising compositional distributed representations . In
Proceedings of the Tenth Annual Conference of the Cognitive Science Society (pp. 33-39). Hillsdale, NJ :
Erlbaum.
Schyns, P. G., & Murphy, G. L. (1992). The ontogeny of units in object categories. In Proceedings of the Fourteenth Annual Conference (!fthe COKnitil'e Science Society (pp. 197-202). Hillsdale, NJ: Erlbaum.
Schyns, P. G. , & Murphy, G. L. (1993). The ontogeny of transformable part representations in object concepts. In
Proceedings of the Fifteenth Annual Conference of the OIKllitive Scien ce Society (pp. 917-922). Hillsdale.
NJ: Erlbaum.
Smolensky, P. (1990). Tensor product variable binding and the representation of symbolic structures in connectionist networks. Artificial Intelligence, 46. 159-2 I 6.
Tenenbaum, J. B. (1994). Functional parts. In A. Ram & K. Eiselt (Eds .). Pmceedings (If the Sixteenth An/lual
Conference of the Cognitive Science Society (pp. 864-869). Hillsdale, NJ : Erlbaum.
von der Malsburg, C. (1981). The correlatioll theory of brain fun ction (Internal Report 81-2) . Goettingen:
Department of Neurobiology, Max Planck Institute for Biophysical Chemistry.
Zemel, R. S., Williams, C. K. I.. & Mozer, M. C. (1995). Lending direction to neural networks . Neural Networks ,
8, 503-512.
| 1482 |@word version:3 briefly:1 stronger:2 seems:1 simulation:7 propagate:1 decomposition:12 recursively:3 carry:1 initial:1 configuration:12 cyclic:1 fragment:3 contains:1 selecting:1 denoting:1 current:2 nowlan:2 written:1 readily:1 must:2 alone:6 intelligence:1 leaf:2 selected:2 nervous:1 devising:1 plane:1 coarse:2 plaut:2 node:2 location:1 lending:1 five:1 burst:1 constructed:1 consists:1 combine:2 fitting:1 interlayer:1 ontogeny:6 tagging:1 indeed:1 behavior:1 themselves:1 brain:1 detects:1 decomposed:2 becomes:1 begin:1 provided:1 panel:3 what:3 developed:1 finding:1 nj:4 configural:1 every:2 act:1 fun:1 stick:13 control:2 unit:26 grant:1 appear:2 planck:1 segmenting:1 positive:2 before:1 local:2 limit:2 encoding:1 firing:1 might:4 minimally:1 studied:2 specifying:2 co:1 scien:1 range:4 faithful:1 recursive:1 procedure:1 empirical:1 thought:1 significantly:1 convenient:1 confidence:1 regular:3 quadrant:1 suggest:1 symbolic:2 cannot:1 equivalent:1 map:3 center:1 conceptualized:1 primitive:5 williams:4 iri:1 resolution:1 pyi:1 reitboek:1 continued:1 array:6 his:2 handle:1 notion:2 hierarchy:2 play:1 colorado:1 target:10 lighter:1 homogeneous:1 pa:1 element:2 labeled:1 role:1 capture:1 region:1 connected:1 deeply:1 mozer:8 environment:8 asked:2 dynamic:2 trained:14 depend:1 colinear:2 segment:9 immaterial:1 tight:1 serve:1 easily:1 represented:1 various:2 cat:1 articulated:2 train:1 instantiated:1 distinct:3 describe:2 london:1 ction:1 artificial:1 zemel:4 labeling:1 outcome:1 quite:2 whose:3 encoded:3 presupposes:1 larger:1 tightness:1 roo:1 statistic:2 transform:1 associative:1 higherlevel:1 indication:1 biophysical:1 net:9 rock:27 propose:3 product:1 neighboring:2 turned:1 achieve:1 representational:1 sixteenth:1 protrusion:3 inducing:1 partwhole:2 regularity:28 requirement:1 cluster:1 sea:1 categorization:2 generating:1 object:25 illustrate:3 recurrent:1 exemplar:4 minor:1 received:1 strong:1 c:1 involves:1 come:3 synchronized:1 direction:5 stochastic:1 centered:1 human:5 crt:1 settle:2 implementing:1 munk:1 hillsdale:4 require:1 assign:2 generalization:1 weren:1 clustered:2 biological:1 goettingen:1 sufficiently:1 around:1 baggage:1 bump:1 continuum:2 vary:1 saw:2 grouped:3 engel:1 establishes:1 reflects:1 brought:1 mit:1 always:3 rather:3 avoid:1 encode:5 consistently:1 indicates:1 likelihood:1 contrast:3 detect:1 am:1 rigid:1 integrated:1 initially:1 hidden:12 relation:6 reproduce:1 issue:3 among:7 orientation:2 classification:1 development:1 spatial:3 special:1 constrained:1 raised:1 field:2 once:1 having:8 extraction:1 x4:1 represents:1 identical:1 broad:1 unsupervised:6 connectionist:2 stimulus:8 report:2 primarily:1 oriented:2 randomly:2 composed:2 tightly:2 homogeneity:7 individual:1 recognize:2 murphy:12 phase:6 highly:1 certainly:1 alignment:1 replicates:1 adjoining:1 closer:1 partial:2 necessary:1 experience:1 tree:10 rotating:1 pollack:2 instance:7 classify:2 earlier:3 modeling:2 contiguous:1 assignment:1 uniform:2 examining:1 erlbaum:4 too:1 spatiotopic:2 teacher:5 combined:4 fundamental:1 twelve:1 off:1 pool:4 michael:1 together:5 offeatures:1 von:2 central:1 venusian:3 again:1 containing:2 possibly:1 corner:1 external:1 cognitive:3 account:3 chemistry:1 includes:2 depends:1 performed:1 root:1 try:1 view:1 observing:1 start:1 sort:5 formed:2 variance:8 characteristic:1 who:1 yield:2 directional:13 weak:1 worth:1 finer:2 cybernetics:1 oscillatory:2 detector:1 strongest:1 touretzky:1 jpmax:1 ed:2 nonetheless:1 energy:1 pp:5 conveys:1 associated:1 adjusting:1 ask:1 knowledge:3 torso:2 back:3 coloring:1 feed:2 higher:1 supervised:2 specify:2 response:2 arranged:1 leen:1 box:2 delineate:3 angular:2 implicit:2 just:1 until:1 koenig:1 horizontal:3 propagation:1 gray:3 perhaps:1 effect:1 concept:1 tagged:2 assigned:3 hence:1 spatially:2 read:1 during:1 branching:1 encourages:1 intercolumnar:1 criterion:1 complete:1 performs:1 image:8 discovers:2 novel:2 common:1 speculative:1 rotation:1 functional:2 physical:1 endpoint:1 belong:1 interpretation:1 linking:1 interpret:1 mellon:1 composition:1 cambridge:1 ai:3 framed:1 tuning:1 pew:1 eckhorn:2 wiii:1 had:2 moving:1 stable:1 specification:1 cortex:2 depiction:1 base:1 align:2 apart:2 tesauro:1 binary:1 continue:1 der:2 additional:3 somewhat:1 impose:1 converge:1 paradigm:5 determine:1 kruse:1 ii:2 multiple:1 technical:1 sphere:1 elaboration:1 divided:2 award:1 watched:1 variant:1 cmu:1 fifteenth:1 iteration:6 represent:1 achieved:1 irregular:1 addition:1 extra:1 subject:15 elegant:1 spirit:1 seem:2 jordan:1 call:2 extracting:1 structural:1 near:1 granularity:2 enough:1 architecture:4 identified:1 idea:1 narrowly:1 bottleneck:1 six:6 whether:1 becker:2 compositional:1 ignored:1 dramatically:1 neocortex:1 tenenbaum:2 category:23 specifies:1 nsf:1 inhibitory:1 neuroscience:1 correctly:1 subpart:2 diverse:1 carnegie:1 transformable:1 express:1 group:5 key:1 four:3 salient:1 ce:1 tenth:1 ram:1 relaxation:1 run:1 fourteenth:1 powerful:2 respond:2 patch:5 oscillation:2 pushed:1 capturing:1 layer:5 guaranteed:1 followed:1 distinguish:1 yielded:3 activity:8 annual:3 strength:1 scene:23 flat:2 encodes:1 tag:51 aspect:1 structured:1 department:2 according:2 mcdonnell:1 belonging:5 across:1 inflexible:1 lid:5 delineated:1 leg:2 invariant:1 boulder:1 resource:1 brosch:1 turn:2 describing:1 discus:1 mechanism:1 needed:1 singer:1 nose:5 serf:1 junction:2 permit:1 hierarchical:11 appropriate:1 alternative:3 reinstate:2 top:1 clustering:3 remaining:1 graphical:1 malsburg:2 establish:2 society:4 tensor:1 objective:2 arrangement:3 occurs:1 fthe:1 said:1 exhibit:1 attentional:1 mapped:1 capacity:1 seven:1 relationship:13 providing:1 magic:25 gated:1 allowing:1 vertical:3 neuron:1 defining:1 hinton:2 incorporated:1 head:4 neurobiology:1 arbitrary:1 dog:4 cast:1 pair:3 connection:3 coherent:1 learned:1 narrow:1 distinction:1 established:1 address:2 beyond:1 suggested:4 proceeds:1 below:1 pattern:10 able:4 recurring:1 smolensky:2 belonged:1 program:1 max:1 memory:2 reciprocally:1 mouth:4 difficulty:1 force:3 arm:2 representing:2 scheme:1 eye:3 autoencoder:3 naive:1 featural:1 auto:1 prior:1 discovery:5 acknowledgement:1 relative:5 synchronization:1 interesting:1 proportional:1 degree:1 teeth:4 x25:1 sufficient:2 consistent:2 imposes:1 principle:24 share:1 translation:1 skeptical:1 supported:1 last:1 side:2 weaker:2 deeper:1 allow:1 pulled:1 institute:1 taking:1 distributed:2 bauer:1 depth:1 dimension:2 superficially:1 contour:2 instructed:1 made:6 qualitatively:1 status:1 ml:2 global:1 incoming:1 containment:1 pittsburgh:1 don:1 spectrum:2 reality:1 lip:4 promising:1 nature:3 learn:1 additionally:1 composing:2 pmceedings:1 schyns:12 complex:3 intralayer:1 spread:1 main:1 whole:13 convey:3 martian:3 body:3 depicts:2 screen:2 position:2 explicit:3 clamped:2 perceptual:3 behrmann:2 third:3 learns:1 subserves:1 down:1 companion:1 list:1 offset:1 virtue:1 concern:1 grouping:2 consist:2 albeit:1 restricting:1 magnitude:1 subtree:4 flavor:1 suited:1 depicted:1 octant:4 simply:1 explore:1 neurophysiological:1 visual:19 contained:3 scalar:1 applies:2 binding:2 corresponds:1 extracted:2 ma:1 shell:4 viewed:1 goal:2 presentation:1 consequently:4 man:9 uniformly:1 invariance:1 deforming:1 indicating:2 select:1 internal:3 support:1 latter:1 incorporate:1 dept:1 outgoing:1 tested:2 scratch:1 |
528 | 1,483 | The Bias-Variance Tradeoff and the Randomized
GACV
Grace Wahba, Xiwu Lin and Fangyu Gao
Dong Xiang
Dept of Statistics
Univ of Wisconsin
1210 W Dayton Street
Madison, WI 53706
wahba,xiwu,fgao@stat.wisc.edu
SAS Institute, Inc.
SAS Campus Drive
Cary, NC 27513
sasdxx@unx.sas.com
Ronald Klein, MD and Barbara Klein, MD
Dept of Ophthalmalogy
610 North Walnut Street
Madison, WI 53706
kleinr,kleinb@epi.ophth.wisc.edu
Abstract
We propose a new in-sample cross validation based method (randomized
GACV) for choosing smoothing or bandwidth parameters that govern the
bias-variance or fit-complexity tradeoff in 'soft' classification. Soft classification refers to a learning procedure which estimates the probability
that an example with a given attribute vector is in class 1 vs class O. The
target for optimizing the the tradeoff is the Kullback-Liebler distance
between the estimated probability distribution and the 'true' probability distribution, representing knowledge of an infinite population. The
method uses a randomized estimate of the trace of a Hessian and mimics
cross validation at the cost of a single relearning with perturbed outcome
data.
1 INTRODUCTION
We propose and test a new in-sample cross-validation based method for optimizing the biasvariance tradeoff in 'soft classification' (Wahba et al1994), called ranG ACV (randomized
Generalized Approximate Cross Validation) . Summarizing from Wahba et al(l994) we are
given a training set consisting of n examples, where for each example we have a vector
t E T of attribute values, and an outcome y, which is either 0 or 1. Based on the training
data it is desired to estimate the probability p of the outcome 1 for any new examples in the
The Bias-Variance TradeofJand the Randomized GACV
621
future. In 'soft' classification the estimate p(t) of p(t) is of particular interest, and might be
used by a physician to tell patients how they might modify their risk p by changing (some
component of) t, for example, cholesterol as a risk factor for heart attack. Penalized likelihood estimates are obtained for p by assuming that the logit f(t), t E T, which satisfies
p(t) = ef(t) 1(1 + ef(t?) is in some space 1{ of functions . Technically 1{ is a reproducing
kernel Hilbert space, but you don't need to know what that is to read on. Let the training
set be {Yi, ti, i = 1,???, n}. Letting Ii = f(td, the negative log likelihood .c{Yi, ti, fd of
the observations, given f is
n
.c{Yi, ti, fd = 2::[-Ydi + b(li)],
(1)
i=1
where b(f) = log(l + ef ). The penalized likelihood estimate of the function f is the
solution to: Find f E 1{ to minimize h. (I):
n
h.(f) = 2::[-Ydi + b(ld) + J>.(I),
(2)
i =1
where 1>.(1) is a quadratic penalty functional depending on parameter(s) A = (AI, ... , Aq)
which govern the so called bias-variance tradeoff. Equivalently the components of A control the tradeoff between the complexity of f and the fit to the training data. In this paper we
sketch the derivation of the ranG ACV method for choosing A, and present some preliminary but favorable simulation results, demonstrating its efficacy. This method is designed
for use with penalized likelihood estimates, but it is clear that it can be used with a variety
of other methods which contain bias-variance parameters to be chosen, and for which minimizing the Kullback-Liebler (K L) distance is the target. In the work of which this is a part,
we are concerned with A having multiple components. Thus, it will be highly convenient
to have an in-sample method for selecting A, if one that is accurate and computationally
convenient can be found.
Let P>. be the the estimate and p be the 'true' but unknown probability function and let
= p(td,p>.i = p>.(ti). For in-sample tuning, our criteria for a good choice of A is
Pi
the KL distance KL(p,p>.) = ~ E~I[PilogP7. + (1- pdlogg~::?)]. We may replace
K L(p,p>.) by the comparative K L distance (C K L), which differs from K L by a quantity
which does not depend on A. Letting hi = h (ti), the C K L is given by
1
CKL(p,p>.) == CKL(A) = ;;,
n
2:: [-pd>'i + b(l>.i)).
(3)
i=)
C K L(A) depends on the unknown p, and it is desired is to have a good estimate or proxy
for it, which can then be minimized with respect to A.
It is known (Wong 1992) that no exact unbiased estimate of CK L(A) exists in this case, so
that only approximate methods are possible. A number of authors have tackled this problem, including Utans and M90dy(1993), Liu(l993), Gu(1992). The iterative U BR method
of Gu(l992) is included in GRKPACK (Wang 1997), which implements general smoothing spline ANOVA penalized likelihood estimates with multiple smoothing parameters. It
has been successfully used in a number of practical problems, see, for example, Wahba
et al (1994,1995). The present work represents an approach in the spirit of GRKPACK
but which employs several approximations, and may be used with any data set, no matter
how large, provided that an algorithm for solving the penalized likelihood equations, either
exactly or approximately, can be implemented.
622
2
G. Wahba et al.
THE GACV ESTIMATE
In the general penalized likelihood problem the minimizer 1>,(-) of (2) has a representation
M
1>.(t)
=L
n
dv<Pv(t)
v=l
+
L CiQ>.(ti, t)
(4)
i=l
where the <Pv span the null space of 1>" Q>.(8, t) is a reproducing kernel (positive definite
function) for the penalized part of 7-1., and C = (Cl' ... , Cn )' satisfies M linear conditions,
so that there are (at most) n free parameters in 1>.. Typically the unpenalized functions
<Pv are low degree polynomials. Examples of Q(ti,') include radial basis functions and
various kinds of splines; minor modifications include sigmoidal basis functions, tree basis
functions and so on. See, for example Wahba( 1990, 1995), Girosi, Jones and Poggio( 1995).
If f>.C) is of the form (4) then 1>,(f>.) is a quadratic form in c. Substituting (4) into (2)
results in h a convex functional in C and d, and C and d are obtained numerically via a
Newton Raphson iteration, subject to the conditions on c. For large n, the second sum on
the right of (4) may be replaced by L~=1 Cik Q>. (ti k , t), where the tik are chosen via one
of several principled methods.
To obtain the CACV we begin with the ordinary leaving-out-one cross validation function
CV(.\) for the CKL:
n
[-i]
CV (.\) -_ -1 ""
LJ-yd>.i
n
+ b( 1>.i) ] ,
(5)
i=1
where fl- i ] the solution to the variational problem of (2) with the ith data point left out
and fti] is the value of fl- i] at ti . Although f>.C) is computed by solving for C and d
the CACV is derived in terms of the values (it"", fn)' of f at the ti. Where there is
no confusion between functions f(-) and vectors (it, ... ,fn)' of values of fat tl, ... ,tn,
we let f = (it, ... " fn)'. For any f(-) of the form (4), J>. (f) also has a representation as
a non-negative definite quadratic form in (it, ... , fn)'. Letting L:>. be twice the matrix of
this quadratic form we can rewrite (2) as
1
n
h(f,Y)
= L[-Ydi + b(/i)] + 2f'L:>.f.
(6)
i=1
Let W = W(f) be the n x n diagonal matrix with (/ii == Pi(l - Pi) in the iith position.
Using the fact that (/ii is the second derivative of b(fi), we have that H = [W + L:>.] - 1
is the inverse Hessian of the variational problem (6). In Xiang and Wahba (1996), several
Taylor series approximations, along with a generalization of the leaving-out-one lemma
(see Wahba 1990) are applied to (5) to obtain an approximate cross validation function
ACV(.\), which is a second order approximation to CV(.\) . Letting hii be the iith entry
of H , the result is
CV(.\)
~ ACV('\) = .!. t[-Yd>.i + b(f>.i)] + .!. t
n
n
i= l
i=1
hiiYi(Yi - P>.i) .
[1 - hiwii]
Then the GACV is obtained from the ACV by replacing h ii by ~ L~1 hii
and replacing 1 - hiWii by ~tr[I - (Wl/2 HWl/2)], giving
CACV('\)
1~
]
tr(H)
== ~tr(H)
L~l Yi(Yi - P>.i)
(Wl/2HWl /2)] ,
= ;; t;;[-Yd>.i + b(1).i) + -n-tr[I _
(7)
(8)
where W is evaluated at 1>.. Numerical results based on an exact calculation of (8) appear
in Xiang and Wahba (1996). The exact calculation is limited to small n however.
623
The Bias-Variance TradeofJand the Randomized GACV
3 THE RANDOMIZED GACV ESTIMATE
Given any 'black box' which, given >., and a training set {Yi, ti} produces f>. (.) as the minimizer of (2), and thence f>. = (fA 1 , " ' , f>.n)', we can produce randomized estimates of
trH and tr[! - W 1 / 2 HW 1/ 2 J without having any explicit calculations of these matrices.
This is done by running the 'black box' on perturbed data {Vi + <5i , td. For the Yi Gaussian, randomized trace estimates of the Hessian of the variational problem (the 'influence
matrix') have been studied extensively and shown to be essentially as good as exact calculations for large n, see for example Girard(1998) . Randomized trace estimates are based
on the fact that if A is any square matrix and <5 is a zero mean random n-vector with independent components with variance
then E<5' A<5 = ~ tr A. See Gong et al( 1998) and
(TJ,
u"
references cited there for experimental results with multiple regularization parameters. Returning to the 0-1 data case, it is easy to see that the minimizer fA(') of 1;.. is continuous in
Y, not withstanding the fact that in our training set the Yi take on only values 0 or 1. Letting
if = UA1,"', f>.n)' be the minimizer of (6) given y = (Y1,"', Yn)', and if+O be the
minimizer given data y+<5 = (Y1 +<51, ... ,Yn +<5n )' (the ti remain fixed), Xiang and Wahba
(1997) show, again using Taylor series expansions, that if+O - ff ,. . ., [WUf) + ~AJ-1<5.
This suggests that ~<5'Uf+O - ff) provides an estimate oftr[W(ff) + ~At1. However,
u"
if we take the solution ff to the nonlinear system for the original data Y as the initial value
for a Newton-Raphson calculation of ff+O things become even simpler. Applying a one
step Newton-Raphson iteration gives
(9)
-<5 + PjfUf,Y) = -<5, and [:;~f(ff,Y + <5)J - 1
y + o ,l
y + o ,l 82 h(fY
8 2 h(fY
[ 8?7if
A' Y)J- 1 ,we have f A
- fYA + [8?7if
A' Y)J- 1 uJ: so that f A
- fYA
[WUf) + E At 1<5. The result is the following ranGACV function:
Since Pjf(ff,y + <5)
ranGACV(>.) =
=
n
.!. ~[n~
'I '+bU .)J+
Yz At
At
<5' (f Y+O,l
A
n
-
fY)
A
",n
(
)
wi=l Yi Yi - PAi
.
[<5'<5 - <5'WUf)Uf+O,l - ff)J
(10)
To reduce the variance in the term after the '+' in (10), we may draw R
independent replicate vectors <51,'" ,<5R , and replace the term after the '+' in
( 1O)b
1... ",R
o:(fr Hr . 1 -ff)
wr=l
n
ranGACV(>.) function.
4
y
R
2:7-1y.(y.-P>..)
[O~Or-O~ W(fn(f~+Ar . l-ff)1
to obtain an R-replicated
NUMERICAL RESULTS
In this section we present simulation results which are representative of more extensive
simulations to appear elsewhere. In each case, K < < n was chosen by a sequential clustering algorithm. In that case, the ti were grouped into K clusters and one member of each
cluster selected at random. The model is fit. Then the number of clusters is doubled and the
model is fit again. This procedure continues until the fit does not change. In the randomized
trace estimates the random variates were Gaussian. Penalty functionals were (multivariate
generalizations of) the cubic spline penalty functional>. fa1U" (X))2, and smoothing spline
ANOVA models were fit.
G. Wahba et at.
624
4.1
EXPERIMENT 1. SINGLE SMOOTHING PARAMETER
In this experiment t E [0,1], f(t) = 2sin(10t), ti = (i - .5)/500, i = 1,???,500. A
random number generator produced 'observations' Yi = 1 with probability Pi = el , /(1 +
eli), to get the training set. QA is given in Wahba( 1990) for this cubic spline case, K = 50.
Since the true P is known, the true CKL can be computed. Fig. l(a) gives a plot of
CK L(A) and 10 replicates of ranGACV(A). In each replicate R was taken as 1, and J
was generated anew as a Gaussian random vector with (115 = .001. Extensive simulations
with different (115 showed that the results were insensitive to (115 from 1.0 to 10- 6 ? The
minimizer of C K L is at the filled-in circle and the 10 minimizers of the 10 replicates of
ranGACV are the open circles. Anyone of these 10 provides a rather good estimate of
the A that goes with the filled-in circle. Fig. l(b) gives the same experiment, except that
this time R = 5. It can be seen that the minimizers ranGACV become even more reliable
estimates of the minimizer of C K L, and the C K L at all of the ranG ACV estimates are
actually quite close to its minimum value.
4.2
EXPERIMENT 2. ADDITIVE MODEL WITH A = (Al' A2)
Here t E [0,1] 0 [0,1]. n = 500 values of ti were generated randomly according to
a uniform distribution on the unit square and the Yi were generated according to Pi =
e li j(l + e l ,) with t = (Xl,X2) and f(t) = 5 sin 27rXl - 3sin27rX2. An additive model
as a special case of the smoothing spline ANOVA model (see Wahba et al, 1995), of the
form f(t) = /-l + h(xd + h(X2) with cubic spline penalties on hand h were used.
K = 50, (115 = .001, R = 5. Figure l(c) gives a plot of CK L(Al' A2) and Figure l(d)
gives a plot of ranGACV(Al, A2). The open circles mark the minimizer of ranGACV
in both plots and the filled in circle marks the minimizer of C K L. The inefficiency, as
measured by CKL()..)/minACKL(A) is 1.01. Inefficiencies near 1 are typical of our
other similar simulations.
4.3
EXPERIMENT 3. COMPARISON OF ranGACV AND UBR
This experiment used a model similar to the model fit by GRKPACK for the risk of
progression of diabetic retinopathy given t = (Xl, X2, X3) = (duration, glycosylated
hemoglobin, body mass index) in Wahba et al(l995) as 'truth'. A training set of 669
examples was generated according to that model, which had the structure f(Xl, X2, X3) =
/-l + fl (xd + h (X2) + h (X3) + fl,3 (Xl, X3). This (synthetic) training set was fit by GRKPACK and also using K = 50 basis functions with ranG ACV. Here there are P = 6
smoothing parameters (there are 3 smoothing parameters in f13) and the ranGACV function was searched by a downhill simplex method to find its minimizer. Since the 'truth' is
known, the CKL for)" and for the GRKPACK fit using the iterative UBR method were
computed. This was repeated 100 times, and the 100 pairs of C K L values appears in Figure l(e). It can be seen that the U BR and ranGACV give similar C K L values about 90%
of the time, while the ranG ACV has lower C K L for most of the remaining cases.
4.4
DATA ANALYSIS: AN APPLICATION
Figure 1(f) represents part of the results of a study of association at baseline of pigmentary
abnormalities with various risk factors in 2585 women between the ages of 43 and 86 in the
Beaver Dam Eye Study, R. Klein et al( 1995). The attributes are: Xl = age, X2 =body mass
index, X3 = systolic blood pressure, X4 = cholesterol. X5 and X6 are indicator variables for
taking hormones, and history of drinking. The smoothing spline ANOVA model fitted was
f(t) = /-l+dlXl +d2X2+ h(X3)+ f4(X4)+ h4(X3, x4)+d5I(x5) +d6I(x6), where I is the
indicator function. Figure l(e) represents a cross section of the fit for X5 = no, X6 = no,
The Bias- Variance Tradeoff and the Randomized GACV
625
X2, X3 fixed at their medians and Xl fixed at the 75th percentile. The dotted lines are the
Bayesian confidence intervals, see Wahba et al( 1995). There is a suggestion of a borderline
inverse association of cholesterol. The reason for this association is uncertain. More details
will appear elsewhere.
Principled soft classification procedures can now be implemented in much larger data sets
than previously possible, and the ranG ACV should be applicable in general learning.
References
Girard, D. (1998), 'Asymptotic comparison of (partial) cross-validation, GCV and randomized GCV in nonparametric regression', Ann. Statist. 126, 315-334.
Girosi, F., Jones, M. & Poggio, T. (1995), 'Regularization theory and neural networks
architectures', Neural Computatioll 7,219-269.
Gong, J., Wahba, G., Johnson, D. & Tribbia, J. (1998), 'Adaptive tuning of numerical
weather prediction models: simultaneous estimation of weighting, smoothing and physical
parameters', MOllthly Weather Review 125, 210-231.
Gu, C. (1992), 'Penalized likelihood regression: a Bayesian analysis', Statistica Sinica
2,255-264.
Klein, R., Klein, B. & Moss, S. (1995), 'Age-related eye disease and survival. the Beaver
Dam Eye Study', Arch Ophthalmol113, 1995.
Liu, Y. (1993), Unbiased estimate of generalization error and model selection in neural
network, manuscript, Department of Physics, Institute of Brain and Neural Systems, Brown
University.
Utans, J. & Moody, J. (1993), Selecting neural network architectures via the prediction
risk: application to corporate bond rating prediction, in 'Proc. First Int'I Conf. on Artificial
Intelligence Applications on Wall Street', IEEE Computer Society Press.
Wahba, G. (1990), Spline Models for Observational Data, SIAM. CBMS-NSF Regional
Conference Series in Applied Mathematics, v. 59.
Wahba, G. (1995), Generalization and regularization in nonlinear learning systems, in
M. Arbib, ed., 'Handbook of Brain Theory and Neural Networks', MIT Press, pp. 426430.
Wahba, G., Wang, Y., Gu, c., Klein, R. & Klein, B. (1994), Structured machine learning
for 'soft' classification with smoothing spline ANOVA and stacked tuning, testing and
evaluation, in J. Cowan, G. Tesauro & J. Alspector, eds, 'Advances in Neural Information
Processing Systems 6', Morgan Kauffman, pp. 415-422.
Wahba, G., Wang, Y., Gu, C., Klein, R. & Klein, B. (1995), 'Smoothing spline AN OVA for
exponential families, with application to the Wisconsin Epidemiological Study of Diabetic
Retinopathy' , Ann. Statist. 23, 1865-1895.
Wang, Y. (1997), 'GRKPACK: Fitting smoothing spline analysis of variance models to data
from exponential families', Commun. Statist. Sim. Compo 26,765-782.
Wong, W. (1992), Estimation of the loss of an estimate, Technical Report 356, Dept. of
Statistics, University of Chicago, Chicago, II.
Xiang, D. & Wahba, G. (1996), 'A generalized approximate cross validation for smoothing
splines with non-Gaussian data', Statistica Sinica 6, 675-692, preprint TR 930 available
via www. stat. wise. edu/-wahba - > TRLIST.
Xiang, D. & Wahba, G. (1997), Approximate smoothing spline methods for large data sets
in the binary case, Technical Report 982, Department of Statistics, University of Wisconsin,
Madison WI. To appear in the Proceedings of the 1997 ASA Joint Statistical Meetings,
Biometrics Section, pp 94-98 (1998). Also in TRLIST as above.
G, Wahba et aI,
626
CKL
ranGACV
10
(0
c:i
(0
c:i
0
o
c:i
c:i
(0
10
10
c:i
CKL
10
(0
10
10
.' .
c:i
0
o
0
c:i
10
~
-8
-7
-6
-5
log lambda
(a)
-3
-4
9.29
-8
"f
\~7
0\
-7
-7
:. ? ..O-:!4!7
:
-5
O. 4
.25
......
"
:0'
-6
log lambda1
(c)
ranGACV
.'
O. 7 O. 9
-3
...0,28
.. ????????r .....
r,,6
~,
-4
-6
-5
log lambda
(b)
0
.. ..
.
.: 0'F5 0'F8 0.[32
0'.2,4
-7
-4
0:\13
':
-6
log lambda1
(d)
-4
-5
o
C\I
(0
c:i
(0
c:i
.~
=...,.
.0
ca O
12!
.
.0
o
e
a..
co
10
C\I
c:i
c:i
(0
o
10
c:i ~--------.-------.--------r--~
0,56
0,58
0,60
ranGACV
(e)
0,62
c:i
~
100
__~____, -____. -__- .____. -__~
150
200
250
300
Cholesterol (mg/dL)
(f)
350
400
Figure 1: (a) and (b): Single smoothing parameter comparison of ranGACV and CK L .
(c) and (d): Two smoothing parameter comparison of ranGACV and CK L. (e): Comparison of ranG ACV and U B R. (f): Probability estimate from Beaver Dam Study
| 1483 |@word polynomial:1 replicate:2 logit:1 open:2 simulation:5 pressure:1 tr:7 ld:1 initial:1 liu:2 series:3 unx:1 efficacy:1 selecting:2 inefficiency:2 com:1 fn:5 additive:2 ronald:1 numerical:3 chicago:2 girosi:2 designed:1 plot:4 v:1 d5i:1 intelligence:1 selected:1 beaver:3 ith:1 compo:1 provides:2 attack:1 sigmoidal:1 simpler:1 along:1 h4:1 become:2 fitting:1 wuf:3 alspector:1 brain:2 td:3 provided:1 begin:1 campus:1 fti:1 mass:2 null:1 what:1 kind:1 ti:15 xd:2 exactly:1 fat:1 returning:1 control:1 unit:1 appear:4 yn:2 positive:1 walnut:1 modify:1 approximately:1 yd:3 might:2 black:2 twice:1 studied:1 suggests:1 co:1 limited:1 practical:1 testing:1 borderline:1 implement:1 differs:1 definite:2 x3:8 epidemiological:1 procedure:3 dayton:1 weather:2 convenient:2 confidence:1 radial:1 refers:1 doubled:1 get:1 close:1 selection:1 risk:5 influence:1 applying:1 dam:3 wong:2 www:1 go:1 duration:1 convex:1 cholesterol:4 population:1 target:2 exact:4 us:1 continues:1 preprint:1 wang:4 principled:2 disease:1 pd:1 govern:2 complexity:2 depend:1 solving:2 rewrite:1 asa:1 technically:1 basis:4 gacv:8 gu:5 joint:1 various:2 rxl:1 derivation:1 stacked:1 univ:1 epi:1 acv:10 artificial:1 tell:1 choosing:2 outcome:3 quite:1 larger:1 statistic:3 mg:1 propose:2 fr:1 ckl:8 cacv:3 cluster:3 produce:2 comparative:1 depending:1 gong:2 stat:2 measured:1 minor:1 sim:1 sa:3 implemented:2 lambda1:2 attribute:3 f4:1 observational:1 generalization:4 wall:1 preliminary:1 drinking:1 substituting:1 a2:3 favorable:1 estimation:2 proc:1 applicable:1 tik:1 bond:1 grouped:1 wl:2 cary:1 successfully:1 mit:1 fya:2 gaussian:4 ck:5 rather:1 derived:1 likelihood:8 baseline:1 summarizing:1 el:1 minimizers:2 typically:1 lj:1 classification:6 smoothing:17 special:1 having:2 x4:3 represents:3 pai:1 jones:2 mimic:1 future:1 report:2 minimized:1 spline:14 simplex:1 employ:1 randomly:1 tradeofjand:2 replaced:1 hemoglobin:1 consisting:1 interest:1 fd:2 highly:1 hiwii:2 evaluation:1 replicates:2 tj:1 accurate:1 partial:1 poggio:2 biometrics:1 tree:1 filled:3 taylor:2 desired:2 circle:5 fitted:1 uncertain:1 soft:6 ar:1 ordinary:1 cost:1 entry:1 uniform:1 gcv:2 johnson:1 perturbed:2 synthetic:1 grkpack:6 cited:1 randomized:13 siam:1 bu:1 dong:1 physician:1 physic:1 moody:1 ophthalmalogy:1 again:2 woman:1 f5:1 lambda:2 conf:1 derivative:1 li:2 north:1 int:1 inc:1 matter:1 depends:1 vi:1 trh:1 minimize:1 square:2 variance:10 bayesian:2 produced:1 rangacv:16 drive:1 liebler:2 history:1 simultaneous:1 ed:2 pp:3 knowledge:1 hilbert:1 actually:1 cik:1 cbms:1 appears:1 manuscript:1 x6:3 evaluated:1 box:2 done:1 arch:1 until:1 sketch:1 hand:1 ubr:2 replacing:2 nonlinear:2 hwl:2 aj:1 contain:1 true:4 unbiased:2 brown:1 regularization:3 read:1 trlist:2 sin:2 x5:3 percentile:1 criterion:1 generalized:2 confusion:1 tn:1 variational:3 wise:1 ef:3 fi:1 functional:3 physical:1 insensitive:1 association:3 numerically:1 ai:2 cv:4 tuning:3 mathematics:1 aq:1 retinopathy:2 had:1 multivariate:1 showed:1 optimizing:2 commun:1 barbara:1 tesauro:1 binary:1 meeting:1 yi:13 seen:2 minimum:1 morgan:1 ii:5 multiple:3 corporate:1 hormone:1 technical:2 calculation:5 cross:9 raphson:3 lin:1 prediction:3 regression:2 patient:1 essentially:1 iteration:2 kernel:2 interval:1 median:1 biasvariance:1 leaving:2 regional:1 subject:1 thing:1 member:1 cowan:1 spirit:1 near:1 abnormality:1 easy:1 concerned:1 variety:1 fit:10 variate:1 architecture:2 wahba:25 bandwidth:1 rang:7 reduce:1 arbib:1 cn:1 tradeoff:7 br:2 penalty:4 hessian:3 clear:1 nonparametric:1 extensively:1 statist:3 iith:2 nsf:1 dotted:1 utans:2 estimated:1 wr:1 klein:9 demonstrating:1 blood:1 wisc:2 changing:1 anova:5 f8:1 sum:1 glycosylated:1 inverse:2 eli:1 you:1 family:2 draw:1 f13:1 fl:4 hi:1 tackled:1 quadratic:4 x2:7 anyone:1 span:1 uf:2 department:2 structured:1 according:3 remain:1 wi:4 modification:1 dv:1 heart:1 taken:1 computationally:1 equation:1 previously:1 know:1 letting:5 available:1 progression:1 hii:2 original:1 running:1 include:2 clustering:1 remaining:1 madison:3 newton:3 giving:1 uj:1 yz:1 society:1 quantity:1 fa:2 md:2 grace:1 diagonal:1 distance:4 street:3 fy:3 reason:1 assuming:1 systolic:1 index:2 minimizing:1 nc:1 equivalently:1 sinica:2 trace:4 negative:2 unknown:2 observation:2 thence:1 y1:2 reproducing:2 rating:1 pair:1 kl:2 extensive:2 qa:1 kauffman:1 including:1 reliable:1 unpenalized:1 indicator:2 hr:1 representing:1 eye:3 moss:1 l992:1 review:1 xiang:6 wisconsin:3 asymptotic:1 loss:1 suggestion:1 generator:1 validation:8 at1:1 age:3 degree:1 proxy:1 pi:5 elsewhere:2 penalized:8 free:1 ovum:1 bias:7 institute:2 taking:1 author:1 adaptive:1 replicated:1 functionals:1 approximate:5 kullback:2 anew:1 handbook:1 don:1 continuous:1 iterative:2 diabetic:2 ca:1 expansion:1 cl:1 statistica:2 ciq:1 withstanding:1 repeated:1 girard:2 body:2 fig:2 representative:1 tl:1 ff:10 cubic:3 position:1 pv:3 explicit:1 downhill:1 exponential:2 xl:6 weighting:1 hw:1 survival:1 dl:1 exists:1 sequential:1 relearning:1 l993:1 gao:1 ua1:1 minimizer:10 satisfies:2 truth:2 ydi:3 ann:2 replace:2 change:1 included:1 infinite:1 except:1 typical:1 lemma:1 called:2 experimental:1 mark:2 searched:1 xiwu:2 dept:3 |
529 | 1,484 | The Bias-Variance Tradeoff and the Randomized
GACV
Grace Wahba, Xiwu Lin and Fangyu Gao
Dong Xiang
Dept of Statistics
Univ of Wisconsin
1210 W Dayton Street
Madison, WI 53706
wahba,xiwu,fgao@stat.wisc.edu
SAS Institute, Inc.
SAS Campus Drive
Cary, NC 27513
sasdxx@unx.sas.com
Ronald Klein, MD and Barbara Klein, MD
Dept of Ophthalmalogy
610 North Walnut Street
Madison, WI 53706
kleinr,kleinb@epi.ophth.wisc.edu
Abstract
We propose a new in-sample cross validation based method (randomized
GACV) for choosing smoothing or bandwidth parameters that govern the
bias-variance or fit-complexity tradeoff in 'soft' classification. Soft classification refers to a learning procedure which estimates the probability
that an example with a given attribute vector is in class 1 vs class O. The
target for optimizing the the tradeoff is the Kullback-Liebler distance
between the estimated probability distribution and the 'true' probability distribution, representing knowledge of an infinite population. The
method uses a randomized estimate of the trace of a Hessian and mimics
cross validation at the cost of a single relearning with perturbed outcome
data.
1 INTRODUCTION
We propose and test a new in-sample cross-validation based method for optimizing the biasvariance tradeoff in 'soft classification' (Wahba et al1994), called ranG ACV (randomized
Generalized Approximate Cross Validation) . Summarizing from Wahba et al(l994) we are
given a training set consisting of n examples, where for each example we have a vector
t E T of attribute values, and an outcome y, which is either 0 or 1. Based on the training
data it is desired to estimate the probability p of the outcome 1 for any new examples in the
The Bias-Variance TradeofJand the Randomized GACV
621
future. In 'soft' classification the estimate p(t) of p(t) is of particular interest, and might be
used by a physician to tell patients how they might modify their risk p by changing (some
component of) t, for example, cholesterol as a risk factor for heart attack. Penalized likelihood estimates are obtained for p by assuming that the logit f(t), t E T, which satisfies
p(t) = ef(t) 1(1 + ef(t?) is in some space 1{ of functions . Technically 1{ is a reproducing
kernel Hilbert space, but you don't need to know what that is to read on. Let the training
set be {Yi, ti, i = 1,???, n}. Letting Ii = f(td, the negative log likelihood .c{Yi, ti, fd of
the observations, given f is
n
.c{Yi, ti, fd = 2::[-Ydi + b(li)],
(1)
i=1
where b(f) = log(l + ef ). The penalized likelihood estimate of the function f is the
solution to: Find f E 1{ to minimize h. (I):
n
h.(f) = 2::[-Ydi + b(ld) + J>.(I),
(2)
i =1
where 1>.(1) is a quadratic penalty functional depending on parameter(s) A = (AI, ... , Aq)
which govern the so called bias-variance tradeoff. Equivalently the components of A control the tradeoff between the complexity of f and the fit to the training data. In this paper we
sketch the derivation of the ranG ACV method for choosing A, and present some preliminary but favorable simulation results, demonstrating its efficacy. This method is designed
for use with penalized likelihood estimates, but it is clear that it can be used with a variety
of other methods which contain bias-variance parameters to be chosen, and for which minimizing the Kullback-Liebler (K L) distance is the target. In the work of which this is a part,
we are concerned with A having multiple components. Thus, it will be highly convenient
to have an in-sample method for selecting A, if one that is accurate and computationally
convenient can be found.
Let P>. be the the estimate and p be the 'true' but unknown probability function and let
= p(td,p>.i = p>.(ti). For in-sample tuning, our criteria for a good choice of A is
Pi
the KL distance KL(p,p>.) = ~ E~I[PilogP7. + (1- pdlogg~::?)]. We may replace
K L(p,p>.) by the comparative K L distance (C K L), which differs from K L by a quantity
which does not depend on A. Letting hi = h (ti), the C K L is given by
1
CKL(p,p>.) == CKL(A) = ;;,
n
2:: [-pd>'i + b(l>.i)).
(3)
i=)
C K L(A) depends on the unknown p, and it is desired is to have a good estimate or proxy
for it, which can then be minimized with respect to A.
It is known (Wong 1992) that no exact unbiased estimate of CK L(A) exists in this case, so
that only approximate methods are possible. A number of authors have tackled this problem, including Utans and M90dy(1993), Liu(l993), Gu(1992). The iterative U BR method
of Gu(l992) is included in GRKPACK (Wang 1997), which implements general smoothing spline ANOVA penalized likelihood estimates with multiple smoothing parameters. It
has been successfully used in a number of practical problems, see, for example, Wahba
et al (1994,1995). The present work represents an approach in the spirit of GRKPACK
but which employs several approximations, and may be used with any data set, no matter
how large, provided that an algorithm for solving the penalized likelihood equations, either
exactly or approximately, can be implemented.
622
2
G. Wahba et al.
THE GACV ESTIMATE
In the general penalized likelihood problem the minimizer 1>,(-) of (2) has a representation
M
1>.(t)
=L
n
dv<Pv(t)
v=l
+
L CiQ>.(ti, t)
(4)
i=l
where the <Pv span the null space of 1>" Q>.(8, t) is a reproducing kernel (positive definite
function) for the penalized part of 7-1., and C = (Cl' ... , Cn )' satisfies M linear conditions,
so that there are (at most) n free parameters in 1>.. Typically the unpenalized functions
<Pv are low degree polynomials. Examples of Q(ti,') include radial basis functions and
various kinds of splines; minor modifications include sigmoidal basis functions, tree basis
functions and so on. See, for example Wahba( 1990, 1995), Girosi, Jones and Poggio( 1995).
If f>.C) is of the form (4) then 1>,(f>.) is a quadratic form in c. Substituting (4) into (2)
results in h a convex functional in C and d, and C and d are obtained numerically via a
Newton Raphson iteration, subject to the conditions on c. For large n, the second sum on
the right of (4) may be replaced by L~=1 Cik Q>. (ti k , t), where the tik are chosen via one
of several principled methods.
To obtain the CACV we begin with the ordinary leaving-out-one cross validation function
CV(.\) for the CKL:
n
[-i]
CV (.\) -_ -1 ""
LJ-yd>.i
n
+ b( 1>.i) ] ,
(5)
i=1
where fl- i ] the solution to the variational problem of (2) with the ith data point left out
and fti] is the value of fl- i] at ti . Although f>.C) is computed by solving for C and d
the CACV is derived in terms of the values (it"", fn)' of f at the ti. Where there is
no confusion between functions f(-) and vectors (it, ... ,fn)' of values of fat tl, ... ,tn,
we let f = (it, ... " fn)'. For any f(-) of the form (4), J>. (f) also has a representation as
a non-negative definite quadratic form in (it, ... , fn)'. Letting L:>. be twice the matrix of
this quadratic form we can rewrite (2) as
1
n
h(f,Y)
= L[-Ydi + b(/i)] + 2f'L:>.f.
(6)
i=1
Let W = W(f) be the n x n diagonal matrix with (/ii == Pi(l - Pi) in the iith position.
Using the fact that (/ii is the second derivative of b(fi), we have that H = [W + L:>.] - 1
is the inverse Hessian of the variational problem (6). In Xiang and Wahba (1996), several
Taylor series approximations, along with a generalization of the leaving-out-one lemma
(see Wahba 1990) are applied to (5) to obtain an approximate cross validation function
ACV(.\), which is a second order approximation to CV(.\) . Letting hii be the iith entry
of H , the result is
CV(.\)
~ ACV('\) = .!. t[-Yd>.i + b(f>.i)] + .!. t
n
n
i= l
i=1
hiiYi(Yi - P>.i) .
[1 - hiwii]
Then the GACV is obtained from the ACV by replacing h ii by ~ L~1 hii
and replacing 1 - hiWii by ~tr[I - (Wl/2 HWl/2)], giving
CACV('\)
1~
]
tr(H)
== ~tr(H)
L~l Yi(Yi - P>.i)
(Wl/2HWl /2)] ,
= ;; t;;[-Yd>.i + b(1).i) + -n-tr[I _
(7)
(8)
where W is evaluated at 1>.. Numerical results based on an exact calculation of (8) appear
in Xiang and Wahba (1996). The exact calculation is limited to small n however.
623
The Bias-Variance TradeofJand the Randomized GACV
3 THE RANDOMIZED GACV ESTIMATE
Given any 'black box' which, given >., and a training set {Yi, ti} produces f>. (.) as the minimizer of (2), and thence f>. = (fA 1 , " ' , f>.n)', we can produce randomized estimates of
trH and tr[! - W 1 / 2 HW 1/ 2 J without having any explicit calculations of these matrices.
This is done by running the 'black box' on perturbed data {Vi + <5i , td. For the Yi Gaussian, randomized trace estimates of the Hessian of the variational problem (the 'influence
matrix') have been studied extensively and shown to be essentially as good as exact calculations for large n, see for example Girard(1998) . Randomized trace estimates are based
on the fact that if A is any square matrix and <5 is a zero mean random n-vector with independent components with variance
then E<5' A<5 = ~ tr A. See Gong et al( 1998) and
(TJ,
u"
references cited there for experimental results with multiple regularization parameters. Returning to the 0-1 data case, it is easy to see that the minimizer fA(') of 1;.. is continuous in
Y, not withstanding the fact that in our training set the Yi take on only values 0 or 1. Letting
if = UA1,"', f>.n)' be the minimizer of (6) given y = (Y1,"', Yn)', and if+O be the
minimizer given data y+<5 = (Y1 +<51, ... ,Yn +<5n )' (the ti remain fixed), Xiang and Wahba
(1997) show, again using Taylor series expansions, that if+O - ff ,. . ., [WUf) + ~AJ-1<5.
This suggests that ~<5'Uf+O - ff) provides an estimate oftr[W(ff) + ~At1. However,
u"
if we take the solution ff to the nonlinear system for the original data Y as the initial value
for a Newton-Raphson calculation of ff+O things become even simpler. Applying a one
step Newton-Raphson iteration gives
(9)
-<5 + PjfUf,Y) = -<5, and [:;~f(ff,Y + <5)J - 1
y + o ,l
y + o ,l 82 h(fY
8 2 h(fY
[ 8?7if
A' Y)J- 1 ,we have f A
- fYA + [8?7if
A' Y)J- 1 uJ: so that f A
- fYA
[WUf) + E At 1<5. The result is the following ranGACV function:
Since Pjf(ff,y + <5)
ranGACV(>.) =
=
n
.!. ~[n~
'I '+bU .)J+
Yz At
At
<5' (f Y+O,l
A
n
-
fY)
A
",n
(
)
wi=l Yi Yi - PAi
.
[<5'<5 - <5'WUf)Uf+O,l - ff)J
(10)
To reduce the variance in the term after the '+' in (10), we may draw R
independent replicate vectors <51,'" ,<5R , and replace the term after the '+' in
( 1O)b
1... ",R
o:(fr Hr . 1 -ff)
wr=l
n
ranGACV(>.) function.
4
y
R
2:7-1y.(y.-P>..)
[O~Or-O~ W(fn(f~+Ar . l-ff)1
to obtain an R-replicated
NUMERICAL RESULTS
In this section we present simulation results which are representative of more extensive
simulations to appear elsewhere. In each case, K < < n was chosen by a sequential clustering algorithm. In that case, the ti were grouped into K clusters and one member of each
cluster selected at random. The model is fit. Then the number of clusters is doubled and the
model is fit again. This procedure continues until the fit does not change. In the randomized
trace estimates the random variates were Gaussian. Penalty functionals were (multivariate
generalizations of) the cubic spline penalty functional>. fa1U" (X))2, and smoothing spline
ANOVA models were fit.
G. Wahba et at.
624
4.1
EXPERIMENT 1. SINGLE SMOOTHING PARAMETER
In this experiment t E [0,1], f(t) = 2sin(10t), ti = (i - .5)/500, i = 1,???,500. A
random number generator produced 'observations' Yi = 1 with probability Pi = el , /(1 +
eli), to get the training set. QA is given in Wahba( 1990) for this cubic spline case, K = 50.
Since the true P is known, the true CKL can be computed. Fig. l(a) gives a plot of
CK L(A) and 10 replicates of ranGACV(A). In each replicate R was taken as 1, and J
was generated anew as a Gaussian random vector with (115 = .001. Extensive simulations
with different (115 showed that the results were insensitive to (115 from 1.0 to 10- 6 ? The
minimizer of C K L is at the filled-in circle and the 10 minimizers of the 10 replicates of
ranGACV are the open circles. Anyone of these 10 provides a rather good estimate of
the A that goes with the filled-in circle. Fig. l(b) gives the same experiment, except that
this time R = 5. It can be seen that the minimizers ranGACV become even more reliable
estimates of the minimizer of C K L, and the C K L at all of the ranG ACV estimates are
actually quite close to its minimum value.
4.2
EXPERIMENT 2. ADDITIVE MODEL WITH A = (Al' A2)
Here t E [0,1] 0 [0,1]. n = 500 values of ti were generated randomly according to
a uniform distribution on the unit square and the Yi were generated according to Pi =
e li j(l + e l ,) with t = (Xl,X2) and f(t) = 5 sin 27rXl - 3sin27rX2. An additive model
as a special case of the smoothing spline ANOVA model (see Wahba et al, 1995), of the
form f(t) = /-l + h(xd + h(X2) with cubic spline penalties on hand h were used.
K = 50, (115 = .001, R = 5. Figure l(c) gives a plot of CK L(Al' A2) and Figure l(d)
gives a plot of ranGACV(Al, A2). The open circles mark the minimizer of ranGACV
in both plots and the filled in circle marks the minimizer of C K L. The inefficiency, as
measured by CKL()..)/minACKL(A) is 1.01. Inefficiencies near 1 are typical of our
other similar simulations.
4.3
EXPERIMENT 3. COMPARISON OF ranGACV AND UBR
This experiment used a model similar to the model fit by GRKPACK for the risk of
progression of diabetic retinopathy given t = (Xl, X2, X3) = (duration, glycosylated
hemoglobin, body mass index) in Wahba et al(l995) as 'truth'. A training set of 669
examples was generated according to that model, which had the structure f(Xl, X2, X3) =
/-l + fl (xd + h (X2) + h (X3) + fl,3 (Xl, X3). This (synthetic) training set was fit by GRKPACK and also using K = 50 basis functions with ranG ACV. Here there are P = 6
smoothing parameters (there are 3 smoothing parameters in f13) and the ranGACV function was searched by a downhill simplex method to find its minimizer. Since the 'truth' is
known, the CKL for)" and for the GRKPACK fit using the iterative UBR method were
computed. This was repeated 100 times, and the 100 pairs of C K L values appears in Figure l(e). It can be seen that the U BR and ranGACV give similar C K L values about 90%
of the time, while the ranG ACV has lower C K L for most of the remaining cases.
4.4
DATA ANALYSIS: AN APPLICATION
Figure 1(f) represents part of the results of a study of association at baseline of pigmentary
abnormalities with various risk factors in 2585 women between the ages of 43 and 86 in the
Beaver Dam Eye Study, R. Klein et al( 1995). The attributes are: Xl = age, X2 =body mass
index, X3 = systolic blood pressure, X4 = cholesterol. X5 and X6 are indicator variables for
taking hormones, and history of drinking. The smoothing spline ANOVA model fitted was
f(t) = /-l+dlXl +d2X2+ h(X3)+ f4(X4)+ h4(X3, x4)+d5I(x5) +d6I(x6), where I is the
indicator function. Figure l(e) represents a cross section of the fit for X5 = no, X6 = no,
The Bias- Variance Tradeoff and the Randomized GACV
625
X2, X3 fixed at their medians and Xl fixed at the 75th percentile. The dotted lines are the
Bayesian confidence intervals, see Wahba et al( 1995). There is a suggestion of a borderline
inverse association of cholesterol. The reason for this association is uncertain. More details
will appear elsewhere.
Principled soft classification procedures can now be implemented in much larger data sets
than previously possible, and the ranG ACV should be applicable in general learning.
References
Girard, D. (1998), 'Asymptotic comparison of (partial) cross-validation, GCV and randomized GCV in nonparametric regression', Ann. Statist. 126, 315-334.
Girosi, F., Jones, M. & Poggio, T. (1995), 'Regularization theory and neural networks
architectures', Neural Computatioll 7,219-269.
Gong, J., Wahba, G., Johnson, D. & Tribbia, J. (1998), 'Adaptive tuning of numerical
weather prediction models: simultaneous estimation of weighting, smoothing and physical
parameters', MOllthly Weather Review 125, 210-231.
Gu, C. (1992), 'Penalized likelihood regression: a Bayesian analysis', Statistica Sinica
2,255-264.
Klein, R., Klein, B. & Moss, S. (1995), 'Age-related eye disease and survival. the Beaver
Dam Eye Study', Arch Ophthalmol113, 1995.
Liu, Y. (1993), Unbiased estimate of generalization error and model selection in neural
network, manuscript, Department of Physics, Institute of Brain and Neural Systems, Brown
University.
Utans, J. & Moody, J. (1993), Selecting neural network architectures via the prediction
risk: application to corporate bond rating prediction, in 'Proc. First Int'I Conf. on Artificial
Intelligence Applications on Wall Street', IEEE Computer Society Press.
Wahba, G. (1990), Spline Models for Observational Data, SIAM. CBMS-NSF Regional
Conference Series in Applied Mathematics, v. 59.
Wahba, G. (1995), Generalization and regularization in nonlinear learning systems, in
M. Arbib, ed., 'Handbook of Brain Theory and Neural Networks', MIT Press, pp. 426430.
Wahba, G., Wang, Y., Gu, c., Klein, R. & Klein, B. (1994), Structured machine learning
for 'soft' classification with smoothing spline ANOVA and stacked tuning, testing and
evaluation, in J. Cowan, G. Tesauro & J. Alspector, eds, 'Advances in Neural Information
Processing Systems 6', Morgan Kauffman, pp. 415-422.
Wahba, G., Wang, Y., Gu, C., Klein, R. & Klein, B. (1995), 'Smoothing spline AN OVA for
exponential families, with application to the Wisconsin Epidemiological Study of Diabetic
Retinopathy' , Ann. Statist. 23, 1865-1895.
Wang, Y. (1997), 'GRKPACK: Fitting smoothing spline analysis of variance models to data
from exponential families', Commun. Statist. Sim. Compo 26,765-782.
Wong, W. (1992), Estimation of the loss of an estimate, Technical Report 356, Dept. of
Statistics, University of Chicago, Chicago, II.
Xiang, D. & Wahba, G. (1996), 'A generalized approximate cross validation for smoothing
splines with non-Gaussian data', Statistica Sinica 6, 675-692, preprint TR 930 available
via www. stat. wise. edu/-wahba - > TRLIST.
Xiang, D. & Wahba, G. (1997), Approximate smoothing spline methods for large data sets
in the binary case, Technical Report 982, Department of Statistics, University of Wisconsin,
Madison WI. To appear in the Proceedings of the 1997 ASA Joint Statistical Meetings,
Biometrics Section, pp 94-98 (1998). Also in TRLIST as above.
G, Wahba et aI,
626
CKL
ranGACV
10
(0
c:i
(0
c:i
0
o
c:i
c:i
(0
10
10
c:i
CKL
10
(0
10
10
.' .
c:i
0
o
0
c:i
10
~
-8
-7
-6
-5
log lambda
(a)
-3
-4
9.29
-8
"f
\~7
0\
-7
-7
:. ? ..O-:!4!7
:
-5
O. 4
.25
......
"
:0'
-6
log lambda1
(c)
ranGACV
.'
O. 7 O. 9
-3
...0,28
.. ????????r .....
r,,6
~,
-4
-6
-5
log lambda
(b)
0
.. ..
.
.: 0'F5 0'F8 0.[32
0'.2,4
-7
-4
0:\13
':
-6
log lambda1
(d)
-4
-5
o
C\I
(0
c:i
(0
c:i
.~
=...,.
.0
ca O
12!
.
.0
o
e
a..
co
10
C\I
c:i
c:i
(0
o
10
c:i ~--------.-------.--------r--~
0,56
0,58
0,60
ranGACV
(e)
0,62
c:i
~
100
__~____, -____. -__- .____. -__~
150
200
250
300
Cholesterol (mg/dL)
(f)
350
400
Figure 1: (a) and (b): Single smoothing parameter comparison of ranGACV and CK L .
(c) and (d): Two smoothing parameter comparison of ranGACV and CK L. (e): Comparison of ranG ACV and U B R. (f): Probability estimate from Beaver Dam Study
Graph Matching for Shape Retrieval
Benoit Huet, Andrew D.J. Cross and Edwin R. Hancock'
Department of Computer Science, University of York
York, YOI 5DD, UK
Abstract
This paper describes a Bayesian graph matching algorithm for
data-mining from large structural data-bases. The matching algorithm uses edge-consistency and node attribute similarity to determine the a posteriori probability of a query graph for each of the
candidate matches in the data-base. The node feature-vectors are
constructed by computing normalised histograms of pairwise geometric attributes. Attribute similarity is assessed by computing
the Bhattacharyya distance between the histograms. Recognition
is realised by selecting the candidate from the data-base which has
the largest a posteriori probability. We illustrate the recognition
technique on a data-base containing 2500 line patterns extracted
from real-world imagery. Here the recognition technique is shown
to significantly outperform a number of algorithm alternatives.
1
Introduction
Since Barrow and Popplestone [1] first suggested that relational structures could be
used to represent and interpret 2D scenes, there has been considerable interest in the
machine vision literature in developing practical graph-matching algorithms [8, 3,
10]. The main computational issues are how to compare relational descriptions when
there is significant structural corruption [8, 10] and how to search for the best match
[3]. Despite resulting in significant improvements in the available methodology for
graph-matching, there has been little progress in applying the resulting algorithms
to large-scale object recognition problems. Most of the algorithms developed in the
literature are evaluated for the relatively simple problem of matching a model-graph
against a scene known to contain the relevant structure. A more realistic problem is
that of taking a large number (maybe thousands) of scenes and retrieving the ones
that best match the model. Although this problem is key to data-mining from large
libraries of visual information, it has invariably been approached using low-level
feature comparison techniques. Very little effort [7,4] has been devoted to matching
? corresponding author erh@cs.york.ac.uk
897
Graph Matching for Shape Retrieval
higher-level structural primitives such as lines, curves or regions. Moreover, because
of the perceived fragility of the graph matching process, there has been even less
effort directed at attempting to retrieve shapes using relational information.
Here we aim to fill this gap in the literature by using graph-matching as a means
of retrieving the shape from a large data-based that most closely resembles a query
shape. Although the indexation images in large data-bases is a problem of current
topicality in the computer vision literature [5, 6, 9], the work presented in this
paper is more ambitious. Firstly, we adopt a structural abstraction of the shape
recognition problem and match using attributed relational graphs. Each shape in
our data-base is a pattern of line-segments. The structural abstraction is a nearest
neighbour graph for the centre-points of the line-segments. In addition, we exploit
attribute information for the line patterns. Here the geometric arrangement of the
line-segments is encapsulated using a histogram of Euclidean invariant pairwise (binary) attributes. For each line-segment in turn we construct a normalised histogram
of relative angle and length with the remaining line-segments in the pattern. These
histograms capture the global geometric context of each line-segment. Moreover,
we interpret the pairwise geometric histograms as measurement densities for the
line-segments which we compare using the Bhattacharyya distance.
Once we have established the pattern representation, we realise object recognition
using a Bayesian graph-matching algorithm. This is a two-step process. Firstly,
we establish correspondence matches between the individual tokens in the query
pattern and each of the patterns in the data-base. The correspondences matches
are sought so as to maximise the a posteriori measurement probability. Once the
MAP correspondence matches have been established, then the second step in our
recognition architecture involves selecting the line-pattern from the data-base which
has maximum matching probability.
2
MAP Framework
Formally our recognition problem is posed as follows. Each ARG in the database is
a triple, G = (Vc, Ec, Ac), where Vc is the set of vertices (nodes), Ec is the edge
set (Ec C Vc x Vc), and Ac is the set of node attributes. In our experimental
example, the nodes represent line-structures segmented from 2D images. The edges
are established by computing the N-nearest neighbour graph for the line-centres.
Each node j E Vc is characterised by a vector of attributes, ~j and hence Ac =
{~j jj E Vc}. In the work reported here the attribute-vector is represents the
contents of a normalised pairwise attribute histogram .
The data-base of line-patterns is represented by the set of ARG's D = {G}. The
goal is to retrieve from the data-base D, the individual ARG that most closely
resembles a query pattern Q = (VQ' EQ, AQ). We pose the retrieval process as one
of associating with the query the graph from the data-base that has the largest a
posteriori probability. In other words, the class identity of the graph which most
closely corresponds to the query is
wQ =
arg max P(G' IQ)
C'EV
However, since we wish to make a detailed structural comparison of the graphs,
rather than comparing their overall statistical properties, we must first establish
a set of best-match correspondences between each ARG in the data-base and the
query Q. The set of correspondences between the query Q and the ARG G is
a relation fc : Vc f-7 VQ over the vertex sets of the two graphs. The mapping
function consists of a set of Cartesian pairings between the nodes of the two graphs,
B. Huet, A. D. 1. Cross and E. R. Hancock
898
i.e. Ie = {(a,a);a E Ve,a E VQ} ~ Ve x VQ . Although this may appear to be a
brute force method, it must be stressed that we view this process of correspondence
matching as the final step in the filtering of the line-patterns. We provide more
details of practical implementation in the experimental section of this paper.
With the correspondences to hand we can re-state our maximum a posteriori probability recognition objective as a two step process. For each graph G in turn, we
locate the maximum a posteriori probability mapping function Ie onto the query
Q. The second step is to perform recognition by selecting the graph whose mapping
function results in the largest matching probability. These two steps are succinctly
captured by the following statement of the recognition condition
wQ
= arg max max P(fe,IG', Q)
e'ED
la'
This global MAP condition is developed into a useful local update formula by applying the Bayes formula to the a posteriori matching probability. The simplification
is as follows
PU IG Q)
e ,
= p(Ae, AQl/e)P(felVe, Ee, VQ, EQ)P(Ve , Ee)P(VQ, EQ)
P(G)P(Q)
The terms on the right-hand side of the Bayes formula convey the following
meaning. The conditional measurement density p(Ae,AQl/e) models the measurement similarity of the node-sets of the two graphs. The conditional probability P(feIEe, EQ) models the structural similarity of the two graphs under
the current set of correspondence matches. The assumptions used in developing our simplification of the a posteriori matching probability are as follows.
Firstly, we assume that the joint measurements are conditionally independent
of the structure of the two graphs provided that the set of correspondences is
known, i.e. P(Ae, AQl/e, Ee, Ve, E Q, VQ) = P(Ae, AQl/e). Secondly, we assume that there is conditional independence of the two graphs in the absence of
correspondences. In other words, P(Ve, Ee, VQ, EQ) = P(VQ, EQ)P(Ve, Ee) and
P(G, Q) = P(G)P(Q). Finally, the graph priors P(Ve, Ee) , P(VQ, EQ) P(G) and
P( Q) are taken as uniform and are eliminated from the decision making process.
To continue our development, we first focus on the conditional measurement density,
p(Ae, AQl/e) which models the process of comparing attribute similarity on the
nodes of the two graphs. Assuming statistical independence of node attributes, the
conditional measurement density p( Ae, AQ lie), can be factorised over the Cartesian
pairs (a, a) E Ve x VQ which constitute the the correspondence match Ie in the
following manner
p(Ae, AQl/e)
=
II
P(~a' ~ol/e(a)
= a)
(a,o)E/a
As a result the correspondence matches may be optimised using a simple node-bynode discrete relaxation procedure. The rule for updating the match assigned to
the node a of the graph G is
le(a) = arg
max
oEVQU{4>}
P(~a'~o)l/(a) = a)P(feIEe,EQ)
In order to model the structural consistency of the set of assigned matches,we turn
to the framework recently reported by Finch, Wilson and Hancock [2}. This work
provides a framework for computing graph-matching energies using the weighted
Hamming distance between matched cliques. Since we are dealing with a large-scale
object recognition system, we would like to minimise the computational overheads
associated with establishing correspondence matches. For this reason, rather than
899
Graph Matchingfor Shape Retrieval
working with graph neighbourhoods or cliques, we chose to work with the relational
units of the smallest practical size. In other words we satisfy ourself with measuring
consistency at the edge level. For edge-units, the structural matching probability
P(fa!Va, Ea, VQ, EQ) is computed from the formula
(a,b)EEG (Ct ,(J )EEQ
where Pe is the probability of an error appearing on one of the edges of the matched
structure. The Sa,Ct are assignment variables which are used to represent the current
state of match and convey the following meaning
Sa
3
Ct
,
= {I
0
if fa (a) = a
otherwise
Histogram-based consistency
We now furnish some details of the shape retrieval task used in our experimental
evaluation of the recognition method. In particular, we focus on the problem of
recognising 2D line patterns in a manner which is invariant to rotation, translation
and scale. The raw information available for each line segment are its orientation
(angle with respect to the horizontal axis) and its length (see figure 1). To illustrate
how the Euclidean invariant pairwise feature attributes are computed, suppose that
we denote the line segments associated with the nodes indexed a and b by the
vectors Ya and Yb respectively. The vectors are directed away from their point of
intersection. The pairwise relative angle attribute is given by
(Ja ,b
= arccos
[I:: 1?1::1]
From the relative angle we compute the directed relative angle. This involves giving
d
~:.~~:
---------c:----;-~:---
b-!
~------.
o..b
---------------.
D;b
Figure 1: Geometry for shape representation
the relative angle a positive sign if the direction of the angle from the baseline Ya to
its partner Yb is clockwise and a negative sign if it is counter-clockwise. This allows
us to extend the range of angles describing pairs of segments from [0,7I"J to [-7I",7I"J.
The directed relative position {}a,b is represented by the normalised length ratio
between the oriented baseline vector Ya and the vector yl joining the end (b) of the
baseline segment (ab) to the intersection of the segment pair (cd).
{}a,b
=
1
D
l+~
2
Dab
B. Huet, A. D. 1. Cross and E. R. Hancock
900
The physical range of this attribute is (0, IJ. A relative position of 0 indicates that
the two segments are parallel, while a relative position of 1 indicates that the two
segments intersect at the middle point of the baseline.
The Euclidean invariant angle and position attributes 8a,b and {)a ,b are binned in a
histogram. Suppose that Sa(J-l, v) = {(a , b)18a,b E All 1\ {)a,b E Rv 1\ bE VD} is the
set of nodes whose pairwise geometric attributes with the node a are spanned by
the range of directed relative angles All and the relative position attribute range
Rv. The contents of the histogram bin spanning the two attribute ranges is given
by Ha(J-l, v) = ISa(J-l, v)l. Each histogram contains nA relative angle bins and nR
length ratio bins. The normalised geometric histogram bin-entries are computed as
follows
Ha(J-l, v)
ha(J-l, v) = "nA "nR H (
)
~Il'=l ~v'=l
a J-l, v
The probability of match between the pattern-vectors is computed using the Bhattacharyya distance between the normalised histograms.
P(f(a) = al~a' ~a) =
I:~~l I:~~l ha(J-l, v)ha(J-l, v)
L j'EQ I:nA
I:nR
h (
)h (
)
Il'=l
v'=l a J-l, V
a J-l, V
= exp[-Ba ,aJ
With this modelling ingredient , the condition for recognition is
WQ
4
= arg~~%
L
L
(a , b}EE~
(a,iJ}EEQ
{-Ba,a-Bb,iJ+ln(I-Pe)Sa,aSb,iJ+lnPe(I-Sa,aSb,/3)}
Experiments
The aim in this section is to evaluate the graph-based recognition scheme on a database of real-world line-patterns. We have conducted our recognition experiments
with a data-base of 2500 line-patterns each containing over a hundred lines. The
line-patterns have been obtained by applying line/edge detection algorithms to the
raw grey-scale images followed by polygonisation. For each line-pattern in the database, we construct the six-nearest neighbour graph . The feature extraction process
together with other details of the data used in our study are described in recent
papers where we have focussed on the issues of histogram representation [4J and the
optimal choice of the relational structure for the purposes of recognition. In order to
prune the set of line-patterns for detailed graph-matching we select about 10% of the
data-base using a two-step process. This consists of first refining the data-base using
a global histogram of pairwise attributes [4J . The top quartile of matches selected
in this way are then further refined using a variant of the Haussdorff distance to
select the set of pairwise attributes that best match against the query.
The recognition task is posed as one of recovering the line-pattern which most closely
resembles a digital map . The original images from which our line-patterns have been
obtained are from a number of diverse sources. However , a subset of the images are
aerial infra-red line-scan views of southern England. Two of these infra-red images
correspond to different views of the area covered by the digital map. These views
are obtained when the line-scan device is flying at different altitudes. The line-scan
device used to obtain the aerial images introduces severe barrel distortions and
hence the map and aerial images are not simply related via a Euclidean or affine
transformation. The remaining line-patterns in the data-base have been extracted
from trademarks and logos. It is important to stress that although the raw images
are obtained from different sources, there is nothing salient about their associated
line-pattern representations that allows us to distinguish them from one-another.
Graph Matchingfor Shape Retrieval
(a) Digital Map
901
(b) Target 1
(c) Target 2
Figure 2: Images from the data-base
Moreover, since it is derived from a digital map rather than one of the images in
the data-base, the query is not identical to any of the line-patterns in the model
library.
We aim to assess the importance of different attributes representation on the retrieval process. To this end, we compare node-based and the histogram-based attribute representation. \Ve also consider the effect of taking the relative angle and
relative position attributes both singly and in tandem. The final aspect of the
comparison is to consider the effects of using the attributes purely for initialisation
purposes and also in a persistent way during the iteration of the matching process.
To this end we consider the following variants of our algorithm .
? Non-Persistent Attributes: Here we ignore the attribute information
provided by the node-histograms after the first iteration and attempt to
maximise the structural congruence of the graphs .
? Local attributes: Here we use only the single node attributes rather than
an attribute histogram to model the a posteriori matching probabilities.
Graph Matching Strategy
ReI. Position Attribute iInitialisation only)
ReI. Angle Attribute (Initialisation only)
ReI. Angle + Position Attributes (Initialisation only)
1D ReI. Position Histogram (Initialisation only)
1D ReI. Angle Histogram (Initialisation only)
2D Histogram (Initialisation only)
ReI. Position Attribute (Persistent)
ReI. Angle Attribute (Persistent)
ReI. Angle + Position Attributes (Persistent)
1D ReI. Position Histogram (Persistent)
1D ReI. Angle Histogram (Persistent)
2D Histogram (Persistent)
Retrieval
Accuracy
39%
45%
58%
42%
59%
68%
63%
89%
98%
66%
92%
100%
Iterations
per recall
5.2
4.75
4.27
4.7
4.2
3.9
3.96
3.59
3.31
3.46
3.23
3.12
Table 1: Recognition performance of various recognition strategies averaged over
26 queries in a database of 260 line-patterns
In Table 1 we present the recognition performance for each of the recognition strategies in turn. The table lists the recall performance together with the average number
B. Huet, A. D. 1. Cross and E. R. Hancock
902
of iterations per recall for each of the recognition strategies in turn. The main features to note from this table are as follows . Firstly, the iterative recall using the full
histogram representation outperforms each of the remaining recognition methods
in terms of both accuracy and computational overheads. Secondly, it is interesting
to compare the effect of using the histogram in the initialisation-only and iteration
persistent modes. In the latter case the recall performance is some 32% better than
in the former case. In the non-persistent mode the best recognition accuracy that
can be obtained is 68%. Moreover, the recall is typically achieved in only 3.12 iterations as opposed to 3.9 (average over 26 queries on a database of 260 images) .
Finally, the histogram representation provides better performance, and more significantly, much faster recall than the single-attribute similarity measure. When the
attributes are used singly, rather than in tandem , then it is the relative angle that
appears to be the most powerful.
5
Conclusions
We have presented a practical graph-matching algorithm for data-mining in large
structural libraries. The main conclusion to be drawn from this study is that the
combined use of structural and histogram information improves both recognition
performance and recall speed. There are a number of ways in which the ideas
presented in this paper can be extended. Firstly, we intend to explore more a perceptually meaningful representation of the line patterns, using grouping principals
derived from Gestalt psychology. Secondly, we are exploring the possibility of formulating the filtering of line-patterns prior to graph matching using Bayes decision
trees.
References
[1] H. Barrow and R. Popplestone. Relational descriptions in picture processing.
Machine Intelligence, 5:377- 396, 1971.
[2] A. Finch, R. Wilson, and E. Hancock. Softening discrete relaxation . Advances
in NIPS 9, Edited by M. Mozer, M. Jordan and T. Petsche, MIT Press, pages
438- 444, 1997.
[3] S. Gold and A. Rangarajan. A graduated assignment algorithm for graph
matching. IEEE PAMI, 18:377- 388, 1996.
[4] B. Huet and E. Hancock. Relational histograms for shape indexing. IEEE
ICC V, pages 563- 569, 1998.
[5] W. Niblack et al.. The QBIC project: Querying images by content using color ,
texture and shape. Image and Vision Storage and Retrieval, 173- 187, 1993.
[6] A. P. Pentland, R. W. Picard, and S. Scarloff. Photobook: tools for contentbased manipulation of image databases. Storage and Retrieval for Image and
Video Database II, pages 34- 47, February 1994.
[7] K. Sengupta and K. Boyer. Organising large structural databases. IEEE PAMI,
17(4):321- 332,1995.
[8] 1. Shapiro and R. Haralick. A metric for comparing relational descriptions.
IEEE PAMI, 7(1):90- 94, 1985.
[9] M. Swain and D. Ballard. Color indexing. International Journal of Computer
Vision, 7(1) :11- 32, 1991.
[10] R. Wilson and E. R. Hancock. Structural matching by discrete relaxation.
IEEE PAMI, 19(6):634- 648, June 1997.
| 1484 |@word middle:1 polynomial:1 replicate:2 logit:1 open:2 grey:1 simulation:5 pressure:1 tr:7 ld:1 initial:1 liu:2 series:3 unx:1 efficacy:1 selecting:5 inefficiency:2 contains:1 initialisation:7 bhattacharyya:3 outperforms:1 current:3 com:1 comparing:3 must:2 ronald:1 fn:5 numerical:3 additive:2 chicago:2 girosi:2 shape:13 realistic:1 designed:1 plot:4 update:1 v:1 d5i:1 intelligence:2 selected:2 device:2 beaver:3 ith:1 compo:1 provides:4 node:18 organising:1 attack:1 sigmoidal:1 simpler:1 firstly:5 along:1 h4:1 constructed:1 become:2 retrieving:2 pairing:1 consists:2 persistent:10 fitting:1 overhead:2 wuf:3 manner:2 pairwise:9 alspector:1 brain:2 ol:1 td:3 little:2 tandem:2 provided:3 begin:1 campus:1 fti:1 moreover:4 mass:2 matched:2 null:1 what:1 barrel:1 kind:1 developed:2 transformation:1 ti:15 xd:2 exactly:1 fat:1 returning:1 uk:2 control:1 unit:3 brute:1 appear:5 yn:2 positive:2 maximise:2 walnut:1 local:2 modify:1 despite:1 joining:1 establishing:1 optimised:1 approximately:1 yd:3 might:2 black:2 twice:1 topicality:1 studied:1 resembles:3 chose:1 suggests:1 logo:1 co:1 limited:1 range:5 averaged:1 popplestone:2 practical:5 directed:5 testing:1 borderline:1 implement:1 differs:1 definite:2 x3:8 epidemiological:1 procedure:4 dayton:1 intersect:1 area:1 significantly:2 weather:2 convenient:2 matching:26 confidence:1 radial:1 refers:1 word:3 doubled:1 get:1 close:1 selection:1 onto:1 storage:2 risk:5 influence:1 applying:4 dam:3 wong:2 www:1 context:1 map:8 go:1 primitive:1 duration:1 convex:1 rule:1 cholesterol:4 fill:1 spanned:1 retrieve:2 population:1 target:4 suppose:2 exact:4 us:2 recognition:26 updating:1 continues:1 database:8 preprint:1 wang:4 capture:1 thousand:1 region:1 counter:1 edited:1 principled:2 disease:1 pd:1 govern:2 complexity:2 mozer:1 depend:1 solving:2 rewrite:1 segment:14 asa:1 technically:1 flying:1 purely:1 basis:4 gacv:8 gu:5 edwin:1 joint:2 various:3 rxl:1 represented:2 derivation:1 stacked:1 univ:1 epi:1 acv:10 hancock:8 artificial:1 query:13 tell:1 approached:1 choosing:2 outcome:3 refined:1 quite:1 whose:2 larger:1 posed:2 distortion:1 otherwise:1 dab:1 statistic:3 final:2 mg:1 aql:6 propose:2 fr:1 ckl:8 relevant:1 gold:1 description:3 cacv:3 cluster:3 rangarajan:1 produce:2 comparative:1 object:3 depending:1 andrew:1 illustrate:2 stat:2 gong:2 pose:1 nearest:3 ac:4 measured:1 ij:4 sim:1 eq:10 minor:1 implemented:2 c:1 lambda1:2 involves:2 recovering:1 progress:1 sa:8 direction:1 rei:10 closely:4 attribute:40 f4:1 quartile:1 vc:7 observational:1 bin:4 ja:1 generalization:4 wall:1 preliminary:1 secondly:3 exploring:1 drinking:1 asb:2 exp:1 congruence:1 mapping:3 substituting:1 sought:1 adopt:1 a2:3 smallest:1 purpose:2 perceived:1 encapsulated:1 favorable:1 proc:1 applicable:1 tik:1 bond:1 estimation:2 grouped:1 wl:2 largest:3 cary:1 successfully:1 tool:1 weighted:1 mit:2 fya:2 gaussian:4 indexation:1 aim:3 ck:5 rather:6 wilson:3 derived:3 focus:2 refining:1 haralick:1 improvement:1 june:1 modelling:1 likelihood:8 indicates:2 baseline:5 summarizing:1 posteriori:9 abstraction:2 el:1 minimizers:2 typically:2 lj:1 relation:1 boyer:1 issue:2 classification:6 orientation:1 overall:1 arg:9 development:1 arccos:1 smoothing:17 special:1 sengupta:1 construct:2 once:2 having:2 extraction:1 eliminated:1 x4:3 represents:4 pai:1 jones:2 identical:1 mimic:1 future:1 minimized:1 spline:14 simplex:1 report:2 employ:1 infra:2 randomly:1 neighbour:3 oriented:1 ve:9 individual:2 tradeofjand:2 replaced:1 hemoglobin:1 consisting:1 geometry:1 ab:1 attempt:1 detection:1 invariably:1 interest:2 fd:2 highly:1 mining:3 possibility:1 picard:1 hiwii:2 evaluation:2 replicates:2 benoit:1 introduces:1 severe:1 tj:1 devoted:1 accurate:1 edge:7 partial:1 poggio:2 biometrics:1 tree:2 filled:3 taylor:2 euclidean:4 indexed:1 desired:2 circle:5 re:1 fitted:1 uncertain:1 soft:6 ar:1 measuring:1 assignment:2 ordinary:1 cost:1 vertex:2 entry:2 subset:1 uniform:2 hundred:1 swain:1 gcv:2 johnson:1 conducted:1 reported:2 perturbed:2 synthetic:1 finch:2 grkpack:6 combined:1 cited:1 density:4 randomized:13 siam:1 ie:3 international:1 bu:1 dong:1 physician:1 physic:1 yl:1 together:2 moody:1 na:3 ophthalmalogy:1 again:2 imagery:1 containing:2 opposed:1 woman:1 f5:1 lambda:2 conf:1 derivative:1 li:2 factorised:1 north:1 int:1 inc:1 matter:1 satisfy:1 depends:1 vi:1 view:4 realised:1 red:2 bayes:3 parallel:1 trh:1 minimize:1 square:2 il:2 ass:1 accuracy:3 variance:10 correspond:1 bayesian:4 raw:3 produced:1 rangacv:16 drive:1 corruption:1 liebler:2 history:1 huet:5 simultaneous:1 ed:3 against:2 realise:1 energy:1 pp:3 associated:3 attributed:1 hamming:1 recall:8 knowledge:1 color:2 improves:1 hilbert:1 actually:1 cik:1 cbms:1 appears:2 manuscript:1 ea:1 higher:1 x6:3 methodology:1 yb:2 evaluated:2 box:2 done:1 arch:1 until:1 sketch:1 hand:3 working:1 ubr:2 replacing:2 horizontal:1 nonlinear:2 hwl:2 mode:2 aj:2 effect:3 contain:2 true:4 unbiased:2 brown:1 former:1 regularization:3 hence:2 assigned:2 read:1 furnish:1 trlist:2 contentbased:1 conditionally:1 sin:2 x5:3 during:1 percentile:1 criterion:1 generalized:2 stress:1 confusion:1 tn:1 pami:4 image:16 variational:3 wise:1 ef:3 fi:1 meaning:2 recently:1 rotation:1 functional:3 physical:2 insensitive:1 association:3 extend:1 numerically:1 interpret:2 significant:2 measurement:7 ai:2 cv:4 iq:1 tuning:3 consistency:4 mathematics:1 centre:2 softening:1 aq:3 retinopathy:2 had:1 similarity:6 base:18 pu:1 multivariate:1 showed:1 recent:1 optimizing:2 commun:1 barbara:1 tesauro:1 manipulation:1 binary:2 continue:1 meeting:1 yi:13 seen:2 minimum:1 morgan:1 captured:1 prune:1 determine:1 clockwise:2 ii:7 rv:2 full:1 multiple:3 corporate:1 isa:1 hormone:1 technical:2 match:18 segmented:1 calculation:5 cross:13 raphson:3 lin:1 retrieval:10 england:1 faster:1 dept:3 va:1 yoi:1 prediction:3 variant:2 regression:2 ae:7 patient:1 metric:1 essentially:1 vision:4 iteration:8 kernel:2 histogram:29 represent:3 achieved:1 addition:1 interval:1 median:1 biasvariance:1 leaving:2 source:2 regional:1 subject:1 thing:1 member:1 cowan:1 spirit:1 jordan:1 structural:14 near:1 ee:7 abnormality:1 easy:1 concerned:1 variety:1 independence:2 fit:10 variate:1 psychology:1 architecture:3 wahba:25 bandwidth:1 rang:7 reduce:1 arbib:1 cn:1 idea:1 tradeoff:7 br:2 associating:1 minimise:1 six:1 eeq:2 effort:2 penalty:4 hessian:3 york:3 jj:1 constitute:1 useful:1 clear:1 detailed:2 covered:1 maybe:1 singly:2 nonparametric:1 extensively:1 statist:3 iith:2 shapiro:1 outperform:1 nsf:1 dotted:1 utans:2 sign:2 estimated:1 wr:1 per:2 klein:9 diverse:1 discrete:3 key:1 salient:1 demonstrating:1 blood:1 drawn:1 wisc:2 changing:1 anova:5 f8:1 graph:39 relaxation:3 sum:1 glycosylated:1 inverse:2 eli:1 you:1 angle:19 powerful:1 family:2 draw:1 decision:2 f13:1 fl:4 hi:1 ct:3 followed:1 simplification:2 tackled:1 correspondence:13 distinguish:1 quadratic:4 binned:1 x2:7 scene:3 aspect:1 speed:1 anyone:1 span:1 formulating:1 attempting:1 relatively:1 uf:2 department:3 structured:1 according:3 developing:2 project:1 aerial:3 remain:1 describes:1 wi:4 modification:1 making:1 dv:1 invariant:4 indexing:2 altitude:1 heart:1 taken:2 computationally:1 equation:1 vq:12 previously:1 ln:1 turn:5 describing:1 know:1 letting:5 end:3 available:3 progression:1 away:1 petsche:1 hii:2 appearing:1 neighbourhood:1 alternative:1 original:2 top:1 running:1 include:2 clustering:1 remaining:4 madison:3 newton:3 exploit:1 giving:2 uj:1 yz:1 establish:2 society:1 february:1 objective:1 intend:1 arrangement:1 quantity:1 fa:4 strategy:4 md:2 grace:1 diagonal:1 nr:3 southern:1 distance:9 street:3 vd:1 ourself:1 partner:1 fy:3 reason:2 spanning:1 assuming:2 systolic:1 length:4 index:2 ratio:2 minimizing:1 nc:1 equivalently:1 sinica:2 fe:1 statement:1 trace:4 negative:3 ba:2 implementation:1 ambitious:1 unknown:2 perform:1 observation:2 thence:1 barrow:2 pentland:1 relational:9 extended:1 y1:2 locate:1 reproducing:2 rating:1 pair:4 kl:2 extensive:2 established:3 nip:1 qa:1 suggested:1 pattern:26 kauffman:1 ev:1 including:1 reliable:1 max:4 unpenalized:1 niblack:1 video:1 force:1 indicator:2 hr:1 representing:1 scheme:1 eye:3 library:3 picture:1 axis:1 moss:1 l992:1 review:1 geometric:6 literature:4 prior:2 icc:1 xiang:6 wisconsin:3 asymptotic:1 loss:1 relative:14 suggestion:1 interesting:1 filtering:2 querying:1 erh:1 ingredient:1 generator:1 validation:8 at1:1 age:3 triple:1 degree:1 digital:4 affine:1 proxy:1 dd:1 pi:5 cd:1 translation:1 qbic:1 elsewhere:2 penalized:8 token:1 succinctly:1 free:1 ovum:1 bias:7 normalised:6 side:1 institute:2 taking:3 focussed:1 curve:1 world:2 author:2 adaptive:1 replicated:1 ig:2 ec:3 gestalt:1 functionals:1 bb:1 approximate:5 ignore:1 kullback:2 clique:2 dealing:1 anew:1 global:3 handbook:1 don:1 continuous:1 iterative:3 search:1 diabetic:2 table:4 ballard:1 fragility:1 ca:1 eeg:1 expansion:1 cl:1 statistica:2 ciq:1 main:3 withstanding:1 nothing:1 repeated:1 girard:2 convey:2 body:2 fig:2 representative:1 tl:1 ff:10 cubic:3 position:13 pv:3 explicit:1 downhill:1 exponential:2 xl:6 candidate:2 wish:1 lie:1 pe:2 weighting:1 hw:1 formula:4 trademark:1 list:1 survival:1 dl:1 exists:1 grouping:1 recognising:1 sequential:1 importance:1 texture:1 perceptually:1 cartesian:2 relearning:1 gap:1 intersection:2 l993:1 fc:1 simply:1 explore:1 gao:1 visual:1 ua1:1 corresponds:1 minimizer:10 satisfies:2 truth:2 extracted:2 conditional:5 ydi:3 goal:1 identity:1 graduated:1 ann:2 replace:2 absence:1 considerable:1 change:1 content:3 included:1 infinite:1 except:1 typical:1 characterised:1 lemma:1 principal:1 called:2 experimental:4 la:1 ya:3 meaningful:1 formally:1 select:2 wq:3 mark:2 searched:1 latter:1 stressed:1 assessed:1 scan:3 photobook:1 xiwu:2 evaluate:1 |
530 | 1,485 | Dynamically Adapting Kernels in Support
Vector Machines
N ello Cristianini
Dept. of Engineering Mathematics
University of Bristol, UK
nello.cristianini@bristol.ac.uk
Colin Campbell
Dept. of Engineering Mathematics
University of Bristol, UK
c.campbell@bristol.ac.uk
John Shawe-Taylor
Dept. of Computer Science
Royal Holloway College
john@dcs.rhbnc.ac.uk
Abstract
The kernel-parameter is one of the few tunable parameters in Support Vector machines, controlling the complexity of the resulting
hypothesis. Its choice amounts to model selection and its value is
usually found by means of a validation set. We present an algorithm which can automatically perform model selection with little
additional computational cost and with no need of a validation set .
In this procedure model selection and learning are not separate,
but kernels are dynamically adjusted during the learning process
to find the kernel parameter which provides the best possible upper
bound on the generalisation error. Theoretical results motivating
the approach and experimental results confirming its validity are
presented.
1
Introduction
Support Vector Machines (SVMs) are learning systems designed to automatically
trade-off accuracy and complexity by minimizing an upper bound on the generalisation error provided by VC theory. In practice, however, SVMs still have a few
tunable parameters which need to be determined in order to achieve the right balance and the values of these are usually found by means of a validation set. One
of the most important of these is the kernel-parameter which implicitly defines the
structure of the high dimensional feature space where the maximal margin hyperplane is found. Too rich a feature space would cause the system to overfit the data,
Dynamically Adapting Kernels in Support Vector Machines
205
and conversely the system can be unable to separate the data if the kernels are too
poor. Capacity control can therefore be performed by tuning the kernel parameter
subject to the margin being maximized. For noisy datasets, yet another quantity
needs to be set, namely the soft-margin parameter C.
SVMs therefore display a remarkable dimensionality reduction for model selection.
Systems such as neural networks need many different architectures to be tested and
decision trees are faced with a similar problem during the pruning phase. On the
other hand SVMs can shift from one model complexity to another by simply tuning
a continuous parameter.
Generally, model selection by SVMs is still performed in the standard way: by
learning different SVMs and testing them on a validation set in order to determine
the optimal value of the kernel-parameter. This is expensive in terms of computing
time and training data. In this paper we propose a different scheme which dynamically adjusts the kernel-parameter to explore the space of possible models at
little additional computational cost compared to fixed-kernel learning. Futhermore
this approach only makes use of training-set information so it is more efficient in a
sample complexity sense.
Before proposing the model selection procedure we first prove a theoretical result,
namely that the margin and structural risk minimization (SRM) bound on the generalization error depend smoothly on the kernel parameter. This can be exploited
by an algorithm which keeps the system close to maximal margin while the kernel
parameter is changed smoothly. During this phase, the theoretical bound given by
SRM theory can be computed. The best kernel-parameter is the one which gives the
lowest possible bound. In section 4 we present experimental results showing that
model selection can be efficiently performed using the proposed method (though we
only consider Gaussian kernels in the simulations outlined).
2
Support Vector Learning
The decision function implemented by SV machines can be written as:
f(x) = sign
(L
Yiai K(x, Xi) -
B)
tESV
where the ai are obtained by maximising the following Lagrangian (where m is the
number of patterns):
m
L
=L
i= l
m
ai - 1/2 L
aiajYiyjK(Xi, Xj)
i,j= l
with respect to the ai, subject to the constraints
m
LaiYi =
a
i=l
and where the functions K( x, x') are called kernels. The kernels provide an expression for dot- products in a high-dimensional feature space [1]:
K( x, x') = (<I> (x) , <I>(x' ))
206
N. Cristianini, C. Campbell and 1. Shawe-Taylor
and also implicitly define the nonlinear mapping <1>( x) of the training data into
feature space where they may be separated using the maximal margin hyperplane.
A number of choices of kernel-function can be made e.g. Gaussians kernels:
K(x, x') = e-ll x-x'1 12/2(T2
The following upper bound can be proven from VC theory for the generalisation
error using hyperplanes in feature space [7, 9J:
where R is the radius of the smallest ball containing the training set, m the number
of training points and 'Y the margin (d. [2J for a complete survey of the generalization properties of SV machines) .
The Lagrange multipliers Qi are usually found by means of a Quadratic Programming optimization routine, while the kernel-parameters are found using a validation
set. As illustrated in Figure 1 there is a minimum of the generalisation error for
that value of the kernel-parameter which has the best trade-off between overfitting
and ability to find an efficient solution.
0 13
012
0 11
01
009
008
0 07
0 06
005
0 04
2
10
Figure 1: Generalization error (y-axis) as a function of (J (x-axis) for the mirror symmetry problem (for Gaussian kernels with zero training error and maximal margin,
m = 200, n = 30 and averaged over 105 examples).
3
A utomatic Model Order Selection
We now prove a theorem which shows that the margin of the optimal hyperplane is a
smooth function of the kernel parameter, as is the upper bound on the generalisation
error. First we state the Implicit Function Theorem.
Implicit Function Theorem [10]: Let F(x, y) be a continuously differentiable
function,
F : U ~ ~ x V ~ ~p --t ~
and let (a, b) E U x V be a solution to the equation F(x, y) = O. Let the partial
derivatives matrix mi ,j = (~:;) w.r.t. y be full rank at (a, b) . Then, near (a, b),
207
Dynamically Adapting Kernels in Support Vector Machines
there exists one and only one function
function is continuous.
y = g(x) such that F(x,g(x)) = 0, and such
Theorem: The margin, of SV machines depends smoothly on the kernel parameter
a.
Proof: Consider the function 9 : ~ <;;; ~ --t A <;;; ~P, 9 : a ~ (aO, A) which given the
data maps the choice of a to the optimal parameters aO and lagrange parameter A
of the SV machine with Kernel matrix Gij = YiYjK(a; Xi, Xj )). Let
p
Wu(a) = l:ai - 1/2 l:aiajYiyjK(a; Xi,Xj)
i,j
i=l
+ A(l: Yiai )
be the functional that the SV machine maximizes. Fix a value of a and let aO(a) be
the corresponding solution of Wu(a). Let I be the set of indices for which aj((J) =1= O.
We may assume that the submatrix of G indexed by I is non-singular since otherwise
the maximal margin hyperplane could be expressed in terms of a subset of indices.
Now choose a maximal set of indices J containing I such that the corresponding
su bmatrix of G is non-singular and all of the points indexed by J have margin 1.
Now consider the function F((J ,a , A)i = (a~)j; ,i 2: 1, F((J,a,A)o = LjYjaj in
the neighbourhood of (J, where ji is an enumeration of the elements of J,
oWu
"'_
oa. = 1 - Yj L.. aiYiK((J; Xi, Xj)
J
+ AYj
.
t
and satisfies the equation F((J, aO((J), A(a)) = 0 at the extremal points of Wu(a) .
Then the SV function is the implicit function, (aO, A) = g((J), and is continuous
(and unique) iff F is continuously differentiable and the partial derivatives matrix
w.r.t. a, A is full rank. But the partial derivatives matrix H is given by
Hij
OP
= oat
= Yj;Yj}K((J;xj; ,Xj})
= Hji,i , j
2: 1,
JJ
for ji,iJ E J, which was non-degenerate by definition of J, while
Hoo
=
oFo
OA
=0
and
HOj
Consider any non-zero a satisfying
(a, Af H(a, A)
oFo
= oajJ = n =
Lj ajYJ
oFj
OA
.
= Hjo,J
2: 1.
= 0, and any A. We have
= aTGa + 2AaT Y = aTGa > O.
Hence, the matrix H is non-singular for a satisfying the given linear constraint.
Hence , by the implicit function theorem 9 is a continuous function of (J. The
following is proven in [2J:
,2 = (t Zif) -1
t=l
which shows that, is a continuous function of (J. As the radius of the ball containing
the points is also a continuous function of (J , and the generalization error bound has
the form f. ~ CR(a)2llaO((J)lll for some constant C, we have the following corollary.
Corollary: The bound on the generalization error is smooth in (J.
This means that, when the margin is optimal, small variations in the kernel parameter will produce small variations in the margin (and in the bound on the
generalisation error). Thus ,u ~ ,uHu and after updating the (J, the system will
N Cristianini, C. Campbell and J. Shawe- Taylor
208
still be in a sub-optimal position. This suggests the following strategy for Gaussian
kernels, for instance:
Kernel Selection Procedure
l. Initialize u to a very small value
2. Maximize the margin, then
? Compute the SRM bound (or observe the validation error)
? Increase the kernel parameter: u +- u + 8u
3. Stop when a predetermined value of u is reached else repeat step 2.
This procedure takes advantage of the fact that for very small (J convergence is
generally very rapid (overfi tting the data, of course), and that once the system is
near the equilibrium, few iterations will always be sufficient to move it back to the
maximal margin situation. In other words, this system is brought to a maximal
margin state in the beginning, when this is computationally very cheap, and then it
is actively kept in that situation by continuously adjusting the a while the kernelparameter is gradually increased.
In the next section we will experimentally investigate this procedure for real-life
datasets. In the numerical simulations we have used the Kernel-Adatron (KA)
algorithm recently developed by two of the authors [4] which can be used to train SV
machines. We have chosen this algorithm because it can be regarded as a gradient
ascent procedure for maximising the Kuhn-Tucker Lagrangian L . Thus the ai for
a sub-optimal state are close to those for the optimum and so little computational
effort will be needed to bring the system back to a maximal margin position:
The Kernel-Adatron Algorithm.
l. exi = l.
2. FOR i
= 1 TO m
? ,i
=
? 8ex i
? IF
YiZi
= 17(1 _ , i )
(ex i + 8ex i ) ::; 0 THEN
ex i
= 0 ELSE ex i
+-
ex i
+ 8ex
t .
? margin = ~ (min(z;) -max(z;))
(4 (z;) = positively (negatively) labelled patterns)
3. IF(margin
4
= 1) THEN
stop, ELSE go to step 2.
Experimental Results
In this section we implement the above algorithm for real-life datasets and plot the
upper bound given by VC theory and the generalization error as functions of (J. In
order to compute the bound, E ::; R 2/m,2 we need to estimate the radius of the ball
in feature space. In general his can be done explicitly by maximising the following
Lagrangian w.r.t. Ai using convex quadratic programming routines:
L
=L
subject to the constraints
AiK(Xi, Xi) -
L AiAjK(Xi,
Xj)
i,j
2:i Ai = 1 and Ai 2: O.
The radius is then found from [3]:
Dynamically Adapting Kernels in Support Vector Machines
i,j
209
i,j
However, we can also get an upper bound for this quantity by noting that Gaussian
kernels always map training points to the surface of a sphere of radius 1 centered on
the origin of the feature space. This can be easily seen by noting that the distance
of a point from the origin is its norm:
11<I>(x)11
= J(<I>(X),<I>(X)) = JK(x,x) = Jellx-xll/2o-2 = 1
In Figure 2 we give both these bounds (the upper bound is Li adm) and generalisation error (on a test set) for two standard datasets: the aspect-angle dependent
sonar classification dataset of Gorman and Sejnowski [5] and the Wisconsin breast
cancer dataset [8]. As we see from these plots there is little need for the additional computational cost of determining R from the above quadratic progamming
problem, at least for Gaussian kernels. In Fig. 3 we plot the bound Li adm and
generalisation error for 2 figures from a United States Postal Service dataset of
handwritten digits [6]. In these, and other instances we have investigated, the minimum of the bound approximately coincides with the minimum of the generalisation
error. This gives a good criterion for the most suitable choice for a. Furthermore,
this estimate for the best a is derived solely from training data without the need
for an additional validation set .
02
., ..
Figure 2: Generalisation error (solid curves) for the sonar classification (left Fig.)
and Wisconsin breast cancer datasets (right Fig.). The upper curves (dotted) show
the upper bounds from VC theory (for the top curves R=l).
Starting with a small a-value we have observed that the margin can be maximised
rapidly. Furthermore, the margin remains close to 1 if a is incremented by a small
amount. Consequently, we can study the performance of the system by traversing
a range of a-values, alternately incrementing a then maximising the margin using
the previous optimal set of a-values as a starting point. We have found that this
procedure does not add a significant computational cost in general. For example,
for the sonar classification dataset mentioned above and starting at a = 0.1 with
increments ~a = 0.1 it took 186 iterations to reach a = 1.0 and 4895 to reach
a = 2.0 as against 110 and 2624 iterations for learning at both these a-values. For
a rough doubling of the learning time it is possible to determine a reasonable value
for a for good generalisation without use of a validation set.
210
N Cristianini, C Campbell and J Shawe-Taylor
.. ......
"
".
O.
-",
'.'.
'.
o.
\
\.
07
07
\'"
O.
'.
O.
-'
\.
\
\\.,
0'
,
02
0
0
\
"- '.
0'
10
0
~
0
12
Figure 3: Generalisation error (solid curve) and upper bound from VC theory
(dashed curve with R=l) for digits 0 and 3 from the USPS dataset of handwritten
digits.
5
Conclusion
We have presented an algorithm which automatically learns the kernel parameter
with little additional cost, both in a computational and sample-complexity sense.
Model selection takes place during the learning process itself, and experimental
results are provided showing that this strategy provides a good estimate of the
correct model complexity.
References
[1] Aizerman, M., Braverman, E ., and Rozonoer , L. (1964) . Theoretical Foundations of
the Potential Function Method in Pattern Recognition Learning, A utomations and
Remote Control, 25:821-837.
[2] Bartlett P., Shawe-Taylor J ., (1998). Generalization Performance of Support Vector
Machines and Other Pattern Classifiers. 'Advances in Kernel Methods - Support Vector
Learning', Bernhard Sch61kopf, Christopher J . C. Burges, and Alexander J . Smola
(eds.), MIT Press, Cambridge, USA.
[3] Burges c., (1998). A tutorial on support vector machines for pattern recognition . Data
Mining and Knowledge Discovery, 2:l.
[4] Friess T., Cristianini N., Campbell C. , (1998) The Kernel-Adatron Algorithm: a Fast
and Simple Learning Procedure for Support Vector Machines, in Shavlik, J. , ed., Machine Learning: Proceedings of the Fifteenth International Conference, Morgan Kaufmann Publishers, San Francisco, CA.
[5] Gorman R.P. & Sejnowski, T.J. (1988) Neural Networks 1:75-89.
[6] LeCun, Y., Jackel, L. D. , Bottou, L., Brunot, A., Cortes, C., Denker, J . S., Drucker, H.,
Guyon, I., Muller, U. A., Sackinger, E ., Simard, P. and Vapnik, V., (1995) . Comparison
of learning algorithms for handwritten digit recognition, International Conference on
Artificial Neural Networks, Fogelman, F. and Gallinari, P. (Ed.), pp. 53-60.
[7] Shawe-Taylor, J ., Bartlett, P., Williamson, R. & Anthony, M. (1996) . Structural Risk
Minimization over Data-Dependent Hierarchies NeuroCOLT Technical Report NCTR-96-053 (ftp://ftp.des .rhbne .ae. uk /pub/neuroeolt/teeh_reports).
[8] Ster, B ., & Dobnikar, A. (1996) Neural networks in medical diagnosis: comparison with
other methods. In A. Bulsari et al. (ed.) Proceedings of the International Conference
EA NN '96, p. 427-430.
[9] Vapnik , V. (1995) The Nature of Statistical Learning Theory, Springer Verlag.
[10] James, Robert C. (1966) Advanced calculus Belmont, Calif. : Wadsworth
| 1485 |@word norm:1 calculus:1 simulation:2 solid:2 reduction:1 united:1 pub:1 ka:1 yet:1 written:1 john:2 belmont:1 numerical:1 predetermined:1 confirming:1 cheap:1 designed:1 plot:3 beginning:1 maximised:1 provides:2 postal:1 hyperplanes:1 ofo:2 prove:2 rapid:1 automatically:3 little:5 enumeration:1 lll:1 provided:2 maximizes:1 lowest:1 developed:1 proposing:1 adatron:3 classifier:1 uk:6 control:2 gallinari:1 medical:1 before:1 service:1 engineering:2 aat:1 solely:1 approximately:1 dynamically:6 conversely:1 suggests:1 range:1 averaged:1 unique:1 lecun:1 testing:1 yj:3 practice:1 implement:1 digit:4 procedure:8 sch61kopf:1 adapting:4 word:1 get:1 close:3 selection:10 risk:2 map:2 lagrangian:3 go:1 starting:3 convex:1 survey:1 adjusts:1 regarded:1 his:1 variation:2 increment:1 rozonoer:1 tting:1 hierarchy:1 controlling:1 aik:1 programming:2 hypothesis:1 origin:2 element:1 expensive:1 satisfying:2 updating:1 jk:1 recognition:3 fries:1 observed:1 remote:1 trade:2 incremented:1 mentioned:1 complexity:6 cristianini:6 depend:1 negatively:1 usps:1 easily:1 exi:1 train:1 separated:1 fast:1 sejnowski:2 artificial:1 otherwise:1 ability:1 noisy:1 itself:1 advantage:1 differentiable:2 took:1 propose:1 maximal:9 product:1 rapidly:1 iff:1 degenerate:1 achieve:1 convergence:1 optimum:1 produce:1 ftp:2 ac:3 ij:1 op:1 implemented:1 kuhn:1 radius:5 correct:1 vc:5 centered:1 ao:5 generalization:7 fix:1 adjusted:1 equilibrium:1 mapping:1 smallest:1 jackel:1 extremal:1 minimization:2 brought:1 rough:1 mit:1 gaussian:5 always:2 cr:1 corollary:2 derived:1 rank:2 sense:2 dependent:2 nn:1 lj:1 fogelman:1 adm:2 classification:3 initialize:1 wadsworth:1 once:1 t2:1 report:1 few:3 phase:2 investigate:1 mining:1 braverman:1 futhermore:1 partial:3 traversing:1 tree:1 indexed:2 taylor:6 calif:1 theoretical:4 instance:2 increased:1 soft:1 cost:5 subset:1 srm:3 too:2 motivating:1 sv:7 international:3 off:2 continuously:3 containing:3 choose:1 derivative:3 simard:1 actively:1 li:2 potential:1 de:1 explicitly:1 depends:1 performed:3 reached:1 accuracy:1 kaufmann:1 efficiently:1 maximized:1 handwritten:3 utomatic:1 bristol:4 reach:2 ed:4 definition:1 against:1 pp:1 tucker:1 james:1 proof:1 mi:1 stop:2 tunable:2 adjusting:1 dataset:5 knowledge:1 dimensionality:1 ster:1 routine:2 ea:1 campbell:6 back:2 done:1 though:1 furthermore:2 implicit:4 smola:1 overfit:1 hand:1 zif:1 christopher:1 su:1 nonlinear:1 sackinger:1 defines:1 aj:1 usa:1 validity:1 multiplier:1 hence:2 illustrated:1 ll:1 during:4 coincides:1 criterion:1 complete:1 bring:1 recently:1 functional:1 ji:2 significant:1 cambridge:1 ai:8 tuning:2 outlined:1 mathematics:2 shawe:6 dot:1 surface:1 add:1 verlag:1 yiai:2 life:2 exploited:1 muller:1 seen:1 minimum:3 additional:5 morgan:1 determine:2 maximize:1 colin:1 dashed:1 full:2 smooth:2 technical:1 af:1 sphere:1 qi:1 breast:2 ae:1 fifteenth:1 iteration:3 kernel:38 else:3 singular:3 publisher:1 ascent:1 subject:3 structural:2 near:2 noting:2 xj:7 architecture:1 drucker:1 shift:1 expression:1 rhbnc:1 bartlett:2 effort:1 cause:1 jj:1 generally:2 amount:2 svms:6 tutorial:1 dotted:1 sign:1 hji:1 diagnosis:1 nctr:1 kept:1 angle:1 place:1 reasonable:1 guyon:1 wu:3 decision:2 submatrix:1 bound:20 display:1 quadratic:3 constraint:3 aspect:1 min:1 ball:3 poor:1 hoo:1 brunot:1 gradually:1 computationally:1 equation:2 remains:1 needed:1 gaussians:1 denker:1 observe:1 neighbourhood:1 top:1 move:1 quantity:2 strategy:2 aiajyiyjk:2 gradient:1 distance:1 separate:2 unable:1 neurocolt:1 capacity:1 oa:3 nello:1 maximising:4 index:3 minimizing:1 balance:1 robert:1 hij:1 xll:1 perform:1 upper:10 datasets:5 situation:2 dc:1 yiyjk:1 namely:2 alternately:1 usually:3 pattern:5 aizerman:1 royal:1 max:1 suitable:1 advanced:1 scheme:1 axis:2 faced:1 discovery:1 determining:1 wisconsin:2 proven:2 remarkable:1 validation:8 foundation:1 ayj:1 sufficient:1 cancer:2 course:1 changed:1 repeat:1 burges:2 shavlik:1 curve:5 rich:1 author:1 made:1 san:1 pruning:1 implicitly:2 bernhard:1 keep:1 overfitting:1 francisco:1 xi:8 continuous:6 sonar:3 nature:1 ca:1 symmetry:1 williamson:1 investigated:1 bottou:1 anthony:1 ofj:1 oat:1 incrementing:1 positively:1 fig:3 sub:2 position:2 learns:1 yizi:1 theorem:5 showing:2 cortes:1 exists:1 vapnik:2 mirror:1 margin:23 gorman:2 smoothly:3 simply:1 explore:1 lagrange:2 expressed:1 doubling:1 springer:1 satisfies:1 ello:1 consequently:1 labelled:1 experimentally:1 generalisation:12 determined:1 hyperplane:4 called:1 gij:1 experimental:4 holloway:1 college:1 support:11 alexander:1 dept:3 tested:1 ex:7 |
531 | 1,486 | Evidence for a Forward Dynamics Model
in Human Adaptive Motor Control
Nikhil Bhushan and Reza Shadmehr
Dept. of Biomedical Engineering
Johns Hopkins University, Baltimore, MD 21205
Email: nbhushan@bme.jhu.edu, reza@bme.jhu.edu
Abstract
Based on computational principles, the concept of an internal
model for adaptive control has been divided into a forward and an
inverse model. However, there is as yet little evidence that learning
control by the eNS is through adaptation of one or the other. Here
we examine two adaptive control architectures, one based only on
the inverse model and other based on a combination of forward and
inverse models. We then show that for reaching movements of the
hand in novel force fields, only the learning of the forward model
results in key characteristics of performance that match the kinematics of human subjects. In contrast, the adaptive control system
that relies only on the inverse model fails to produce the kinematic
patterns observed in the subjects, despite the fact that it is more
stable. Our results provide evidence that learning control of novel
dynamics is via formation of a forward model.
1 Introduction
The concept of an internal model, a system for predicting behavior of a controlled
process, is central to the current theories of motor control (Wolpert et al. 1995) and
learning (Shadmehr and Mussa-Ivaldi 1994). Theoretical studies have proposed
that internal models may be divided into two varieties: forward models, which
simulate the causal flow of a process by predicting its state transition given a motor
command, and inverse models, which estimate motor commands appropriate for a
desired state transition (Miall and Wolpert, 1996). This classification is relevant for
adaptive control because based on computational principles, it has been proposed
that learning control of a nonlinear system might be facilitated if a forward model
of the plant is learned initially, and then during an off-line period is used to train
an inverse model (Jordan and Rumelhart, 1992). While there is no experimental
evidence for this idea in the central nervous system, there is substantial evidence
4
N. Bhushan and R. Shadmehr
that learning control of arm movements involves formation of an internal model.
For example, practicing arm movements while holding a novel dynamical system
initiates an adaptation process which results in the formation of an internal model:
upon sudden removal of the force field, after-effects are observed which match the
expected behavior of a system that has learned to predict and compensate for the
dynamics of the imposed field (Shadmehr and Brashers-Krug, 1997). However, the
computational nature of this internal model, whether it be a forward or an inverse
model, or a combination of both, is not known.
Here we use a computational approach to examine two adaptive control architectures: adaptive inverse model feedforward control and adaptive forward-inverse
model feedback control. We show that the two systems predict different behaviors
when applied to control of arm movements. While adaptation to a force field is
possible with either approach, the second system with feedback control through an
adaptive forward model, is far less stable and is accompanied with distinct kinematic
signatures, termed "near path-discontinuities". We observe remarkably similar instability and near path-discontinuities in the kinematics of 16 subjects that learned
force fields. This is behavioral evidence that learning control of novel dynamics is
accomplished with an adaptive forward model of the system.
2 Adaptive Control using Internal Models
Adaptive control of a nonlinear system which has large sensory feedback delays,
such as the human arm, can be accomplished by using two different internal model
architectures. The first method uses only an adaptive inverse dynamics model to
control the system (Shadmehr and Mussa-Ivaldi, 1994). The adaptive controller
is feedforward in nature and ignores delayed feedback during the movement. The
control system is stable because it relies on the equilibrium properties of the muscle
and the spinal reflexes to correct for any deviations from the desired trajectory.
The second method uses a rapidly adapting forward dynamics model and delayed
sensory feedback in addition to an inverse dynamics model to control arm movements (Miall and Wolpert, 1996). In this case, the corrections to deviations from
the desired trajectory are a result of a combination of supraspinal feedback as well
as spinal/muscular feedback. Since the two methods rely on different internal model
and feedback structures, they are expected to behave differently when the dynamics
of the system are altered.
The Mechanical Model of the Human Arm
For the purpose of simulating arm movements with the two different control architectures, a reasonably accurate model of the human arm is required. We model the
arm as a two joint revolute arm attached to six muscles that act in pairs around
the two joints. The three muscle pairs correspond to elbow joint, shoulder joint
and two joint muscles and are assumed to have constant moment arms. Each muscle is modeled using a Hill parametric model with nonlinear stiffness and viscosity
(Soechting and Flanders, 1997). The dynamics of the muscle can be represented by
a nonlinear state function f M, such that,
(1)
where, Ft is the force developed by the muscle, N is the neural activation to the
muscle, and X m, xm are the muscle length and velocity. The passive dynamics
related to the mechanics of the two-joint revolute arm can be represented by fD,
such that,
x = fD(T, x, x) = D- 1 (x)[T - C(x, x)x + JT Fxl
(2)
Evidence for a Forward Dynamics Model in Human Adaptive Motor Control
5
where, x is the hand acceleration, T is the joint torque generated by the muscles,
x, x are the hand position and velocity, D and C are the inertia and the coriolis
matrices of the arm, J is the Jacobian for hand position and joint angle, and Fx is
the external dynamic interaction force on the hand.
Under the force field environment, the external force Fx acting on the hand is
equal to Bx, where B is a 2x2 rotational viscosity matrix. The effect of the force
field is to push the hand perpendicular to the direction of movement with a force
proportional to the speed of the hand. The overall forward plant dynamics of the
arm is a combination of JM and JD and can be repff~sented by the function Jp ,
(3)
Adaptive Inverse Model Feedforward Control
The first control architecture uses a feedforward controller with only an adaptive
inverse model. The inverse model computes the neural activation to the muscles
for achieving a desired acceleration, velocity and position of the hand. It can be
represented as the estimated inverse, 1;1, of the forward plant dynamics, and maps
the desired position Xd, velocity Xd, and acceleration Xd of the hand, into descending
neural commands N c.
Nc = 1;1 (Xd, Xd, Xd)
(4)
Adaptation to novel external dynamics occurs by learning a new inverse model of the
altered external environment. The error between desired and actual hand trajectory
can be used for training the inverse model. When the inverse model is an exact
inverse of the forward plant dynamics, the gain of the feedforward path is unity and
the arm exactly tracks the desired trajectory. Deviations from the desired trajectory
occur when the inverse model does not exactly model the external dynamics. Under
that situation, the spinal reflex corrects for errors in desired (Xmd, Xmd) and actual
(xm,x m) muscle state, by producing a corrective neural signal NR based on a linear
feedback controller with constants K1 and K 2 ?
(5)
Adaptive Forward-Inverse Model Feedback Control
The second architecture provides feedback control of arm movements in addition
to the feedforward control described above. Delays in feedback cause instability,
therefore, the system relies on a forward model to generate updated state estimates
of the arm. An estimated error in hand trajectory is given by the difference in
desired and estimated state, and can be used by the brain to issue corrective neural
signals to the muscles while a movement is being made. The forward model, written
Inverse Arm
Dynamics Model Td
Desired
Trajectory
6d(t+60)
A?1
to
Muscle
Inverse Muscle
Model f;:.,'
fM
T
Arm
Dynamics
6
to
/
r+----------------~
! '".-...----v-----~
.
".,
l. . ...._. _. . . _. . __._..._!e.. . . . .__ . . . _. ........ .
A=gO ms
6 d (I.30)
Fx
(external force)
+
.
6(1.30)
A=30ms
Figure 1: The adaptive inverse model feedforward control system.
6
N. Bhushan and R. Shadmehr
1\
A=120ms
1\
x, X (t+60)
Desired
Trajectory
Td Inverse Muscle Nc
Model
f--L-..~
f?'
M
A=60 ms +
NR
L........_---'
A~30ms
A-9Oms
Figure 2: A control system that provides feedback control with the use of a forward
and an inverse model.
as jp, mimics the forward dynamics of the plant and predicts hand acceleration
from neural signal Nc, and an estimate of hand state x, ?.
i,
(6)
U sing this equation, one can solve for x, ?at time t, when given the estimated state
at some earlier time t - T, and the descending neural commands N c from time t - T
to t. If t is the current time and T is the time delay in the feedback loop, then sensory
feedback gives the hand state x, x at t-T. The current estimate of the hand position
and velocity can be computed by assuming initial conditions x(t - T)=X(t - T) and
?(t - T)=X(t - T), and then solving Eq. 6. For the simulations, T has value of 200
msec, and is composed of 120 msec feedback delay, 60 msec descending neural path
delay, and 20 msec muscle activation delay.
Based on the current state estimate and the estimated error in trajectory, the desired
acceleration is corrected using a linear feedback controller with constants Kp and
Kv. The inverse model maps the hand acceleration to appropriate neural signal
for the muscles Nc. The spinal reflex provides additional corrective feedback N R ,
when there is an error in the estimated and actual muscle state.
+ Xc = Xd + Kp(Xd - x) + Kv(Xd 1;1 (x new , x, ?)
K 1 (x m - xm) + K 2 (?m d - xm)
Xd
?)
(7)
(8)
(9)
When the forward model is an exact copy of the forward plant dynamics jp= jp, and
the inverse model is correct j;l =1;1, the hand exactly tracks the desired trajectory.
Errors due to an incorrect inverse model are corrected through the feedback loop.
However, errors in the forward model cause deviations from the desired behavior
and instability in the system due to inappropriate feedback action.
3 Simulations results and comparison to human behavior
To test the two control architectures, we compared simulations of arm movements
for the two methods to experimental human results under a novel force field environment. Sixteen human subjects were trained to make rapid point-to-point reaching
Evidence for a Forward Dynamics Model in Human Adaptive Motor Control
(1)
Inverse Model
Feedforward Control
('.::::)(:::?).
...-
.."'-
.,~
."""..
r<:. .~
t>4J(:: ::~
.,../ .).. ?./
!A
c. .... .....v )
i.. .. :;. ..: . . . .,
:.~
_.i .')-'"
(2)
Forward?lnverse Model
Feedback Control
Typical Subject
~.
",A
......
7
?o..
""'....
~
021N\l
~:~
o:lJIffl:J
Q2~
O'IT[] 04[m:J
0.5
1
1.5
O.S
1
1.5
0.5
1
1.5
0.3
0.3
0.2
0.'
0,
Ql
~ffi1TI
O.5 sec 1
15
o
~w
o
O.5 sec 1
1.5
Figure 3: Performance in field B2 after a typical subject (middle column) and each of
the controllers (left and right columns) had adapted to field B 1 . (1) hand paths for
8 movement directions , (2-5) hand velocity, speed, derivative of velocity direction,
and segmented hand path for the -90 0 downward movement . The segmentation in
hand trajectory that is observed in our subjects is almost precisely reproduced by
the controller that uses a forward model.
movements with their hand while an external force field , Fx = Bx, pushed on the
hand. The task was to move the hand to a target position 10 cm away in 0.5
sec. The movement could be directed in any of eight equally spaced directions.
The subjects made straight-path minimum-jerk movements to the targets in the
absence of any force fields. The subjects were initially trained in force field Bl
with B=[O 13;-130]' until they had completely adapted to this field and converged
to the straight-path minimum-jerk movement observed before the force field was
applied. Subsequently, the force field was switched to B2 with B=[O -13;13 0] (the
new field pushed anticlockwise, instead of clockwise), and the first three movements
in each direction were used for data analysis. The movements of the subjects in
field B2 showed huge deviations from the desired straight path behavior because
the subjects expected clockwise force field B 1 ? The hand trajectories for the first
movement in each of the eight directions are shown for a typical subject in Fig. 3
(middle column).
Simulations were performed for the two methods under the same conditions as
the human experiment. The movements were made in force field B 2 , while the
internal models were assumed to be adapted to field B 1 . Complete adaptation
to the force field Bl was found to occur for the two methods only when both
N. Bhushan and R. Shadmehr
8
?
Expenmental
data from
?
16 subjects
Forward
Model
Control
(a)
:[[[1 I~ ~ III
&'
=
Q
A1(")
d l(m)
An
t,(s)
A,(")
cJ(m/s' )
Ns
Figure 4: The mean and standard deviation for segmentation parameters for each
type of controller as compared to the data from our subjects. Parameters are
defined in Fig. 3: Ai is angle about a seg. point, d i is the distance to the i-th
seg. point, ti is time to reach the i-th seg. point, Cj is cumulative squared jerk
for the entire movement, Ns is number of seg. point in the movement. Up until
the first segmentation point (AI and dd, behavior of the controllers are similar
and both agree with the performance of our subjects. However, as the movement
progresses, only the controller that utilizes a forward model continues to agree with
the movement characteristics of the subjects.
the inverse and forward models expected field B I . Fig. 3 (left column) shows the
simulation of the adaptive inverse model feedforward control for movements in field
B2 with the inverse model incorrectly expecting B I . Fig. 3 (right column) shows the
simulation of the adaptive forward-inverse model feedback control for movements
in field B2 with both the forward and the inverse model incorrectly expecting B I .
Simulations with the two methods show clear differences in stability and corrective
behavior for all eight directions of movement. The simulations with the inverse
model feedforward control seem to be stable, and converge to the target along a
straight line after the initial deviation. The simulations with the forward-inverse
model feedback control are more unstable and have a curious kinematic pattern
with discontinuities in the hand path. This is especially marked for the downward
movement. The subject's hand paths show the same kinematic pattern of near
discontinuities and segmentation of movement as found with the forward-inverse
model feedback control.
To quantify the segmentation pattern in the hand path, we identified the "near
path-discontinuities" as points on the trajectory where there was a sudden change
in both the derivative of hand speed and the direction of hand velocity. The hand
path was segmented on the basis of these near discontinuities. Based on the first
three segments in the hand trajectory we defined the following parameters: AI,
angle between the first segment and the straight path to the target; d l , the distance
covered during the first segment; A2, angle between the second segment and straight
path to the target from the first segmentation point; t2, time duration of the second
Evidence for a Forward Dynamics Model in Human Adaptive Motor Control
9
segment; A3, angle between the second and third segments; Ns, the number of
segmentation points in the movement . We also calculated the cumulative jerk CJ
in the movements to get a measure of the instability in the system.
The results of the movement segmentation are presented in Fig. 4 for 16 human subjects, 25 simulations of the inverse model and 20 simulations of the forward model
control for three movement directions (a) -900 downward, (b) 90 0 upward and (c)
135 0 upward. We performed the different simulations for the two methods by systematically varying various model parameters over a reasonable physiological range.
This was done because the parameters are only approximately known and also vary
from subject to subject. The parameters of the second and third segment, as represented by A2, t2 and A3, clearly show that the forward model feedback control
performs very differently from inverse model feedforward control and the behavior
of human subjects is very well predicted by the former. Furthermore, this characteristic behavior could be produced by the forward-inverse model feedback control
only when the forward model expected field B 1 . This could be accomplished only
by adaptation of the forward model during initial practice in field B 1 ? This provides
evidence for an adaptive forward model in the control of human arm movements in
novel dynamic environments.
We further tried to fit adaptation curves of simulated movement parameters (using
forward-inverse model feedback control) to real data as subjects trained in field B 1 .
We found that the best fit was obtained for a rapidly adapting forward and inverse
model (Bhushan and Shadmehr, 1999). This eliminated the possibility that the
inverse model was trained offline after practice. The data, however, suggested that
during learning of a force field, the rate of learning of the forward model was faster
than the inverse model. This finding could be paricularly relevant if it is proven
that a forward model is easier to learn than an inverse model (Narendra, 1990),
and could provide a computational rationale for the existence of forward model in
adaptive motor control.
References
Bhushan N, Shadmehr R (1999) Computational architecture of the adaptive controller
during learning of reaching movements in force fields. Biol Cybern, in press.
Jordan MI, Flash T, Arnon Y (1994) A model of learning arm trajectories from spatial
deviations Journal of Cog Neur 6:359-376 .
Jordan MI, Rumelhart DE (1992) Forward model: supervised learning with a distal
teacher. Cog Sc 16:307-354.
Miall RC, Wolpert DM (1996) Forward models for phySiological motor control. Neural
Networks 9:1265-1279.
Narendra KS (1990) Identification and control of dynamical systems using neural networks.
Neural Networks 1:4-27.
Shadmehr R, Brashers-Krug T (1997) Functional stages in the formation of human longterm memory. J Neurosci 17:409-19.
Shadmehr R, Mussa-Ivaldi FA (1994) Adaptive representation of dynamics during learning
of a motor task. The Journal of Neuroscience 14:3208-3224.
Soechting JF, Flanders M (1997) Evaluating an integrated musculoskeletal model of the
human arm J Biomech Eng 9:93-102 .
Wolpert DM, Ghahramani Z, Jordan MI (1995) An internal model for sensorimotor integration. Science 269:1880-82.
| 1486 |@word longterm:1 middle:2 simulation:12 tried:1 eng:1 moment:1 initial:3 ivaldi:3 current:4 activation:3 yet:1 written:1 john:1 motor:10 nervous:1 sudden:2 provides:4 rc:1 along:1 incorrect:1 behavioral:1 expected:5 rapid:1 behavior:10 examine:2 mechanic:1 brain:1 torque:1 td:2 little:1 jm:1 actual:3 inappropriate:1 elbow:1 arnon:1 cm:1 q2:1 developed:1 finding:1 act:1 ti:1 xd:10 exactly:3 control:51 producing:1 before:1 engineering:1 despite:1 krug:2 path:16 approximately:1 might:1 k:1 perpendicular:1 range:1 directed:1 practice:2 coriolis:1 jhu:2 adapting:2 get:1 cybern:1 instability:4 descending:3 imposed:1 map:2 go:1 duration:1 stability:1 fx:4 updated:1 target:5 exact:2 us:4 velocity:8 rumelhart:2 continues:1 predicts:1 observed:4 ft:1 seg:4 movement:37 substantial:1 expecting:2 environment:4 dynamic:25 signature:1 trained:4 solving:1 segment:7 upon:1 completely:1 basis:1 joint:8 differently:2 represented:4 various:1 corrective:4 train:1 distinct:1 kp:2 sc:1 formation:4 solve:1 nikhil:1 reproduced:1 interaction:1 adaptation:7 relevant:2 loop:2 rapidly:2 kv:2 produce:1 bme:2 progress:1 eq:1 predicted:1 involves:1 quantify:1 direction:9 musculoskeletal:1 correct:2 subsequently:1 human:17 correction:1 around:1 equilibrium:1 predict:2 narendra:2 vary:1 a2:2 purpose:1 soechting:2 clearly:1 reaching:3 varying:1 command:4 contrast:1 bhushan:6 entire:1 integrated:1 initially:2 upward:2 overall:1 classification:1 issue:1 spatial:1 integration:1 field:30 equal:1 eliminated:1 mimic:1 t2:2 composed:1 delayed:2 mussa:3 huge:1 fd:2 possibility:1 kinematic:4 accurate:1 desired:16 causal:1 theoretical:1 column:5 earlier:1 deviation:8 delay:6 teacher:1 off:1 corrects:1 hopkins:1 squared:1 central:2 external:7 derivative:2 bx:2 de:1 accompanied:1 sec:3 b2:5 performed:2 characteristic:3 correspond:1 spaced:1 identification:1 produced:1 trajectory:15 straight:6 converged:1 reach:1 email:1 sensorimotor:1 dm:2 mi:3 gain:1 segmentation:8 cj:3 supervised:1 done:1 furthermore:1 biomedical:1 stage:1 until:2 hand:33 nonlinear:4 effect:2 concept:2 former:1 distal:1 during:7 m:5 hill:1 complete:1 performs:1 passive:1 novel:7 functional:1 spinal:4 reza:2 attached:1 jp:4 ai:3 had:2 stable:4 showed:1 termed:1 accomplished:3 muscle:19 minimum:2 additional:1 converge:1 period:1 signal:4 clockwise:2 segmented:2 match:2 faster:1 compensate:1 divided:2 equally:1 a1:1 controlled:1 xmd:2 controller:10 addition:2 remarkably:1 baltimore:1 subject:22 flow:1 seem:1 jordan:4 curious:1 near:5 feedforward:11 iii:1 variety:1 jerk:4 fit:2 architecture:8 fm:1 identified:1 idea:1 whether:1 six:1 cause:2 action:1 clear:1 covered:1 viscosity:2 generate:1 estimated:6 neuroscience:1 track:2 key:1 achieving:1 practicing:1 inverse:46 facilitated:1 angle:5 almost:1 reasonable:1 utilizes:1 sented:1 pushed:2 adapted:3 occur:2 precisely:1 x2:1 simulate:1 speed:3 neur:1 combination:4 unity:1 equation:1 agree:2 kinematics:2 initiate:1 brashers:2 stiffness:1 eight:3 observe:1 away:1 appropriate:2 simulating:1 jd:1 existence:1 xc:1 ghahramani:1 especially:1 k1:1 bl:2 move:1 occurs:1 parametric:1 fa:1 md:1 nr:2 distance:2 simulated:1 unstable:1 assuming:1 length:1 modeled:1 rotational:1 nc:4 ql:1 holding:1 sing:1 behave:1 incorrectly:2 situation:1 shoulder:1 pair:2 mechanical:1 required:1 learned:3 discontinuity:6 suggested:1 dynamical:2 pattern:4 xm:4 memory:1 force:22 rely:1 predicting:2 arm:23 altered:2 removal:1 revolute:2 plant:6 rationale:1 oms:1 proportional:1 proven:1 sixteen:1 switched:1 principle:2 dd:1 systematically:1 copy:1 offline:1 feedback:27 calculated:1 curve:1 transition:2 cumulative:2 evaluating:1 computes:1 sensory:3 forward:48 ignores:1 adaptive:27 inertia:1 made:3 miall:3 far:1 assumed:2 nature:2 reasonably:1 learn:1 neurosci:1 fig:5 en:1 n:3 fails:1 position:6 msec:4 flanders:2 jacobian:1 third:2 cog:2 jt:1 physiological:2 evidence:10 a3:2 downward:3 push:1 easier:1 wolpert:5 reflex:3 relies:3 marked:1 acceleration:6 flash:1 jf:1 absence:1 change:1 muscular:1 typical:3 corrected:2 shadmehr:11 acting:1 experimental:2 internal:11 dept:1 biol:1 |
532 | 1,487 | A Reinforcement Learning Algorithm
in Partially Observable Environments
Using Short-Term Memory
Nobuo Suematsu and Akira Hayashi
Faculty of Computer Sciences
Hiroshima City University
3-4-1 Ozuka-higashi, Asaminami-ku, Hiroshima 731-3194 Japan
{suematsu,akira} @im.hiroshima-cu.ac.jp
Abstract
We describe a Reinforcement Learning algorithm for partially observable environments using short-term memory, which we call BLHT. Since
BLHT learns a stochastic model based on Bayesian Learning, the overfitting problem is reasonably solved. Moreover, BLHT has an efficient
implementation. This paper shows that the model learned by BLHT converges to one which provides the most accurate predictions of percepts
and rewards, given short-term memory.
1
INTRODUCTION
Research on Reinforcement Learning (RL) problem for partially observable environments is gaining more attention recently. This is mainly because
the assumption that perfect and complete perception
of the state of the environment is available for the
learning agent, which many previous RL algorithms
require, is not valid for many realistic environments.
model-free
Figure I: Three approaches
One of the approaches to the problem is the model-free approach (Singh et al. 1995;
Jaakkola et al. 1995) (arrow a in the Fig.l) which gives up state estimation and uses
memory-less policies. We can not expect the approach to find a really effective policy when
it is necessary to accumulate information to estimate the state. Model based approaches are
superior in these environments.
A popular model based approach is via a Partially Observable Markov Decision Process
(POMDP) model which represents the decision process of the agent. In Fig.1 the approach
is described by the route from "World" to "Policy" through "POMDP". The approach has
two serious difficulties. One is in the learning of POMDPs (arrow b in Fig. I). Abe and
N. Suematsu and A. Hayashi
1060
Warmuth (1992) shows that learning of probabilistic automata is NP-hard, which means
that learning of POMDPs is also NP-hard. The other difficulty is in finding the optimal
policy of a given POMDP model (arrow c in Fig. I ). Its PSAPCE-hardness is shown in
Papadimitriou and Tsitsiklis (1987). Accordingly, the methods based on this approach
(Chrisman 1992; McCallum 1993), will not scale well to large problems.
The approach using short-term memory is computationally more tractable. Of course we
can construct environments in which long-term memory is essential. However, in many
environments, because of their stochasticity, the significance of the past information decreases exponentially fast as the time goes. In such environments, memories of moderate
length will work fine.
McCallum (1995) proposes "utile suffix memory" (USM) algorithm. USM uses a tree
structure to represent short-term memories with variable length. USM's model learning is
based on a statistical test, which requires time and space proportional to the learning steps.
This makes it difficult to adapt USM to the environments which require long learning steps.
USM suffers from the overfitting problem which is a difficult problem faced by most of
model based learning methods. USM may overfit or underfit up to the significance level
used for the statistical test and we can not know its proper level in advance.
In this paper, we introduce an algorithm called BLHT (Suematsu et al. 1997), in which the
environment is modeled as a history tree model (HTM), a stochastic model with variable
memory length. Although BLHT shares the tree structured representation of short-term
memory with USM, the computational time required by BLHT is constant in each step and
BLHT copes with environments which require large learning steps. In addition, because
BLHT is based on Bayesian Learning, the overfitting problem is solved reasonably in it. A
similar version of HTMs was introduced and has been used for learning of Hidden Markov
Models in Ron et at. (1994). In their learning method, a tree is grown in a similar way with
USM. If we try to adapt it to our RL problem, it will face the same problems with USM.
This paper shows that the HTM learned by BLHT converges to the optimal one in the
sense that it provides the most accurate predictions of percepts and rewards, given shortterm memory. BLHT can learn a HTM in an efficient way (arrow d in Fig.l). And since
HTMs compose a subset of Markov Decision Processes (MDPs), it can be efficiently solved
by Dynamic Programming (DP) techniques (arrow e in Fig. I). So, we can see BLHT as an
approach to follow an easy way from "World" to "Policy" which goes around "POMDP".
2
THE POMDP MODEL
The decision process of an agent in a partially observable environment can be formulated
as a POMDP. Let the finite set of states of the environment be S, the finite set of agent's
actions be A, and the finite set of all possible percepts be I. Let us denote the probability
of and the reward for making transition from state 8 to 8' using action a by Ps' lsa and W sas'
respectively. We also denote the probability of obtaining percept i after a transition from 8
to 8' using action a by 0ilsas" Then, a POMDP model is specified by (S, A,I, P, 0, W,
xo), where P = {Ps/l sa 18,8' E S,a E A}, 0 = {oilsas,18,8' E S,a E A,i E I}, W
= {W sas,18, 8' E S, a E A}, and Xo = (X~l" .. , x~I SI_l) is the probability distribution of
the initial state.
We denote the history of actions and percepts of the agent till time t, ( ... , at-2, it-I, at-I,
it) by D t . If the POMDP model, M = (S, A,I, P, 0, W, Xi) is given, one can compute
the belief state, Xt = (X~l"'" x~ISI_l) from Df, which is the state estimation at time t.
We denote the mapping from histories to belief states defined by POMDP model M by
X M( .), that is, Xt = X M(Dt). The belief state Xt is the most precise state estimation
and it is known to be the sufficient statistics for the optimal policy in POMDPs (Bertsekas
1987). It is also known that the stochastic process {Xt, t 2:: O} is an MDP in the continuous
1061
An RL Algorithm in Partially Observable Environments Using Memory
3
BAYESIAN LEARNING OF HISTORY TREE MODELS (BLHT)
In this section. we summarize our RL algorithm for partially observable environments.
which we call BLHT (Suematsu et at. 1997).
3.1
HISTORY TREE MODELS
BLHT is Bayesian Learning on a hypothesis space which is composed of predictive models.
which we call History Tree Models (HTMs). Given short-term memory. a HTM provides
the probability disctribution of the next percept and the expected immediate reward for
each action. A HTM is represented by a tree structure called a history tree and parameters
given for each leaf of the tree.
A history tree h associates history D t with a leaf as follows. Starting from the root of h.
we check the most recent percept. it and follow the appropriate branch and then we check
the action at-l and follow the appropriate branch. This procedure is repeated till we reach
a leaf. We denote the reached leaf by Ah(D t ) and the set of leaves of h by Lh.
Each leaf l E Lh has parameters Billa and Wla. Billa denotes the probability of observing
i at time t + 1 when Ah(Dt} = l and the last action at was a. Wla denotes the expected
immediate reward for performing a when Ah(D t ) = l. Let 8 h = {Billa liE T, l E
Lh,a E A}.
b
(a)
---"--../--..--
~
(b)
2 f-~- - - - it
/'-....
a
b f-~- - - at-l
1
a / " - . . ........
1
2 1
2
~ it-l
Figure 2: (a) A three-state environment. in which the agent receives percept 1 in state 1 and
percept 2 in states 2a and 2b. (b) A history tree which can represent the environment.
Fig. 2 shows a three-state environment (a) and a history tree which can represent the
environment (b). We can construct a HTM which is equivalent with the environment by
setting appropriate parameters in each leaf of the history tree.
3.2 BAYESIAN LEARNING
BLHT is designed as Bayesian Learning on the hypothesis space. 11.. which is a set of
history trees. First we show the posterior probability of a history tree h E 11. given history
D t . To derive the posterior probability we set the prior density of 8h as
p(8 h lh) =
II II
Kia
IELh aEA
II B~:~a-l,
iEI
where Kia is the normalization constant and ailla is a hyper parameter to specify the prior
density. Then we can have the posterior probabili,ty of h.
P(hID 11.) =
t,
Ct
P(hI1l.)
II II
IELh aEA
K
n?
la
,E
r(N~1 + a'll )
, la
~ a
r(Nt + a)
,
I
la
(I)
la
where Ct is the normalization constant. r(?) is the gamma function. Nflla is the number
of times i is observed after executing a when Ah(Dt ,) = l in the history D t ? N/ =
a
L.JiEI N illa ? and ala = L.JiEI ailla'
"t
"
Next. we show the estimates of the parameters. We use the average of Billa with its posterior
1062
N. Suematsu and A. Hayashi
density as the estimate, 8~lla' which is expressed as
~t
+ ailla
t a N / a + a'a .
()"II
Nflla
= -'-;----
W'a is estimated just by accumulating rewards received after executing a when Ah(Dt ) =
and dividing it by the number of times a was performed when Ah (D t ) = l, N/a ? That is,
wIa =
1
Nt
l,
N/'a
L
Ttk+1,
la k=l
where tk is the k-th occurrence of execution of a when Ah(D t ) = l.
3.3
LEARNING ALGORITHM
In principle, by evaluating Eq.( J) for all h E 11., we can extract the MAP model. However,
it is often impractical, because a proper hypothesis space 11. is very large when the agent has
little prior knowledge concerning the environment. Fortunately, we can design an efficient
learning algorithm by assuming that the hypothesis space, 11., is the set of pruned trees of a
large history tree h1i and the ratio of prior probabilities of a history tree h and hi obtained
by pruning off subtree Llh from h is given by a known function q( Llh) I .
We define function g(hIDt,1I.) by taking logarithm of the R.H.S. of Eq.(J) without the
normalization constant, which can be rewritten as
g(hIDt,1I.) = log P(hI1l.)
+
L At,
(2)
IEC h
where
[Kla ItEIreNt
r(Nfl/a + a i ll a )]
AIt = ""'1
~ og
+).
aEA
la
(3)
ala
Then, we can extract the MAP model by finding the history tree which maximizes g. Eq.(2)
shows that g(hIDt, 11.) can be evaluated by summing up At over Lh. Accordingly, we can
implement an efficient algorithm using the tree h1i whose each (internal or leaf) node 1
stores AI, N i l/a , ail/a, and Wla?
Suppose that the agent observed it+l when the last action was at. Then, from Eq.(3),
At+l
-I
{
At
I
AI
+I
og
Nt,tl l/ a , +o<;,tll/ a ,
N'la, +0</ a,
cor lEND,
.'
(4)
otherwise
where N D, is the set of nodes on the path from the root to leaf Ah~ (D t ). Thus, h1i is
updated just by evaluating Eq(4), adding I to Nil /a ' and recalculating Wla in nodes of N D ,.
After h1i is updated, we can extract the MAP model using the procedure "Find-MAPSubtree" shown in Fig. 3(a). We show the learning algorithm in Fig.3(b), in which the
MAP model is extracted and policy 7r is updated only when a given condition is satisfied.
4
LIMIT THEOREMS
In this section, we describe limit theorems of BLHT. Throughout the section, we assume
that policy 7r is used while learning and the stochastic process {(st, at, it+d, t ~ O} is
ergodic under 7r ?
First we show a theorem which ensures that the history tree model learned by BLHT does
not miss any relevant memories (see Suematsu et al. (1997) for the proof).
I The condition is satisfied, for example, when P(hl1i) ex ")'Ikl where 0
the size of h.
< ")'
~ 1 and Ihl denotes
An RL Algorithm in Partially Observable Environments Using Memory
10 -
-
1063
Mam-Loop(condltlOn C)
I: t f- O. D t f- ()
2: rr f- "policy selecting action at random"
3: at f- rr(Dt) or exploratory action
4: perform at and receive it+l and rt+l
5: update hll.
6: if (condition C is satisfied) do
7:
h f- Find-MAP-Subtree(Root(hll?
8:
rr f- Dynamic-Programming(h)
9: end
10: Dt+l f- (Dt ,at,i t +l), t f- t + 1
II: goto 3
(b)
u tree no e
I:
hf- .Af-O
2: C f- {all child nodes of node l}
3: if ICI = 0 then return {l, Ad
4: for each c E C do
5:
{Llhc, Ac} f- Find-MAP-Subtree( c)
6:
Llh f- Llh U Llhc
7:
A f- A+ Ac
8: end
9: Llg f-logq(Llh) + A - Al
10: if Llg > 0 then return {Llh, A}
11: else return l, Al
(a)
Figure 3: The procedure to find MAP subtree (a) and the main loop (b).
Theorem 1 For any h E 11..
lim !g(hID t ,11.) = -Hh(IIL, A),
t
where Hh(IIL, A) is the conditional entropy ofi t+1 given It = Ah(D t ) and at defined by
t-too
Hh(IIL,A) == Err
{z:
-Prr (it+l = i I lt,at)logPrr (i t+1 = i Ilt,at)},
iEI
where Prr (.) and Err (.) denotes probability and expected value under 7r respectively.
Let the history tree shown in Fig.2(b) be h* and a history tree obtained by pruning a subtree
of h* be h-. Then, for the environment shown in Fig.2(a) H h- (IlL, A) > H h? (IlL, A),
because h - misses some relevant memories and it makes the conditional entropy increase.
Since BLHT learns the history tree which maximizes g(hID t , 11.) (minimizes Hh(IIL , A),
the learned history tree does not miss any relevant memory.
Next we show a limit theorem concerning the estimates of the parameters. We denote the
true POMDP model by M = (S, A, I, P, 0, W, Xi) and define the following parameters,
P(it+l
O'i lsa
J-Lsa
=
= i I St = s,at = a) =
E(rt+ll s t
= s,at = a) =
z:
z:
Ps'l saOi lsas'
s'ES
wsas'Ps'lsa'
s'ES
Then, the following theorem holds.
Theorem 2 For any leaf I E Ch, a E A. i E I
(5)
lim
t-too
w:
a
= '"'
~ J-LsaY:lla'
sES
where Y:lla == Prr(St = SIAh(Dt) = I, at = a).
Outline of proof: Using the Ergodic Theorem, We have
lim
t-too
O! lla
= Prr (it+l
= ilAh(Dd = I, at = a).
(6)
N. Suematsu and A. Hayashi
1064
By expanding R.H.S of the above equation using the chain rule, we can derive Eq.(5).
?
Eq.(6) can be derived in a similar way.
To explain what Theorem 2 means clearly, we show the relationship between Y;lla and the
belief state Xt.
P7r (St
= SIAh(Dd = i, at = a, Xo = Xi)
L P(St = SIDt = D, at = a, Xo = xi)P
7r
(Dt
= Dlit = i, at = a, Xo = Xi)
DEDI
=
1 L :n.D~
(D){ X M(D)}s P7r (Dt
= Dlit = i, at = a, Xo = xi)dx
X DEDI
Ix
where Vi
I
XS P7r (Xt
= xlit = i, at = a, Xo = xi)dx,
== {DtIAh(D t ) = I}, :n.B(-) is the indicator function of a set B,
{DtIX M(Dd
= x}, and dx = dXl'"
limt-too of the above equation, we have
Yla =
where Yla
V~
==
dXISI-l' Under the ergodic assumption, by taking
Ix
xCPia(x)dx
= (Y;llla' ... , Y;ISI-I Ila) and CPia (x) = P
7r
(Xt
(7)
= xIAh(Dt) = i, at = a).
We see from Eq.(7) that Yla is the average of belief state Xt with conditional density CPia,
that is, the belief states distributed according to CPla are represented by Yia' When shortterm memory of i gives the dominant information of Dt. CPia is concentrated and Yla is
a reasonable approximation of the belief states. An extreme of the case is when CPia is
non-zero only at a point in X. Then YIa = Xt when Ah(Dd = i.
Please note that given short-term memory represented by i and a, YIa is the most accurate
state estimation. Consequently, Theorem 1 and 2 ensure that learned HTM converges to
the model which provides the most accurate predictions of percepts and rewards among 1/..
This fact provides a solid basis for BLHT, and we believe BLHT can be compared favorably
with other methods using short-term memory. Of course, Theorem 1 and 2 also say that
BLHT will find the optimal policy if the environment is Markovian or semi-Markovian
whose order is small enough for the equivalent model to be contained in 1/..
5
EXPERIMENT
We made experiments in various environments. In this paper, we show one of them to
demonstrate the effectiveness of BLHT. The environment we used is the grid world shown
in Fig.4(a). The agent has four actions to change its location to one of the four neighboring
grids, which will fail with probability 0.2. On failure, the agent does not change the location
with probability 0.1 or goes to one ofthe two grids which are perpendicular to the direction
the agent is trying to go with probability 0.1. The agent can detect merely the existence of
the four surrounding walls. The agent receives a reward of 10 when he reaches the goal
which is the grid marked with "G" and - 1 when he tries to go to a grid occupied by an
obstacle. At the goal, any action will relocate the agent to one of the starting states which
are marked with "S" at random. In order to achieve high performance in the environment,
the agent has to select different actions for an identical immediate percept, because many of
the states are aliased (i.e. they look identical by the immediate percepts). The environment
has 50 states, which is among the largest problems shown in the literature of the model
based RL techniques for partially observable environments.
Fig.4(b) shows the learning curve which is obtained by averaging over 10 independent runs.
While learning, the agent updated the policy every 10 trials (10 visits to the goal) and the
1065
An RL Algorithm in Partially Observable Environments Using Memory
policy was evaluated through a run of 100,000 steps. Actions were selected using the policy or at random and the probability of selecting at random was decreased exponentially as
the time goes. We used the tree which has homogeneous depth of 5 as h1i.. In Fig.4(b), the
horizontal broken line indicates the average reward for the MOP model obtained by assuming perfect and complete perception. It gives an upper bound for the original problem, and
it will be higher than the optimal one for the original problem. The learning curve shown
there is close to the upper bound in the later stage.
(a)
(b)
1 - .- .- .--- --.-.-.. -.---.-.-..-- .- - - ---.-._-.
0.8
0.6
0.4
0.2
o
2000
4000
6000
8000
10000
trials
Figure 4: The grid world (a) and the learning curve (b).
6
SUMMARY
This paper has described a RL algorithm for partially observable environments using shortterm memory, which we call BLHT. We have proved that the model learned by BLHT
converges to the optimal model in given hypothesis space, 1{, which provides the most
accurate predictions of percepts and rewards, given short-term memory. We believe this
fact provides a solid basis for BLHT, and BLHT can be compared favorably with other
methods using short-term memory.
References
Abe, N. and M. K. Warmuth (1992). On the computational compleixy of apporximating distributions
by probabilistic automata. Machine Learning, 9:205-260.
Bertsekas, D. P. (1987). Dyanamic Programming. Prentice-Hall.
Chrisman, L. (1992). Reinforcemnt learning with perceptual aliasing: The perceptual distinctions
approach. In Proc. the 10th National Conference on Artificial Intelligence.
Jaakkola, T. , S. P. Singh, and M. I. Jordan (1995). Reinforcement learning algorithm for parially
observable markov decision problems. In Advances in Neural Information Processing Systems 7,
pp. 345-352.
McCallum, R. A. (1993). Overcoming incomplete perception with utile distiction memory. In Proc.
the 10th International Conference on Machine Learning.
McCallum, R. A. (1995). Instance-based utile distinctions for reinforcement learning with hidden
state. In Proc. the 12th International Conference On Machine Learning.
Papadimitriou, C. H. and J. N. Tsitsiklis (1987). The compleXity of markov decision processes.
Mathematics of Operations Research, 12(3):441-450.
Ron, D., Y. Singer, and N. Tishby (1994). Learning probabilistic automata with variable memory
length. In Proc. of Computational Learning Theory, pp. 35-46.
Singh, S. P., T. Jaakkola, and M. I. Jordan (1995). Learning without state-estimation in partially
observable markov decision processes. In Proc. the 12th International Conference on Machine
Learning, pp. 284-292.
Suematsu, N., A. Hayashi, and S. Li (1997). A Bayesian approch to model learning in nonmarkovian environments. In Proc. the 14th International Conference on Machine Learning, pp.
349-357.
| 1487 |@word trial:2 cu:1 version:1 faculty:1 wla:4 solid:2 dedi:2 initial:1 selecting:2 ala:2 past:1 err:2 nt:3 dx:4 realistic:1 designed:1 update:1 intelligence:1 leaf:10 selected:1 warmuth:2 accordingly:2 mccallum:4 short:11 utile:3 provides:7 node:5 ron:2 location:2 compose:1 introduce:1 expected:3 hardness:1 isi:1 aliasing:1 little:1 moreover:1 maximizes:2 aliased:1 what:1 ail:1 minimizes:1 finding:2 impractical:1 every:1 lsa:4 bertsekas:2 limit:3 path:1 perpendicular:1 implement:1 procedure:3 lla:5 ila:1 close:1 prentice:1 accumulating:1 equivalent:2 map:7 go:6 attention:1 starting:2 automaton:3 ergodic:3 pomdp:10 h1i:5 rule:1 htms:3 exploratory:1 updated:4 suppose:1 programming:3 homogeneous:1 us:2 hypothesis:5 associate:1 logq:1 observed:2 higashi:1 solved:3 probabili:1 ensures:1 decrease:1 environment:33 broken:1 complexity:1 reward:10 dynamic:2 singh:3 predictive:1 basis:2 htm:7 represented:3 various:1 grown:1 hiroshima:3 surrounding:1 approch:1 fast:1 describe:2 effective:1 artificial:1 hyper:1 whose:2 say:1 s:1 otherwise:1 statistic:1 rr:3 relocate:1 hid:3 relevant:3 loop:2 neighboring:1 till:2 achieve:1 p:4 perfect:2 converges:4 executing:2 tk:1 derive:2 ac:3 received:1 eq:8 sa:3 dividing:1 direction:1 stochastic:4 require:3 really:1 wall:1 im:1 hidt:3 hold:1 around:1 hall:1 iil:4 mapping:1 mop:1 dxl:1 estimation:5 proc:6 largest:1 city:1 ttk:1 clearly:1 occupied:1 og:2 jaakkola:3 tll:1 derived:1 check:2 mainly:1 indicates:1 nonmarkovian:1 sense:1 detect:1 suffix:1 hidden:2 among:2 ill:2 proposes:1 construct:2 identical:2 represents:1 look:1 papadimitriou:2 np:2 serious:1 composed:1 gamma:1 national:1 llg:2 recalculating:1 extreme:1 chain:1 accurate:5 necessary:1 lh:5 tree:28 incomplete:1 logarithm:1 instance:1 obstacle:1 markovian:2 subset:1 too:4 tishby:1 st:5 density:4 international:4 probabilistic:3 off:1 satisfied:3 yia:3 return:3 li:1 japan:1 iei:2 ad:1 vi:1 performed:1 try:2 root:3 later:1 observing:1 reached:1 hf:1 percept:13 efficiently:1 ofthe:1 bayesian:7 pomdps:3 history:24 ah:10 explain:1 reach:2 suffers:1 failure:1 ty:1 pp:4 proof:2 proved:1 popular:1 ihl:1 knowledge:1 lim:3 higher:1 dt:12 follow:3 specify:1 evaluated:2 just:2 stage:1 overfit:1 receives:2 horizontal:1 ikl:1 mdp:1 believe:2 true:1 ll:3 please:1 trying:1 outline:1 complete:2 demonstrate:1 recently:1 superior:1 rl:9 jp:1 exponentially:2 he:2 accumulate:1 ai:2 grid:6 mathematics:1 stochasticity:1 dominant:1 posterior:4 recent:1 moderate:1 route:1 store:1 suematsu:9 fortunately:1 akira:2 p7r:3 ii:7 branch:2 semi:1 adapt:2 af:1 long:2 concerning:2 visit:1 prediction:4 df:1 represent:3 normalization:3 limt:1 receive:1 addition:1 fine:1 decreased:1 else:1 goto:1 effectiveness:1 jordan:2 call:4 easy:1 enough:1 nfl:1 aea:3 action:14 prr:4 concentrated:1 estimated:1 four:3 merely:1 run:2 ilt:1 throughout:1 reasonable:1 decision:7 bound:2 ct:2 hi:1 pruned:1 performing:1 structured:1 according:1 making:1 xo:7 computationally:1 equation:2 fail:1 hh:4 singer:1 know:1 tractable:1 end:2 cor:1 available:1 operation:1 rewritten:1 appropriate:3 occurrence:1 existence:1 original:2 mam:1 denotes:4 jiei:2 ensure:1 rt:2 dp:1 assuming:2 length:4 modeled:1 relationship:1 ratio:1 difficult:2 favorably:2 implementation:1 design:1 proper:2 policy:13 perform:1 upper:2 markov:6 finite:3 immediate:4 precise:1 yla:4 abe:2 overcoming:1 introduced:1 required:1 specified:1 learned:6 distinction:2 chrisman:2 perception:3 summarize:1 gaining:1 memory:27 lend:1 belief:7 difficulty:2 indicator:1 hll:2 mdps:1 shortterm:3 extract:3 faced:1 prior:4 literature:1 expect:1 proportional:1 agent:16 sufficient:1 principle:1 dd:4 share:1 course:2 kia:2 summary:1 last:2 free:2 tsitsiklis:2 face:1 taking:2 distributed:1 curve:3 depth:1 valid:1 world:4 transition:2 evaluating:2 llh:6 made:1 reinforcement:5 ici:1 cope:1 pruning:2 observable:13 usm:9 overfitting:3 summing:1 xi:7 continuous:1 wia:1 ku:1 reasonably:2 learn:1 iec:1 expanding:1 obtaining:1 significance:2 main:1 arrow:5 underfit:1 ait:1 repeated:1 child:1 fig:14 tl:1 lie:1 perceptual:2 learns:2 kla:1 ix:2 theorem:11 xt:9 x:1 essential:1 adding:1 execution:1 subtree:5 entropy:2 lt:1 expressed:1 contained:1 partially:12 hayashi:5 ch:1 extracted:1 conditional:3 goal:3 formulated:1 marked:2 consequently:1 hard:2 change:2 averaging:1 miss:3 called:2 nil:1 e:2 la:7 ofi:1 select:1 internal:1 ex:1 |
533 | 1,488 | A Model for Associative Multiplication
G. Bjorn Christianson*
Department of Psychology
McMaster University
Hamilton,Ont. L8S 4Kl
bjorn@caltech.edu
Suzanna Becker
Department of Psychology
McMaster University
Hamilton, Onto L8S 4Kl
becker@mcmaster.ca
Abstract
Despite the fact that mental arithmetic is based on only a few hundred basic facts and some simple algorithms, humans have a difficult time mastering the subject, and even experienced individuals
make mistakes. Associative multiplication, the process of doing
multiplication by memory without the use of rules or algorithms,
is especially problematic. Humans exhibit certain characteristic
phenomena in performing associative multiplications, both in the
type of error and in the error frequency. We propose a model for
the process of associative multiplication, and compare its performance in both these phenomena with data from normal humans
and from the model proposed by Anderson et al (1994).
1
INTRODUCTION
Associative mUltiplication is defined as multiplication done without recourse to
computational algorithms, and as such is mainly concerned with recalling the basic
times table. Learning up to the ten times table requires learning at most 121
facts; in fact, if we assume that normal humans use only four simple rules, the
number of facts to be learned reduces to 39. In theory, associative multiplication is
therefore a simple problem. In reality, school children find it difficult to learn, and
even trained adults have a relatively high rate of error, especially in comparison
to performance on associative addition, which is superficially a similar problem.
There has been surprisingly little work done on the methods by which humans
perform basic multiplication problems; an excellent review of the current literature
is provided by McCloskey et al (1991).
If a model is to be considered plausible, it must have error characteristics similar to
* Author to whom correspondence should be addressed. Current address: Computation
and Neural Systems, California Institute of Technology 139-74, Pasadena, CA 91125.
G. B. Christianson and S. Becker
18
those of humans at the same task. In arithmetic, this entails accounting for, at a
minimum, two phenomena. The first is the problem size effect, as noted in various
studies (e.g. Stazyk et ai, 1982), where response times and error rates increase for
problems with larger operands. Secondly, humans have a characteristic distribution
in the types of errors made. Specifically, errors can be classified as one of the
following five types, as suggested by Campbell and Graham (1985), Siegler (1988),
McCloskey et al (1991), and Girelli et al (1996): operand, where the given answer is
correct with one of the operands replaced (e.g. 4 x 7 = 21; this category accounts
for 66.4% of all errors made by normal adults); close-miss, where the result is within
ten percent of the correct response (4 x 7 = 29; 20.0%); table, where the result is
correct for a problem with both operands replaced (4 x 7 = 25; 3.9%); non-table,
where the result is not on the times table (4 x 7 = 17; 6.7%); or operation, where
the answer would have been correct for a different arithmetic operation, such as
addition (4 x 7 = 11; 3.0%)1.
It is reasonable to assume that humans use at least two distinct representations
when dealing with numbers. The work by Mandler and Shebo (1982) on modeling
the performance of various species (including humans, monkeys, and pigeons) on
numerosity judgment tasks suggests that in such cases a coarse coding is used. On
the other hand, humans are capable of dealing with numbers as abstract symbolic
concepts, suggesting the use of a precise localist coding. Previous work has either
used only one of these coding ideas (for example, Sokol et ai, 1991) or a single
representation which combined aspects of both (Anderson et ai, 1994).
Warrington (1982) documented DRC, a patient who suffered dyscalculia following
a stroke. DRC retained normal intelligence and a grasp of numerical and arithmetic
concepts. When presented with an arithmetic problem, DRC was capable of rapidly
providing an approximate answer. However, when pressed for a precise answer, he
was incapable of doing so without resorting to an explicit computational algorithm
such as counting. One possible interpretation of this case study is that D RC retained
the ability to work with numbers in a magnitude-related fashion, but had lost the
ability to treat numbers as symbolic concepts. This suggests the hypothesis that
humans may use two separate, concurrent representations for numbers: both a
coarse coding and a more symbolic, precise coding in the course of doing associative
arithmetic in general, and multiplication in particular, and switch between the
codings at various points in the process. This hypothesis will form the basis of our
modeling work. To guide the placement of these transitions between representations,
we assume the further constraint that the coarse coding is the preferred coding (as
it is conserved across a wide variety of species) and will tend to be expressed before
the precise coding.
1
1
1
1
1
1
1
1
1
1
1
?
6
?
1
I
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
,
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
I
I
1
1
1
1
1
1
1
1
1
1
1
1 1 1 1 1
1 1 1 1 1 1
123450188111111111122222222223333333333444"44444456056(i
012345'181012346178001234056780012345'18801234
Figure 1: The coarse coding for digits. Numbers along the left are the digit; numbers
along the bottom are position numbers. Blank regions in the grid represent zero
activity.
IData taken from Girelli et al (1996).
19
A Model for Associative Multiplication
2
METHODOLOGY
Following the work of Mandler and Shebo (1982), our coarse coding consists of a
54-dimensional vector, with a sliding "bump" of ones corresponding to the magnitude of the digit represented. The size of the bump decreases and the degree of
overlap increases as the magnitude of the digit increases (Figure 1). Noise in this
representation is simulated by the probability that a given bit will be in the wrong
state. The precise representation, intended for symbolic manipulation of numbers,
consists of a 10-dimensional vector with the value of the coded digit given by the
dimension of greatest activity. Both of these representations are digit-based: each
vector codes only for a number between 0 and 9, with concatenations of vectors
used for numbers greater than 9.
o
?o
o
o
o
?o
o
o
direction of flow
Figure 2: Schematic of the network architecture. (A) The coarse coding. (B) The
winner-take-all network. (C) The precise coding. (D) The feed-forward look-up
table. See text for details.
The model is trained in three distinct phases. A simple one-layer perceptron
trained by a winner-take-all competitive learning algorithm is used to map the
input operands from the original coarse coding into the precise representation.
The network was trained for 10 epochs, each with a different set of 5 samples
of noisy coarse-coded digits. At the end of training, the winner-take-all network
performed at near-perfect levels. The translated operands are then presented to a
two-layer feed-forward network with a logistic activation function trained by backpropagation. The number of hidden units was equal to the number of problems in
the training set (in this case, 32) to force look-up table behaviour. The look-up
table was trained independently for varying numbers of iterations, using a learning rate constant of 0.01. The output of the look-up table is coarse coded as in
Figure 1. In the final phase, the table output is translated by the winner-take-all
network to provide the final answer in the precise coding. A schematic of the network architecture is given in Figure 2. The operand vectors used for training of
both networks had a noise parameter of 5%, while the vectors used in the analysis
had 7.5% noise. Both the training and the testing problem set consisted of ten
copies of each of the problems listed in Table 2, which are the problems used in
G. B. Christianson and S. Becker
20
Anderson et al (1994). Simulations were done in MATLAB v5.1 (Mathworks, Inc.,
24 Prime Park Way, Natick MA, 01760-1500).
3
RESULTS
OO r---------~====================~
i
80
o
operand
close-miss
Normal humans (Girelli eta/1996)
Model 01 Anderson eta/(1994)
Model, 200 iterations training
Model, 400 iterations training
Model, 600 iterations training
table
non-table
operation
Error Category
Figure 3: Error distributions for human data (Girelli et al 1996), the model of
Anderson et al (1994), and our model.
Once a model has been trained, its errors on the training data can be categorized
according to the error types listed in the Introduction section; a summary of the
performance of our model is presented in Table 1. For comparison, we plot data
generated by our model, the model of Anderson et al (1994), and human data from
Girelli et al (1996) in Figure 3. In no case did the model generate an operation error.
This is to be expected, as the model was only trained on multiplication, it should
permit no way in which to make an operation error, other than by coincidence. A
full set of results obtained from the model with 400 training iterations is presented
in Table 22.
Table 1: Error rates generated by our model. A column for operation errors is not
included, as in no instance did our model generate an operation error.
Iterations
200
400
600
Errors in
320 trials
114
85
65
Operand
Close-miss
Table
Non-table
(%)
(%)
(%)
(%)
61.4
65.9
63.7
21.0
20.0
16.9
8.8
7.1
9.2
8.8
7.1
10.8
2 As in Anderson et al (1994) , we have set 8 x 9
only problem with an answer greater than 70.
= 67 deliberately so that it is not the
21
A Model for Associative Multiplication
Table 2: Results from ten trials run with the model after 400 training iterations.
Errors are marked in boldface.
I Problem I
2x2
2x4
2x5
3x7
3x8
3x9
4x2
4x5
4x6
4x8
4x9
5x2
5x7
5x8
6x3
6x4
6x5
6x6
6x7
6x8
7x3
7x4
7x5
7x6
7x7
7x8
8x3
8x4
8x6
8x7
8x8
8x9
1
4
8
10
21
24
27
8
20
24
32
36
I
2
4
8
10
21
24
27
8
20
24
32
36
10
10
30
30
24
24
30
36
42
64
24
22
35
42
29
64
24
32
44
56
64
67
42
30
18
24
30
42
32
49
21
28
35
42
49
64
24
32
49
52
64
67
I
3
4
8
10
21
24
27
8
20
24
32
36
30
30
30
18
24
30
36
49
42
21
28
35
42
49
56
21
32
49
56
64
67
I
4
4
8
10
21
64
27
8
20
20
32
36
10
35
35
24
18
30
36
42
49
21
28
35
42
49
64
24
32
44
49
64
67
I
5Trta~
4
8
10
21
24
27
8
30
20
22
21
10
35
30
28
24
30
36
42
44
21
28
35
42
49
56
34
32
44
62
54
67
4
8
10
21
24
27
8
20
24
32
36
10
35
34
12
24
30
36
42
44
21
28
30
42
52
64
24
32
46
46
64
67
I
7
4
8
10
21
21
21
8
20
24
32
36
10
30
30
18
24
30
36
42
64
21
28
35
42
49
56
24
64
42
64
64
67
I
8
4
8
10
21
24
27
10
20
20
32
30
10
30
30
18
24
30
36
42
48
21
28
35
42
42
56
24
32
49
64
64
67
I
9
4
8
10
21
24
27
8
20
24
32
36
10
35
40
24
18
30
36
42
40
21
28
35
49
49
64
24
32
44
49
64
67
I 10 I
4
8
10
21
21
27
8
20
35
32
36
10
35
34
24
18
30
36
42
44
24
32
35
42
42
56
24
32
56
56
64
67
The convention in the current arithmetic literature is to test for the existence of a
problem-size effect by fitting a line to the errors made versus the sum of operands
in the problem. Positive slopes to such fits would demonstrate the existence of a
problem size effect. The results of this analysis are shown in Figure 4. The model
had a problem size effect in all instances. Note that no claims are made of the
appropriateness of a linear model for the given data, nor should any conclusions be
drawn from the specific parameters of the fit, especially given the sparsity of the
data. The sole point of this analysis is to highlight a generally increasing trend.
4
DISCUSSION
As noted in the Results section above, our model demonstrates the problem-size
effect in number of errors made (see Figure 4), though the chosen architecture does
not permit a response time effect. The presence of this effect is hardly surprising,
as all models which use a representation similar to our coarse coding (Mandler &
Shebo, 1982; Anderson et al, 1994) display a problem-size effect.
G. B. Christianson and S. Becker
22
80
?
70
60
y=3 .6x-13
?
~
u
50
~
?
-
840
c::
~30
20
? ?
?
10
? ?
10
12
14
16
18
Sum of Operands
Figure 4: Demonstration of the problem size effect. The data plotted here is for
the model trained for 400 iterations, as it proved the best fit to the distribution of
errors in humans (Figure 3); a similar analysis gives a best-fit slope of 1.9 for 200
training iterations and 1.1 for 600 training iterations.
It has been suggested by a few researchers (e .g. Campbell & Graham, 1985) that
the problem-size effect is simply a frequency effect, as humans encounter problems
involving smaller operands more often in real life. While there is some evidence to
the contrary (Hamman and Ashcraft , 1986) , it remains a possibility.
It is immediately apparent from Figure 3 that our model has much the same distribution of errors as seen in normal humans, and is superior to the model of Anderson
et al (1994) in this regard. That model, implemented as an auto-associative network
using a Brain State in a Box (BSB) architecture (Anderson et al, 1994; Anderson
1995) generates too many operand errors, and no table, non-table or operation
errors. These deficiencies can be predicted from the attractor nature of an autoassociative network. It is the process of translating between representations for
digits, and the possibility for error in doing so, which we believe allows our model
to produce its various categories of errors .
An interesting aspect of our model is revealed by Figure 3 and Table 1. While increased training of the look-up table improves the overall performance of the model,
the error distribution remains relatively constant across the length of training studied. This suggests that in this model, the error distribution is an inherent feature
of the architecture, and not a training artifact. This corresponds with data from
normal humans , in which the error distribution remains relatively constant across
individuals (Girelli et al, 1996). As noted above, the design of our model should
permit the occurrence of all the various error types, save for operation errors. However, at this point, we do not have a clear understanding of the exact architectural
features that generate the error distribution itself.
Defining a model for associative multiplication is only a single step towards the goal
of understanding how humans perform general arithmetic. Rumelhart et al (1986)
proposed a mechanism for multi-digit arithmetic operations given a mechanism for
single-digit operations , which addresses part of the issue; this model has been implemented for addition by Cottrell and T 'sung (1991). The fact that humans make
operation errors suggests that there might be interactions between the mechanisms
A Model for Associative Multiplication
23
of associative multiplication and associative addition; conversely, errors on these
tasks may occur on different processing levels entirely.
In summary, this model , despite several outstanding questions, shows great potential
as a description of the associative multiplication process. Eventually, we expect it
to form the basis for a more complete model of arithmetic in human cognition.
Acknowledgements
The first author acknowledges financial support from McMaster University and Industry Canada. The second author acknowledges financial support from the Natural
Sciences and Engineering Research Council of Canada. We would like to thank J .
Linden, D. Meeker, J. Pezaris, and M. Sahani for their feedback and comments on
this work.
References
Anderson J.A. et al. (1994) In Neural Networks for Knowledge Inference and Representation, Levine D.S. & Aparcicio M., Eds. (Lawrence Erlbaum Associates,
Hillsdale NJ) pp. 311-335.
Anderson J.A. (1995) An Introduction to Neural Networks. (MIT Press/Bradford,
Cambridge MA) pp. 493-544.
Campbell J.I.D. & Graham D.J. (1985) Canadian Journal of Psychology. 39338.
Cottrell G.W. & T'sung F.S . (1991) In Advances in Connectionist and Neural Computation Theory, Burnden J.A. & Pollack J.B ., Eds. (Ablex Publishing Co., Norwood NJ) pp. 305-321.
Girelli L. et al. (1996) Cortex. 32 49.
Hamman M.S. & Ashcraft M.H. (1986) Cognition and Instruction. 3 173.
Mandler G. & Shebo B.J. (1982) Journal of Experimental Psychology: General. 111
1.
McCloskey M. et al. (1991) Journal of Experimental Psychology: Learning, Memory, and Cognition. 17 377.
Rumelhart D.E. et al. (1986) In Parallel distributed processing: Explorations in the
microstructure of cognition. Vol. 2: Psychological and biological models, McClelland
JL, Rumelhart DE, & the PDP Research Groups, Eds. (MIT Press/Bradford,
Cambridge MA) pp. 7-57.
Siegler R. (1988) Journal of Experimental Psychology: General. 117 258.
Stazyk E.H. et al. (1982) Journal of Experimental Psychology: Learning, Memory,
and Cognition. 8 355.
Warrington E.K. (1982) Quarterly Journal of Experimental Psychology. 34A 31.
| 1488 |@word trial:2 instruction:1 simulation:1 accounting:1 pressed:1 current:3 blank:1 surprising:1 activation:1 must:1 cottrell:2 numerical:1 plot:1 intelligence:1 mental:1 coarse:10 five:1 rc:1 along:2 consists:2 fitting:1 expected:1 nor:1 multi:1 brain:1 ont:1 little:1 increasing:1 provided:1 sokol:1 monkey:1 nj:2 sung:2 wrong:1 demonstrates:1 unit:1 hamilton:2 before:1 positive:1 engineering:1 treat:1 mistake:1 despite:2 might:1 studied:1 suggests:4 conversely:1 co:1 testing:1 bsb:1 lost:1 backpropagation:1 x3:3 digit:10 symbolic:4 onto:1 close:3 map:1 independently:1 immediately:1 suzanna:1 rule:2 financial:2 exact:1 hypothesis:2 drc:3 trend:1 rumelhart:3 associate:1 bottom:1 levine:1 coincidence:1 region:1 numerosity:1 decrease:1 trained:9 ablex:1 basis:2 translated:2 various:5 represented:1 distinct:2 apparent:1 larger:1 plausible:1 ability:2 noisy:1 itself:1 final:2 associative:16 propose:1 interaction:1 rapidly:1 description:1 produce:1 perfect:1 oo:1 sole:1 school:1 implemented:2 predicted:1 convention:1 appropriateness:1 direction:1 correct:4 exploration:1 human:21 translating:1 hillsdale:1 behaviour:1 microstructure:1 biological:1 secondly:1 considered:1 normal:7 great:1 lawrence:1 cognition:5 bump:2 claim:1 council:1 concurrent:1 mit:2 varying:1 mainly:1 inference:1 pasadena:1 hidden:1 overall:1 issue:1 equal:1 once:1 x4:4 park:1 look:5 connectionist:1 inherent:1 few:2 individual:2 replaced:2 intended:1 phase:2 attractor:1 recalling:1 possibility:2 grasp:1 capable:2 plotted:1 pollack:1 psychological:1 instance:2 column:1 modeling:2 eta:2 increased:1 industry:1 localist:1 hundred:1 erlbaum:1 too:1 answer:6 combined:1 x9:3 mcmaster:4 account:1 suggesting:1 potential:1 de:1 coding:16 inc:1 performed:1 doing:4 competitive:1 parallel:1 slope:2 characteristic:3 who:1 judgment:1 researcher:1 classified:1 stroke:1 ed:3 frequency:2 pp:4 proved:1 knowledge:1 improves:1 campbell:3 feed:2 x6:4 methodology:1 response:3 done:3 though:1 box:1 anderson:13 hand:1 warrington:2 logistic:1 artifact:1 believe:1 effect:11 concept:3 consisted:1 deliberately:1 pezaris:1 x5:4 noted:3 complete:1 demonstrate:1 percent:1 superior:1 operand:13 winner:4 jl:1 he:1 interpretation:1 cambridge:2 ai:3 resorting:1 grid:1 had:4 entail:1 cortex:1 prime:1 manipulation:1 certain:1 incapable:1 life:1 caltech:1 conserved:1 seen:1 minimum:1 greater:2 arithmetic:10 mandler:4 sliding:1 full:1 reduces:1 coded:3 schematic:2 involving:1 basic:3 patient:1 natick:1 iteration:10 represent:1 addition:4 addressed:1 suffered:1 comment:1 subject:1 tend:1 contrary:1 flow:1 near:1 counting:1 presence:1 revealed:1 canadian:1 concerned:1 switch:1 variety:1 fit:4 psychology:8 architecture:5 idea:1 becker:5 hardly:1 matlab:1 autoassociative:1 generally:1 clear:1 listed:2 ten:4 category:3 mcclelland:1 documented:1 generate:3 problematic:1 vol:1 group:1 four:1 drawn:1 idata:1 sum:2 run:1 reasonable:1 architectural:1 graham:3 bit:1 entirely:1 layer:2 display:1 correspondence:1 activity:2 occur:1 placement:1 constraint:1 deficiency:1 x2:3 x7:5 aspect:2 generates:1 performing:1 relatively:3 department:2 according:1 across:3 smaller:1 mastering:1 taken:1 recourse:1 remains:3 eventually:1 mechanism:3 mathworks:1 end:1 operation:12 permit:3 quarterly:1 occurrence:1 save:1 encounter:1 existence:2 original:1 publishing:1 especially:3 v5:1 question:1 exhibit:1 separate:1 thank:1 simulated:1 concatenation:1 whom:1 boldface:1 length:1 code:1 retained:2 providing:1 demonstration:1 difficult:2 design:1 perform:2 defining:1 precise:8 pdp:1 canada:2 kl:2 california:1 learned:1 adult:2 address:2 suggested:2 sparsity:1 including:1 memory:3 greatest:1 overlap:1 natural:1 force:1 technology:1 x8:6 acknowledges:2 auto:1 sahani:1 text:1 review:1 literature:2 epoch:1 understanding:2 acknowledgement:1 multiplication:17 expect:1 highlight:1 interesting:1 christianson:4 versus:1 norwood:1 degree:1 course:1 summary:2 surprisingly:1 copy:1 guide:1 perceptron:1 institute:1 wide:1 distributed:1 regard:1 feedback:1 dimension:1 superficially:1 transition:1 author:3 made:5 forward:2 approximate:1 preferred:1 dealing:2 meeker:1 table:23 reality:1 learn:1 nature:1 ca:2 excellent:1 did:2 noise:3 child:1 categorized:1 fashion:1 experienced:1 position:1 explicit:1 specific:1 linden:1 evidence:1 l8s:2 magnitude:3 siegler:2 pigeon:1 simply:1 bjorn:2 expressed:1 mccloskey:3 corresponds:1 ma:3 marked:1 goal:1 towards:1 included:1 specifically:1 miss:3 specie:2 bradford:2 experimental:5 support:2 outstanding:1 phenomenon:3 |
534 | 1,489 | A N euromorphic Monaural Sound
Localizer
John G. Harris, Chiang-Jung Pu, and Jose C. Principe
Department of Electrical & Computer Engineering
University of Florida
Gainesville, FL 32611
Abstract
We describe the first single microphone sound localization system
and its inspiration from theories of human monaural sound localization. Reflections and diffractions caused by the external ear (pinna)
allow humans to estimate sound source elevations using only one
ear. Our single microphone localization model relies on a specially
shaped reflecting structure that serves the role of the pinna. Specially designed analog VLSI circuitry uses echo-time processing to
localize the sound. A CMOS integrated circuit has been designed,
fabricated, and successfully demonstrated on actual sounds.
1
Introduction
The principal cues for human sound localization arise from time and intensity differences between the signals received at the two ears. For low-frequency components
of sounds (below 1500Hz for humans), the phase-derived interaural time difference
(lTD) can be used to localize the sound source. For these frequencies, the sound
wavelength is at least several times larger than the head and the amount of shadowing (which depends on the wavelength of the sound compared with the dimensions of
the head) is negligible. lTD localization is a well-studied system in biology (see e.g.,
[5]) and has even been mapped to neuromorphic analog VLSI circuits with limited
success on actual sound signals [6] [2]. Above 3000Hz, interaural phase differences
become ambiguous by multiples of 3600 and are no longer viable localization cues.
For these high frequencies, the wavelength of the sound is small enough that the
sound amplitude is attenuated by the head. The intensity difference of the log magnitudes at the ears provides a unique interaural intensity difference (lID) that can
be used to localize.
Many studies have shown that when one ear is completely blocked, humans can
still localize sounds in space, albeit at a worse resolution in the horizontal direc-
693
A Neuromorphic Monaural Sound Localizer
Sound Signal
---- --,
Detecting
Model :
Generating
Onset
Pulse
i
Neuromorphic
Microphone
Adaptive
Threshold
Computing
I
Delay
(a)
Reflector
Sl
C
:
-,~:Il
. ""
~"
r2
. ; d1
~ k'
Mic.
.. 2
.. 1
Source
,
f
'
d
S2
Reflector
(b )
Figure 1: (a) Proposed localization model is inspired from the biological model (b)
Special reflection surface to serve the role of the pinna
tion. Monaural localization requires that information is somehow extracted from
the direction-dependent effects of the reflections and diffractions of sound off of the
external ear (pinna), head, shoulder, and torso. The S<rcalled "Head Related Transfer Function" (HRTF) is the effective direction-dependent transfer function that is
applied to the incoming sound to produce the sound in the middle ear. Section
2 of this paper introduces our monaural sound localization model and Section 3
discusses the simulation and measurement results.
2
Monaural Sound Localization Model
Batteau [1] was one of the first to emphasize that the external ear, specifically the
pinna, could be a source of spatial cues that account for vertical localization. He
concluded that the physical structure of the external ear introduced two Significant
echoes in addition to the original sound. One echo varies with the azimuthal position
of the sound source, having a latency in the 0 to 80t'S range, while the other varies
with elevation in the lOOt'S to 300t'S range. The output y(t) at the inner ear is
related to the original sound source x(t) as
y(t)
= x(t) + atx(t -
Ta)
+ a2x(t -
Tv)
(1)
where T a , Tv refer to azimuth and elevation echoes respectively; at and a2 are two
reflection constants. Other researchers subsequently verified these results [11] [4].
Our localizer system (shown in Figure l(a)) is composed of a special reflection
surface that encodes the sound source's direction, a silicon cochlea that functions
as a band-pass filter bank, onset detecting circuitry that detects and amplifies the
energy change at each frequency tap, pulse generating circuitry that transfers analog
sound signals into pulse signals based on adaptively thresholding the onset signal,
and delay time computation circuitry that computes the echo's time delay then
decodes the sound source's direction.
Since our recorded signal is composed of a direct sound and an echo, the sound is a
simplified version of actual HRTF recordings that are composed of the direct sound
J. G. Harris, c.-J. Pu and J. C. Principe
694
YIn
Figure 2: (a) Sound signal's onset is detected by taking the difference 01 two low-pass
filters with different time constants. (b) Pulse generating circuit.
and its reflections from the external ear, head, shoulder, and torso. To achieve
localization in a ID plane, we may use any shape of reflection surface as long as the
reflection echo caused by the surface provides a one-to-one mapping between the
echo's delay time and the source's direction. Thus, we propose two flat surfaces to
compose the reflection structure in our proposed model depicted in Figure l(b). A
microphone is placed at distances a1 and lI2 from two flat surfaces (81 and 8 2 ), dis
the distance between the microphone and the sound source moving line (the dotted
line in Figure l(b). As shown in Figure l(b), a sound source is at L~ position.
If the source is far enough from the reflection surface, the ray diagram is valid to
analyze the sound's behavior. We skip the complete derivation but the echo's delay
time can be expressed as
(2)
c
where d 1 is the length of the direct path, r1 + r2 is reflected path length, and c
is the speed of sound. The path distance are easily solved in terms of the source
direction and the geometry of the setup (see [9] for complete details).
The echo's delay time T decreases as the source position ~ moves from 0 to 90
degrees. A similar analysis can be made if the source moves in the opposite direction,
and the reflection is caused by the other reflection surface 8 2 ? Since the reflection
path is longer for reflection surface 8 2 than for reflection surface 8 1 , the echo's delay
time can be segmented into two ranges. Therefore, the echo's delay time encodes
the source's directions in a one-to-one mapping relation.
In the setup, an Earthworks M30 microphone and Labl amplifier were used to record
and amplify the sound signals [3]. For this preliminary study of monaurallocalization, we have chosen to localize simple impulse sounds generated through speakers
and therefore can drop the silicon cochlea from our model. In the future, more
complicated signals, such as speech, will require a silicon cochlea implementation.
Inspired by ideas from visual processing, onset detection is used 'to segment sounds
[10]. The detection of an onset is produced by first taking the difference of two
first-order, low-pass filters given by [10]
OCt, k, r)
=
lot Iz(t - x, k)s(x)dx -lot Iz(t - x, k/r)s(x)dx
where r>l, k is a time constant, sex) is the input sound signal, and /z(x, k)
kexp(-kx).
(3)
=
A hardware implementation of the above equation is depicted in Figure 2a. In our
model, sound signals from the special reflection surface microphone are fed into
two low-pass filters which have different time constants determined by two bias
695
A Neuromorphic Monaural Sound Localizer
Vref2
Figure 3: Adaptive threshold circuit used to remove unwanted reflections.
A(t- 't)
A(t-2 't)
A(t-m't)
A(t)
!,,
t
Dl
t
02
+
03
Om
Figure 4: Neural signal processing model
voltages V Onb1 and V onb2 . The bias voltage V onbS determines the amplification of the
difference. The output of the onset detecting circuit is Vonouc. The onset detection
circuit determines significant increases in the signal energy and therefore segments
sound events. By computing the delay time between two sound events (direct
sound and its echo caused by the reflection surface), the system is able to decode
the source's direction. Each sound event is then transformed into a fixed-width
pulse so that the delay time can be computed with binary autocorrelators.
The fixed-width pulse generating circuit is depicted in Figure 2b. The pulse generating circuit includes a self-resetting neuron circuit [8] that controls the pulse duration
based on the bias voltage Vneubs' As discussed above, an appropriate threshold is
required to discriminate sound events from noise. One input of the pulse generating
circuit is the output of the onset detecting signal, Vonouc. vthreBh is set properly in
the pulse generating circuit in order to generate a fixed width pulse when Vonouc
exceeds vthreBh. Unfortunately the system may be confused by unwanted sound
events due to extraneous reflections from the desks and walls. However, since we
know the expect range of echo delays, we can inhibit many of the environmental
echoes that fall outside this range using an adaptive threshold circuit.
In order to cancel unwanted signals, we need to design an inhibition mechanism
which suppresses signals arriving to our system outside of the expected time range.
This inhibition is implemented in Figure 3. As the pulse generating circuit detects
the first sound event (which is the direct sound signal), the threshold becomes high
in a certain period of time to suppress the detection of the unwanted reflections (not
from our reflection surfaces). The input of the adaptive threshold circuit is Vneuouc
which is the output of the pulse generating circuit. The output of the threshold
circuit is vthreBh which is the input of the pulse generating circuit. When the pulse
generating circuit detects a sound event, Vneuouc becomes high, which increases
vthreBh from V re / 2 to V re / 1 as shown in Figure 3. The higher vthreBh suppresses the
detection. The suppression time is determined by the other self-resetting neuron
circuit.
1. G. Harris. c.-1. Pu and 1. C. Principe
696
2 . G5':2
,,=-
2 . 551 :-
2 . 't 51 ;-
241:2 . )51 ~
2 It;-,
,
I
I
i
I
I
,
,
i
,
5 ? ~
.,.
~"r
qs.?
I
I
~
... ~ .
I
,
,
j
,
,
,
Ii,
, -;
:: 'i'f 'i5'U N2 no
~ ~
t
j
rJ!
:.: t-~.-- -~ - -:-~ It ----:------.--4~,=-=-=-~
"~~--~~~~~~~--~
,,:,::. l . ..
II
I! . ..? . ?
1
~~~~T~~L~~~~~~~~~~r.7~~
o.
TH~[
( LBO
2? i511"
Figure 5: (a) The input sound signal: impulse signal recorded in typical office environment (b) HSPICE simulation of the output of the detecting onset circuit (label
61), the output of the pulse generating circuit {label 12), and the adaptive threshold
circuit response (label 11)
The nervous system likely uses a running autocorrelation analysis to measure the
time delay between signals. The basic neural connections are shown in Figure 4 [7].
A(t) is the input neuron, A(t - r), A(t - 2r), ... A(t - mr) is a delay chain. The
original signal and the delayed signal are multiplied when A(t) and A(t - kr) feed
Ck. Assuming the state of neuron A is NA(t). H each synaptic delay in the chain
is r, the chain gives us NA(t) under various delays. Ck fires simultaneously when
both A(t) and A(t - kr) fire. Neuron Ck connects neuron Dk. Excitation is built
up at Dk by the charge and discharge of Ck' The excitation at Dk is therefore
(4)
Viewing the arrangement of Figure 4 as a neuron autocorrelator, the time-varying
excitation at Db D2, .. Dk provides a spatial representation of the autocorrelation
function. The localization resolution of this system depends on the delay time r,
and the number of the correlators. As r decreases, the localization resolution is
improved provided there are enough correlators. In this paper, 30 unit delay taps,
and 10 correlators have been implemented on chip. The outputs of the 10 correlators
display the time difference between two sound events. The delay time decodes the
source's direction. Therefore, the 10 correlators provide a unit encoding of the
source location in the ID plane.
3
Simulation and Measurement Results
The complete system has been successfully simulated in HSPICE using database we
have recorded. Figure 5(a) shows the input sound signal which is an impulse signal
recording in our lab (a typical student office environment). Figure 5(b) shows the
output of the onset detector (labeled 61), the pulse generating output (labeled 12),
and the adaptive threshold (labeled 11). When the onset output exceeds the threshold, the output of the pulse generating circuit becomes high. Simultaneously, the
high value of the generated pulse turns on the adaptive threshold circuit to increase
the threshold voltage. The adaptive threshold voltage suppresses the unwanted re-
A Neuromorphic Monaural Sound Localizer
697
Reflection Surface
,
"
LED 1
dl
? a1
M30
Speaker
~
o LED 2
Labl
Amp.
Localizer
Chip
d2
- - -- >
LED 3
LED 4
Speaker
moving
direction
Figure 6: Block diagram of the test setup
flection which can be seen right after the direct signal (we believe the unwanted
reflection is caused by the table). Further simulation results are discussed in [9].
The single microphone sound localizer circuit has been fabricated through the MOSIS 2J.'m N-well CMOS process. Impulse signals are played through speakers to
test the fabricated localizer chip. Figure 6 depicts the block diagram of the test
setup. The M30 microphone picks up the direct impulse signal and echoes from the
reflection surface. Since the reflection surface in our test is just a single flat surface,
localization is only tested in one-half of the ID plane. The composite signals are
fed into the input of the sound localizer after amplification. Our sound localizer
chip receives the composite signal, computes the echo time delay, and sends out the
localization result to a display circuit. The display circuit is composed of 4 LEDs
with each LED representing a specific sound source location. The sound localizer
sends the computational result to turn on a specific LED signifying the echo time
delay. In the test, the M30 microphone and the reflection surface are placed at fixed
locations. The speaker is moved along the dotted line shown in Figure 6. The M30
microphone is d1 (33cm) from the reflection surface and al (24cm) from the speaker
moving line. The speaker's location is defined as ch as depicted in Figure 6.
Figure 7(a) shows the theoretical echo's delay at various speaker locations. Figure 7(b) is the measurement of the setup depicted in Figure 6. The y-axis indicates
LED 1 through LED 4. The x-axis represents the distance between the speaker's
location (ch in Figure 6). The solid horizontal line in Figure 7(b) represents the
theoretical results for which LED should respond for each displacement. The results
show that localization is accurate within each region with possibilities of two LEDs
responding in the overlap regions.
4
Conclusion
We have developed the first monaural sound localization system. This system provides a real-time model for human sound localization and has potential use in such
applications as low-cost teleconferencing. More work is needed to further develop
the system. We need to characterize the accuracy of our system and to test more
interesting sound signals, such as speech. Our flat reflection surface is straightforward and simple, but it lacks sufficient flexibility to encode the source's direction in
more than a I-D plane. We plan to replace the flat surfaces with a more complicated
surface to provide more reflections to encode a richer set of source directions.
698
J G. Harris, c.-J Pu and J C. Principe
1
0
10
20
30
40
80
50
sound so ..... d i _ f r o m _ (em)
70
80
localizer clip ...............
4
00
fil3
....0
c:
9
9
9
9
r
9
EJ 0
0
C,l
....
0
9
99
9
e
00
e
1
0
9 9
10
20
30
40
50
80
sound so..... di_lrom _ _ _ (em)
e
e
70
0
80
Figure 7: Sound localizer chip test result
Acknowledgments
This work was supported by an ONR contract #NOOOI4-94-1-0858 and an NSF CAREER award #MIP-9502307. We gratefully acknowledge MOSIS chip fabrication
and Earthworks Inc. for loaning the M30 microphone and amplifier.
References
[1] D. W. Batteau. The role of the pinna in human localization. Proc. R. Soc.
London, Ser. B, 168:158-180,1967.
[2] Neal A. Bhadkamkar. Binaural source localizer chip using subthreshold analog
cmos. In Proceeding of JCNN, pages 1866-1870, 1994.
[3] Earthworks, Inc., P.O. Box 517, Wilton, NH 03086. M90 Microphone.
[4] y. Hiranaka and H. Yamasaki. Envelop representations of pinna impulse responses relating to three-dimensional localization of sound sources. J. Acoust.
Soc. Am., 73:29, 1983.
[5] E. Knudsen, G. Blasdel, and M. Konishi. Mechanisms of sound localization in
the barn owl (tyto alba). J. Compo Physiol, 133:13-21, 1979.
[6] J. Lazzaro and C. A. Mead. A silicon model of auditory localization. Neural
Computation, 1:47-57, 1989.
[7] J.C. Licklider. A duplex theory of pitch perception. Experientia, 7:128-133,
1951.
[8] C. Mead. Analog VLSJ and Neural Systems. Addison-Wesley, 1989.
[9] Chiang-Jung Pu. A neuromorphic microphone for sound localization. PhD
thesis, University of Florida, Gainesville, FL, May 1998.
[10] L.S. Smith. Sound segmentation using onsets and offsets. J. of New Music
Research, 23, 1994.
[11] A.J. Watkins. Psychoacoustical aspects of synthesized vertical locale cues. J.
Acoust. Soc. Am., 63:1152-1165, 1978.
| 1489 |@word version:1 middle:1 sex:1 d2:2 pulse:19 simulation:4 gainesville:2 azimuthal:1 pick:1 solid:1 amp:1 dx:2 john:1 physiol:1 shape:1 remove:1 designed:2 drop:1 cue:4 half:1 nervous:1 plane:4 smith:1 record:1 chiang:2 compo:1 provides:4 detecting:5 location:6 lbo:1 along:1 direct:7 become:1 viable:1 interaural:3 compose:1 ray:1 autocorrelation:2 expected:1 behavior:1 inspired:2 detects:3 actual:3 becomes:3 confused:1 provided:1 circuit:27 duplex:1 tyto:1 cm:2 suppresses:3 developed:1 acoust:2 fabricated:3 locale:1 charge:1 unwanted:6 ser:1 control:1 unit:2 negligible:1 engineering:1 encoding:1 id:3 mead:2 path:4 studied:1 hspice:2 limited:1 range:6 unique:1 acknowledgment:1 atx:1 block:2 displacement:1 composite:2 amplify:1 demonstrated:1 straightforward:1 duration:1 resolution:3 q:1 d1:2 konishi:1 discharge:1 decode:1 us:2 pinna:7 mic:1 database:1 labeled:3 role:3 electrical:1 solved:1 region:2 decrease:2 inhibit:1 environment:2 segment:2 serve:1 localization:24 completely:1 teleconferencing:1 binaural:1 easily:1 chip:7 various:2 derivation:1 describe:1 effective:1 london:1 detected:1 outside:2 richer:1 larger:1 echo:19 propose:1 flexibility:1 achieve:1 amplification:2 moved:1 amplifies:1 r1:1 produce:1 generating:14 cmos:3 develop:1 received:1 soc:3 implemented:2 skip:1 direction:13 filter:4 subsequently:1 human:7 viewing:1 owl:1 require:1 wall:1 preliminary:1 elevation:3 biological:1 experientia:1 barn:1 mapping:2 blasdel:1 circuitry:4 a2:1 proc:1 shadowing:1 label:3 successfully:2 ck:4 ej:1 varying:1 voltage:5 office:2 encode:2 derived:1 properly:1 indicates:1 suppression:1 am:2 dependent:2 integrated:1 vlsi:2 relation:1 transformed:1 extraneous:1 plan:1 spatial:2 special:3 shaped:1 having:1 biology:1 represents:2 cancel:1 future:1 composed:4 simultaneously:2 delayed:1 phase:2 geometry:1 connects:1 fire:2 envelop:1 amplifier:2 detection:5 possibility:1 introduces:1 chain:3 accurate:1 m30:6 re:3 mip:1 theoretical:2 neuromorphic:6 cost:1 delay:21 fabrication:1 azimuth:1 characterize:1 varies:2 adaptively:1 contract:1 off:1 na:2 thesis:1 ear:11 recorded:3 worse:1 external:5 account:1 potential:1 student:1 includes:1 alba:1 inc:2 caused:5 depends:2 onset:13 tion:1 lot:2 lab:1 analyze:1 complicated:2 om:1 il:1 accuracy:1 resetting:2 subthreshold:1 decodes:2 produced:1 researcher:1 detector:1 yamasaki:1 synaptic:1 energy:2 frequency:4 auditory:1 noooi4:1 torso:2 segmentation:1 amplitude:1 reflecting:1 feed:1 wesley:1 ta:1 higher:1 reflected:1 response:2 improved:1 box:1 just:1 receives:1 horizontal:2 lack:1 somehow:1 impulse:6 believe:1 effect:1 bhadkamkar:1 inspiration:1 neal:1 width:3 self:2 ambiguous:1 speaker:9 excitation:3 complete:3 reflection:29 physical:1 nh:1 analog:5 he:1 discussed:2 relating:1 synthesized:1 measurement:3 blocked:1 significant:2 silicon:4 refer:1 gratefully:1 moving:3 longer:2 surface:22 inhibition:2 pu:5 certain:1 psychoacoustical:1 binary:1 success:1 onr:1 seen:1 mr:1 period:1 signal:31 ii:2 multiple:1 sound:70 rj:1 segmented:1 exceeds:2 long:1 award:1 a1:2 pitch:1 basic:1 cochlea:3 addition:1 diagram:3 source:24 concluded:1 sends:2 specially:2 hz:2 recording:2 db:1 correlators:5 enough:3 opposite:1 inner:1 idea:1 attenuated:1 jcnn:1 ltd:2 speech:2 lazzaro:1 latency:1 amount:1 desk:1 band:1 hardware:1 clip:1 generate:1 sl:1 nsf:1 dotted:2 li2:1 iz:2 threshold:13 localize:5 verified:1 mosis:2 jose:1 i5:1 respond:1 diffraction:2 fl:2 played:1 display:3 kexp:1 flat:5 encodes:2 aspect:1 speed:1 department:1 tv:2 em:2 lid:1 equation:1 discus:1 turn:2 mechanism:2 needed:1 know:1 addison:1 fed:2 serf:1 multiplied:1 appropriate:1 florida:2 original:3 responding:1 running:1 music:1 move:2 arrangement:1 g5:1 distance:4 mapped:1 simulated:1 earthwork:3 assuming:1 length:2 setup:5 unfortunately:1 a2x:1 suppress:1 implementation:2 design:1 vertical:2 neuron:7 acknowledge:1 knudsen:1 shoulder:2 head:6 monaural:9 intensity:3 introduced:1 required:1 connection:1 tap:2 autocorrelator:1 able:1 below:1 perception:1 built:1 event:8 overlap:1 representing:1 axis:2 hrtf:2 expect:1 interesting:1 degree:1 sufficient:1 thresholding:1 bank:1 licklider:1 jung:2 placed:2 supported:1 arriving:1 dis:1 bias:3 allow:1 fall:1 taking:2 dimension:1 valid:1 computes:2 made:1 adaptive:8 simplified:1 far:1 emphasize:1 incoming:1 table:1 transfer:3 career:1 s2:1 noise:1 arise:1 n2:1 depicts:1 localizer:14 position:3 watkins:1 specific:2 r2:2 dk:4 offset:1 dl:2 albeit:1 kr:2 phd:1 magnitude:1 direc:1 kx:1 depicted:5 yin:1 led:11 wavelength:3 likely:1 visual:1 expressed:1 ch:2 determines:2 relies:1 harris:4 extracted:1 environmental:1 oct:1 replace:1 change:1 specifically:1 determined:2 typical:2 principal:1 microphone:14 pas:4 discriminate:1 principe:4 signifying:1 tested:1 |
535 | 149 | 57
Self Organizing Neural Networks for the
Identification Problem
Manoel Fernando Tenorio
VVei-Tsih Lee
School of Electrical Engineering
Purdue University
School of Electrical Engineering
Purdue University
VV. Lafayette, UN. 47907
VV. Lafayette, UN. 47907
lwt@ed.ecn.purdue.edu
tenoriQ@ee.ecn.purdue.edu
ABSTRACT
This work introduces a new method called Self Organizing
Neural Network (SONN) algorithm and demonstrates its use in a
system identification task. The algorithm constructs the network,
chooses the neuron functions, and adjusts the weights. It is compared to
the Back-Propagation algorithm in the identification of the chaotic time
series. The results shows that SONN constructs a simpler, more
accurate model. requiring less training data and epochs. The algorithm
can be applied and generalized to appilications as a classifier.
I. INTRODUCTION
1.1 THE SYSTEM IDENTIFICATION PROBLEM
In various engineering applications, it is important to be able to estimate, interpolate,
and extrapolate the behavior of an unknown system when only its input-output pairs are
available. Algorithms which produce an estimation of the system behavior based on these
pairs fall under the category of system identification techniques.
1.2 SYSTEM IDENTIFICATION USING NEURAL
NETWORKS
A general form to represent systems, both linear and nonlinear, is the KolmogorovGarbor polynomial tGarbor. 19611 shown below:
y = ao +
L aixi + L L aijxiXj + ...
1
i
J
(1)
58
Tenorio and Lee
where the y is the output. and x the input to the system. [Garbor .1961] proposed a
learning method that adjusted the coefficient of (1) by minimizing the mean square error
between each desired output sample and the actual output
This paper describes a supervised learning algorithm for structure construction and
adjustment Here. systems which can be described by (1) are presented. The computation
of the function for each neuron performs a choice from a set of possible functions
previously assigned to the algorithm. and it is general enough to accept a wide range of
both continuous and discrete functions. In this work. the set is taken from variants of the
2-input quadratic polynomial for simplicity. although there is no requirement making it
so. This approach abandons the simplistic mean-square error for perfonnance measure in
favor of a modified Minimum Description Length (MOL) criterion [Rissanen,1975]. with
provisions to measure the complexity of the model generated. The algorithm searches for
the simplest model which generates the best estimate. The modified MDL. from hereon
named the Structure Estimation Criterion (SEC). is applied hierarchically in the selection
of the optimal neuron transfer function from the function set. and then used as an optimal
criterion to guide the construction of the structure. The connectivity of the resulting
structure is arbitrary. and under the correct conditions [Geman&Geman. 84] the estimation
of the struCture is optimal in tenns of the output error and low function complexity. This
approach shares the same spirit of GMDH-type algorithms. However, the concept of
parameter estimation from Information Theory. combined with a stochastic search
algorithm - Simulated Annealing. was used to create a new tool for system identification.
This work is organized as follows: section II presents the problem formulation and the
Self Organizing Neural Network (SONN) algorithm description; section III describes the
results of the application of SONN to a well known problem tested before using other
neural network algorithms [Lapede8&Farber. 1987; Moody. 1988]; and fmally, section IV
presents a discussion of the results and future directions for this work.
II. THE SELF ORGANIZING NEURAL NETWORK
ALGORITHM
11.1 SELF ORGANIZING STRUCTURES
The Self Organizing Neural Network (SONN) algorithm performs a search on the model
space by the construction of hypersurfaces. A network of nodes. each node representing a
hypersurface. is organized to be an approximate model of the real system. SONN can be
fully characterized by three major components. which can be modified to incorporate
knowledge about the process: (1) a generating rule of the primitive neuron transfer
functions. (2) an evaluation method which accesses the quality of the model. and. (3) a
structure search strategy. Below. the components of SONN are discussed.
ll.2 THE ALGORITHM STRUCTURE
Self Organizing Neural Networks
11.2.1 The Generating Rule
Given a set of observations S:
S
= {(Xl,
Yl),(Xl, Yl)",,(XI, YI)}
Yi =f(XV + 11
generated by
(2)
where f(.) is represented by a Kolmogorov-Garbor polynomial. and the random variable
11 is nonnally distributed. N(O.l). The dimensions of Y is m. and the dimensions of X is
n. Every component Yk of Y fonns a hypersurface Yk = fk(X) in the space of dim (X) +
1. The problem is to fmd f(.). given the observations S. which is a corrupted version of
the desired function. In this work. the model which estimates f(.) is desired to be as
accurate and simple (small number of parameters. and low degree of non linearity) as
possible.
The approach taken here is to estimate the simplest model which best describes f(.) by
generating optimal functions for each neuron. which can be viewed as the construction of
a hypersurface based on the observed data. It can be described as follows: given a set of
observations S; use p components of the n dimensional space of X to create a
hypersurface which best describes Yk =f(X). through a three step process. First, given X
= [xl' x2' x3' .... xn) and Yk' and the mapping '? n: [Xl' x2' x3' .... Xn) -> [x'?(1)'
x'?(2)' x,?(3)' .... x'?(n?)' construct the hypersurface hi (x'?(1)' x'?(2)' x,?(3)' ....
x'?(n? (hi after the fIrst iteration) of p+ 1 dimensions. where '?n is a projection from n
dimensions to p dimensions. The elements of the domain of '?n are called tenninals.
Second. If the global optimality criterion is reached by the construction of hi(x'?(l)'
x'?(2)' x,?(3)' .... x'?(n?' then stop. otherwise continue to the third step. Thud.
generate from [Xl' x2' x3' .... xn.hl(x'?(l)' x'?(2)' x,?(3)' .... x'?(n?) a new p+l
dimensional hypersurface hi+ I through the extended mapping '?n+ 1(.). and reapply the
second step.The resulting model is a multilayered neural network whose topology is
arbitrarily complex and created by a stochastic search guided by a structure estimation
criterion. For simplicity in this work. the set of prototype functions (F) is restricted to be
2-input quadratic surfaces or smaller. with only four possible types:
y
y
= 8o+alxl +a2x 2
= 8o+alxl +a2x2+a3 x l x2
(3)
(4)
Y = 3o+a l x l+ a2x1
(5)
Y = 8o+alxl+a2x2+a3xlx2+~x1+a5x~
(6)
11.2.2 Evaluation or the Model Based on the MDL Criterion
The selection rule (T) of the neuron transfer function was based on a modifIcation of the
Minimal Description Length (MOL) information criterion. In [Rissanen. 1975] the
principle of minimal description for statistical estimation was developed. The MDL
provides a trade-off between the accuracy and the complexity of the model by including
the structure estimation tenn of the fInal model. The final model (with the minimal
59
60
Tenorio and Lee
MOL) is optimum in the sense of being a consistent estimate of the number of
parameters while achieving the minimum error [Rissanen.1980]. Given a sequence of
observation xl,x2,x3 ?...?xN from the random variable X. the dominant tenn of the MDL
in [Rissanen. 1975] is:
MDL =- log f(xI8) + 0.5 k log N
where f(xI8) is the estimated probability density function of the model. k is the number
of parameters. and N is the number of observations. The first tenn is actually the negative
of the maximum likelihood (ML) with respect to the estimated parameter. The second
term describes the structure of the models and it is used as a penalty for the complexity of
the model. In the case of linear polynomial regression. the MOL is:
MDL = - 0.5 N log S~ + 0.5 k log N
(8)
where k is the number of coefficients in the model selected.
In the SONN algorithm. the MDL criterion is modified to operate both recursively and
hierarchically. First. the concept of the MDL is applied to each candidate prototype
surface for a given neuron. Second. the acceptance of the node. based on Simulated
Annealing. uses the MDL measure as the system energy. However. since the new neuron
is generated from terminals which can be the output of other neurons. the original
defmition of the MDL is unable to compute the true number of system parameters of the
final function. Recall that due to the arbitrary connectivity. feedback loops and other
configurations it is non trivial to compute the number of parameters in the entire
structure. In order to reflect the hierarchical nature of the model. a modified MDL called
Structure Estimation Criterion (SEC) is used in conjunction with an heuristic estimator
of the number of parameters in the system at each stage of the algorithm. A
computationally efficient heuristic for the estimation of the number of parameters in the
model is based on the fact that SONN creates a tree-like structure with multiple roots at
the input terminals. Then k. in expression (8). can be estimated recursively by:
k = kL + kR + (no. of parameters of the current node)
(9)
where kL and kR are the estimated number of parameters of the left and right parents of
the current node. respectively. This heuristic estimator is neither a lower bound nor an
upper bound of the true number of parameter in the model.
11.2.3 The SONN Algorithm
To explain the algorithm. the following definitions are necessary: Node - neuron and the
associated function. connections. and SEC; BASIC NODE - A node for the system input
variable; FRONT NODE - A node without children; IN1ERMIDIATE NODE - The nodes
that are neither front or basic nodes; STATE - The collection of nodes. and the
configuration of their interconnection; INITIAL STATE (SO - The state with only basic
nodes; PARENT AND CHILD STATE - The child state is equal to the parent state except
for f a new node and its interconnection generated on the parent state structure;
NEIGHBOR STATE - A state that is either a child or a parent state of another; ENERGY
Self Organizing Neural Networks
OF THE STATE (SEC-Si) - The energy of the state is defined as the minimum SEC of
all the front nodes in that state.
In the SONN algorithm. the search for the correct model structure is done via Simulated
Annealing. Therefore the algorithm at times can accept partial structures that look less
than ideal. In the same way. it is able to discard partially constructed substructures in
search for better results. The use of this algorithm implies that the node accepting rule
(R) varies at run-time according to a cooling temperature
schedule. The SONN
algorithm is as follows:
m
Initialize T, and S[
Repeat
Repeat
Sj = generate (Si),
If accept ( SEC_Sj. SEC_Si, T) then Si = Sj.
- application of P.
- application ofR.
WUiI the number of new neurons is greater than N.
Decrease the temperature T.
until The temperature T is smaller than tend (Terminal temperature for Simulated
Annealing).
Each neuron output and the system input variables are called terminals. Tenninals are
viewed as potential dimensions from which a new hypersurface can be constructed. Every
tenninal represents the best tentative to approximate the system function with the
available infmnatioo. and are therefore treated equally.
lll. EXAMPLE - THE CHAOTIC TIME SERIES
In the following results. the chaotic time series generated by the Mackay-Glass
differential equations was used. The SONN with the SEC. and its heuristic variant were
used to obtain the approximate model of the system. The result is compared with those
obtained by using the nonlinear signal processing method [LapedeS&Farber. 1987] . The
advantages and disadvantages of both approaches are analyzed in the next section.
111.1 Structure of the Problem
The MacKay-Glass differential equation used here can be described as:
dX(t)
at
=
a x(t - t) _ b x(t)
1 + x10(t - t)
(10)
By setting a =0.2. b =0.1. and t = 17. a chaotic time series with a' strange attractor of
fractal dimension about 3.5 will be produced [Lapedes&Farber. 1987] . To compare the
accuracy of prediction the nonnalized root mean square error is used as a perfonnance
index:
61
62
Tenorio and Lee
no
RMSE
nnalized RMSE - Standard Deviation
(ll)
111.2. SONN WITH THE HEURISTIC SEC (SONN.H)
In the following examples, a modified hewistic version of the SEC is used. The estimator
of the number of parameters is given by (9), and the fmal configuraion is shown in figure
1.
111.2.1 Node 19
In this subsection, SONN is allowed to generate up to the 19th accepted node. In this
first version of the algorithm, all neurons have the same number of interconnections. and
therefore draw their transfer function from the same pool of functions .. Generalizations of
the algorithm can be easily made to accommodate multiple input functions, and neuron
transfer function assignment being drawn from separate pools. In this example, the first
one hundred points of the time series was used for training, and samples 101 through 400
used for prediction testing. The total number of weights in the network is 27. The
performance index average 0.07. The output of the network is overlapped in the figure 2
with the original time series.
For comparison purposes, a GDR network with the structure used in [LapedeS&Farber,
1987] is trained for 6500 epochs. The training data consisted of the first 500 points of the
time series, and the testing data ran from the 501st sample to the 832nd. The total
number of weights is 165. and the fmal performance index equal to 0.12. This was done
to give both algorithms similar computational resources. Figure 3 shows the original
time series overlapped with the GDR network output.
Ill.2.2 NODE 37
In this subsection, the model chosen was formed by the 37th accepted node. The network
was trained in a similar manner to the flfSt example, sioce it is part of the same run. The
final number of weights is 40, and the performance index 0.018. Figure 4 shows the
output of the network overlapped with the original time series. Figure 5 shows the GDR
with 11,500 epochs. Notice that in both cases, the GDR network demands 150
connections and 150 weights. as compared to 12 connections and 27 weights for the first
example and 10 connections and 40 weights for the second example. The comparison of
the performance of different models is listed in figure 6.
IV. Conclusion and Future Work
In this study, we proposed a new approach for the identification problem based on a
flexible, self-organizing neural network (SONN) structure. The variable structure provides
the opportunity to search and construct the optimal model based on input-output
observations. The hierarchical version of the MDL, called the structure estimation criteria,
Self Organizing Neural Networks
was used to guide the trade-off between the model complexity and the accuracy of the
estimation. The SONN approach demonstrates potential usefulness as a tool for system
identification through the example of modeling a chaotic time series.
REFERENCE
Garber, D .? eL al. ,"A universal nonlinear filter, predicator and simulator which
optimizes itsekf by a learning process," IEE Proc.,18B, pp. 422-438, 1961
RissanenJ. "Modeling by shortest data description," Automatica. vo1.14, pp. 465471.1975
Gemen, S, and Gernen D., "Stochastic relaxation, gibbs deisribution, and the bayesian
restoration of images." IEEE PAMI.? PAMI-6,pp.721-741. 1984
Lapedes.A. and Farber, R. ,"Nonlinear signal processing using neural networks:
Predication and system modeling," TR.. LA-UR-87-2662. 1987
Moody. J. This volume
Rissanen:,,J. "Consistent order estimation of autoregressive processing by shortest
description of data." Analysis and optimization of stochastic system. Jacobs et. al. Eds.
N.Y. Academic. 1980
Figure 1. The 37th State Generated
--------_ _ - .._---_.
.. ..
S"s._IO.?SOHNI_ 191
0>
o.
o.
,
??.f..---.....---....---__.-----.------I
,.,.
,IIA
'.0
0'"
'.
Figure 2. SONN 19th Model. P.I. = 0.06
63
64
Tenorio and Lee
..------------ -- ------.
.-
+----~-- - _-
.- ~
-.
--l
nn.
~nn
'lft
Figure 3. GDR after 6.500 Epochs. P.I.
no
=0.12
".0
'. 00
=0.038
Figure 4. SONN 37th Model. P.I.
~------------------<; ??1_ 10.. "ICk P'''Dlq1I!O" -III II 500 EDOCM. ------
.
..
,\
~
r r
\ Ii \ I
,
,
,
.0
\
i
. no
l\'
~ ~o
-.
. no
,?...-()t,...
?,-...C)o . . . .
OOD
Figure 5. GDR after 11.500 Epochs. P.I. = 0.018
Comparison ollhe Perform alice Index
014~-------------------------.
W'
I>:'UU t: I'''~''~
0.12
~Vt~tllfl""Uaj l~,
0. 10
SOliN (o .. ~j ~7)
BP \ I :;W Ep~I"
002
.,.-._._.;,_.", . . . . ...
.o. _._._ . _~ .. _._._._._.- . -.~.,..~
000
.............. ........ . ......... ..... . ...................... .... -_.... ..... .. ... .-. .. .. ..... .
~
~
002
_ _ _ _ _ _ _- - - - - - - - ' - - -
12u
Figure 6. Perfonnance Index Versus the Number of Predicted Points
| 149 |@word version:4 polynomial:4 nd:1 jacob:1 tr:1 accommodate:1 recursively:2 initial:1 configuration:2 series:10 lapedes:4 current:2 si:3 dx:1 tenn:3 selected:1 accepting:1 provides:2 node:22 simpler:1 constructed:2 differential:2 ood:1 manner:1 behavior:2 nor:1 simulator:1 terminal:4 actual:1 lll:1 linearity:1 developed:1 lft:1 every:2 demonstrates:2 classifier:1 before:1 engineering:3 xv:1 io:1 pami:2 alice:1 range:1 lafayette:2 testing:2 x3:4 chaotic:5 universal:1 projection:1 selection:2 primitive:1 simplicity:2 adjusts:1 rule:4 fmally:1 estimator:3 construction:5 aixi:1 us:1 overlapped:3 element:1 cooling:1 geman:2 observed:1 ep:1 electrical:2 trade:2 decrease:1 yk:4 ran:1 complexity:5 trained:2 creates:1 easily:1 various:1 represented:1 kolmogorov:1 whose:1 heuristic:5 garber:1 otherwise:1 interconnection:3 favor:1 abandon:1 final:4 sequence:1 advantage:1 xi8:2 loop:1 organizing:10 description:6 gmdh:1 parent:5 requirement:1 optimum:1 produce:1 generating:3 school:2 sonn:20 predicted:1 implies:1 uu:1 direction:1 guided:1 farber:5 correct:2 filter:1 stochastic:4 ao:1 generalization:1 adjusted:1 mapping:2 major:1 purpose:1 estimation:12 proc:1 create:2 tool:2 ick:1 modified:6 conjunction:1 likelihood:1 sense:1 glass:2 dim:1 el:1 nn:2 entire:1 accept:3 ill:1 flexible:1 initialize:1 mackay:2 equal:2 construct:4 represents:1 look:1 future:2 interpolate:1 attractor:1 acceptance:1 evaluation:2 mdl:12 introduces:1 nonnally:1 analyzed:1 accurate:2 partial:1 necessary:1 perfonnance:3 tree:1 iv:2 desired:3 minimal:3 a2x2:2 modeling:3 disadvantage:1 restoration:1 assignment:1 deviation:1 hundred:1 usefulness:1 front:3 iee:1 varies:1 corrupted:1 chooses:1 combined:1 st:1 density:1 lee:5 yl:2 off:2 pool:2 moody:2 connectivity:2 reflect:1 tenninals:2 potential:2 sec:8 coefficient:2 root:2 reached:1 substructure:1 rmse:2 square:3 formed:1 accuracy:3 identification:9 bayesian:1 tenninal:1 produced:1 explain:1 ed:2 definition:1 energy:3 pp:3 associated:1 stop:1 recall:1 knowledge:1 subsection:2 provision:1 organized:2 schedule:1 actually:1 back:1 supervised:1 formulation:1 done:2 fmal:2 stage:1 until:1 nonlinear:4 propagation:1 quality:1 requiring:1 true:2 concept:2 consisted:1 assigned:1 ll:2 self:10 criterion:10 generalized:1 performs:2 hereon:1 temperature:4 image:1 ecn:2 volume:1 discussed:1 defmition:1 gibbs:1 fk:1 iia:1 access:1 surface:2 dominant:1 optimizes:1 discard:1 arbitrarily:1 continue:1 tenns:1 vt:1 yi:2 ofr:1 minimum:3 greater:1 fernando:1 shortest:2 signal:2 ii:4 multiple:2 x10:1 characterized:1 academic:1 equally:1 prediction:2 variant:2 simplistic:1 regression:1 fonns:1 basic:3 iteration:1 represent:1 annealing:4 operate:1 tend:1 spirit:1 ee:1 ideal:1 iii:2 enough:1 topology:1 prototype:2 expression:1 flfst:1 penalty:1 fractal:1 listed:1 category:1 simplest:2 generate:3 notice:1 estimated:4 discrete:1 four:1 rissanen:5 achieving:1 drawn:1 neither:2 relaxation:1 run:2 named:1 strange:1 draw:1 bound:2 hi:4 quadratic:2 bp:1 x2:5 generates:1 optimality:1 according:1 describes:5 smaller:2 ur:1 making:1 modification:1 hl:1 restricted:1 taken:2 computationally:1 resource:1 equation:2 previously:1 available:2 hierarchical:2 reapply:1 original:4 opportunity:1 predicator:1 strategy:1 unable:1 separate:1 simulated:4 trivial:1 length:2 index:6 minimizing:1 negative:1 a2x:1 unknown:1 perform:1 upper:1 neuron:14 observation:6 purdue:4 predication:1 solin:1 extended:1 nonnalized:1 arbitrary:2 pair:2 kl:2 connection:4 tentative:1 manoel:1 able:2 below:2 including:1 treated:1 representing:1 created:1 thud:1 epoch:5 uaj:1 aijxixj:1 fully:1 versus:1 degree:1 consistent:2 principle:1 share:1 repeat:2 guide:2 vv:2 fall:1 wide:1 neighbor:1 distributed:1 feedback:1 dimension:7 xn:4 autoregressive:1 collection:1 made:1 hypersurface:7 sj:2 approximate:3 ml:1 global:1 automatica:1 xi:1 un:2 continuous:1 search:8 nature:1 transfer:5 mol:4 complex:1 domain:1 hierarchically:2 multilayered:1 child:4 allowed:1 fmd:1 x1:1 xl:6 candidate:1 third:1 a3:1 kr:2 gdr:6 demand:1 lwt:1 tenorio:5 adjustment:1 partially:1 hypersurfaces:1 viewed:2 except:1 vo1:1 called:5 total:2 accepted:2 la:1 incorporate:1 tested:1 extrapolate:1 |
536 | 1,490 | Very Fast EM-based Mixture Model
Clustering using Multiresolution kd-trees
Andrew W. Moore
Robotics Inst.i t. ut.e, Carnegie tl1plloll University
Pittsburgh , PA 15:21:3. a\\'l11'9.'cs.cll1u.eciu
Abstract
Clust ering is impor ta nt in m any fi elds including m a nufac tlll'ing ,
bio l og~' , fin a nce , a nd astronomy. l\Iixturp models a rp a popula r approach due to their st.atist.ical found a t.ions, and EM is a very popular l1wthocl for fillding mixture models. EM, however, requires
lllany accesses of the dat a , a nd thus h as been dismissed as impract ical (e .g. [9]) for d ata mining of enormous dataset.s. We present a
nt' \\? algorit.hm, baspd on thp l1lultiresolution ~.'Cl- trees of [5] , whi ch
dramatically reeluc ps th e cost of EtlI-baspd clusteriug , wit.h savings
rising linearl:; wit.h the numb er of datapoints. Although prespnt.pd
lwre for maximum likplihoocl estimation of Gaussian mixt.ure modf' ls , it. is also applicable to non-(~aussian models (provided class
d ensit.i es are monotonic in Mahala nobis dist.ance ), mixed categorical/ nUllwric clusters. anel Bayesian nwthocls such as Antoclass [1] .
1
Learning Mixture Models
In a Gaussian mixtur e lllod f' l (e.g. [3]) , we aSSUI1W t.hat d ata points {Xl .. . XR} ha\'p
bef'n ge lw r<lt ecl incl e p e ncl e lltl~ by the following process. For each X I in turn, natlll'f'
begius by randomly picking a class, c}' from a discrf' t e set of classf's {('I . . ' Cs }.
T lwn nat m e dr aws X I from an .II-dimension a l Gallss ia n whosf' m ea n fI i and cO\'a riallce ~i depe nd 0 11 th e class, Thus we have
.
wh ere 8 d en ot ps all the p ar ameters of the mixture: the class probabilities Vi (w lwre
P(Cj 18)) , the class centers fl j and th e class covariances ~j'
Vi =
Tlw job of a mixture m od el learn er is to find a good estimat e of t.he mod eL and
Expectation MaximizRtion (EM) , also known a::l "Fuzzy ~' -m e a n::l", i::l a popular
544
A. W Moore
algorit.hm for doing so. TIlt' Ith iteration of El\I begins \vith an estimatp (/ of tllP
model , and ends with all il11prO\'ed pstimate ll+1. 'Write
(2)
E;\I iteratps over parh point.-class combination, comput.ing for pach dass Cj and Pnch
datapoint Xi, thp pxtent to which Xi is "owned" by Cj. The ownership is simply
tl'i) = P(Cj I Xi, (/). Throughout this paper \\iP will use thp following notation :
(lij
P(Xi
Wlj
P(Cj
I Cj Jl)
I Xi , (i) = (/ ijJ!.d Lt~ l (/iI.. )JJ.. (by
Bayes ' Rule)
Then tl1P new value of thp cpntroid, J.ij' of the jt.b rlass in thp npw modpl (l+l is
sim ply tilt' \\"Pightpel t11pan of all the da t<'l point:,; , using the values {LV 1), W~j, . .. lIB.i }
as t.he weights. A similar weight.eel procedure gives the new est.imat.e's of the class
probabilities and the dass cov<'Iriances:
sw?
p' f- _
_
.J
.I
R
1 R
tt} f- - - ~ U:i}Xi
.
s\\"
.I
L
i= l
.
w)wre S\\'j = L~= l tl;ij . Thus each iteration of EM visit:-; ewry datapoint-rla:,;:,; pair.
meaning "YR evaluations of a l\l-dilllensional Ganssian, and so needing O(J/.!.SR)
arithnwtic operations ]wr iteration. This paper aims to reduce that cost.
An IIIrkd-tree (Multiresolution J~D-tree), introduced in [2] and developed further
in [5], is a binary tree in which each node is associateel wit.h a subset of the elatapoints. The root node owns all the datapoints. Each non-leaf-noelp has t \VO rhilelren.
d efineel by a splitting dimension NO.SPLITOIM and a splitting valup NO.sPLITVAL. The
two children divide their parent's datapoints between them , with the left child owing those dat.apoillts that are strictly less than the splitting \"alue in the splitting
dimension, and t he right child owning tllP remaindpr of the parent's cia t apoinb:
Xi E NO.LEFT
<=}
x i [No.sPLITDIM]
Xi E NO.RIGHT
<=}
xdNo .SPLlTOIM]
<
2:
l\O.SPLITVAL anel Xi E No
(4)
NO .SPLITVAL and Xi E ND
(.5)
The distinguishing featur e of mrkcl-trees is that their nodes contain the following:
? NO.NUMPOINTS: The number of points owned by No (equivalently, the average density in No).
? NO.CENTROID: The centroid of the points owned by No (equivalently, the
fir:st 1lI0ment of the density below ND) .
? No.(?ov: The cov<'lriance of the points owned by No (equivalently. the second
lI10lllent. of the clensi t.y below No).
? NO. HYPERRECT: The bounding hyper-rectangle of the points below No
\\"1' construct. mrkcl-trpes top-down, id ent.ifying tilt' bounding box of th e current.
node , and splitting in t.hf' cent er of t.he widest dimension. A node is d eclared to be
a lea L and is left unsplit , if the widest dimension of its bounding box is
SOIllP
threshold, JIB IV. If MB W is zero, t.hen all leaf nodes denote singleton or coincident
points , the tree bas O(R) nodes and so requires O(M~ R) memory, and (with some
care) the cons t ruction co ~t is O(",J'2 R+ M R log R). In practice , we :-;et MB IT' t.o 1%
of t he range of the datapoint components . The tree size and construction thus cost
:s
Very Fast EM-Based Mixture Model Clustering Using Multiresolution Kd-Trees
545
co ns icle ra bl," If'sS t.l Ja n these b Ollnc\s b f'ca use in c\ f' nsf' reg io ns , tin~' lea f n o d es we rf'
a ble t o summari ze cl ozf'lls of cl at apo ints , No t f' t oo th a t. t h e cost of trf'e-builcling is
a m ol' t izecl - th f' tree must I)f' built. o nef' , .ve t E ~I p f'l' fo rms m a llY it er a tio ns ,
To p e rform an it.eratio n ofE1\1 with tllf' IIIrkd- tl'f'e , we ca ll t.h e functio n l\ L\K ESTAT S
(d f'scril)f'c\ b e lo w) o n t. h f' root of t h e trf'f', \L-\K ESn,Ts ( No , tl) o u t puts :i N va lu es:
(S\\'l , S\\':! , ' , , SWx , SWX 1, ' , ,S"'X ,V, SWXX 1, , , ,S\\,XX N ) wh f're
I:::
I:::
Wij
tl 'ij X ,
SWXX j
=
x, E NO
x , E NO
((j )
X,
E NO
e + 1:
TI1f' res ults of 1\ 1A KES TATS (RooT) pro \'idf' sU ffi c if' nt. stat isti c8 to COIl St rue t
]I) f--- S \\'
t
,iI R
(I)
If l\ [AK ESTATS is ca llf'cl o n a If'a f n o c\ f' ,
\ \'P
s impl)' co mpu tf', fo r f'aeh j,
S
te)
= P (c) I x,e = P(x I t'j ,(l)p( c) I et)/ I::: P (x I Ck , et )p (Ck l et)
t )
(8)
1.' =1
=
whe rf' x
NO, ('ENTROID , a nd when ' a ll the it f'lI1s in tllf' rig ht h a nd
a r f' eas il~ ' co mput.ed , \Vf' thell r f't urn S\\'j = Il 'j X =" O,NTJ 1\IPOI NTS,
Il 'j x NO,N TI1\!POINTS x X a nd s \\'xx)
Il'j x );O,f\ Tli\!POINTS x No, co \' ,
SOil Wf' ca ll d o this is th a t, if the If'a f noclf' is W J') ' sm a ll, t.llf' rf' willlw lit tle
in te l ) fo r th f' po int s o \\'l lf'd by t. h f' nod f' a nd so , fo r f'xa lll p lf'
Il',j X i ~
III t h e eXpf'r ill lf'llts be lo w \\'e li se \'e r)' t iny lea f nod f' s , f' lIsurin g acc u rac,\',
=
L:
f' qu at io n
s \\,x) =
The rea va ri atlo n
11'.1 LX"
If \IAKEST.-\TS is ca ll e d o n a no n-If'a f-n odf', it (' a n easily co mputf' its a ns \Vf'r by
r f'c llrsivf' l~' ca llin g l\ IAK ESTAT s o n its tw o chilclrf' n and t.hf' n l'f' turning the :-:UIll of
tllf' t\Vo Sf' ts
was th f' end
co n venti o nal
O(R) n od es,
of a n s \\'f'l'S, In genf'r a l. th a t is f'xact ly how we will pro cef'd, If t h a t
of th e sto ry , W P wo uld h aw little computa tion a l improvel1lf' nt o \'er
E1\l, b ecRuse o n f' pass wo uld fully t.r al'E'rsf' t h e trf'f', which con ta ins
doing O( S .11:!) \\'o rk p er node ,
\Ve will win if we evf'l' sp ot th at. a t so m f' int f' rm edi ate no d f' , \\'f' ca n jll '1111 t , i,f',
e\'a lu a t f' th f' n o d e as if it were a If'a f. witho u t. sf' al't'hin g it.s d f' sce nd en ts , but witho u t.
introduc in g si gnifica nt e rro r int o tllf' compu tatio n,
T o do t hi:-: , \\,f' will CO tllput f' , fo r f'ac h j , the minimulIl a nd m aximum U'ij th a t any
p o int ins idp th f' no d f' could h a \'f' , This pl'o cf'cl lll'f' is m or e complf' x than in tllf' casf'
o f l o" all~ ' we ig ht pd r f'g l'eSSiOll [.5] ,
I{'T ax
u',~nln
u'T
ll1 is a lo\\'f' l' b ound
a nd
fo r each j , wh e re
o n minxi E NO /i 'i) a nd u'T <lX is a n uppf' r b o und on m a xx , E :\O (I'i,/ , T hi s is h ard
b ec ause wjl11 11 is d e t ermin ed Bo t o nly by t Ilf' m ea n and co\'aria nce o f tl lf' jt.h d a s;-;
but also t h e o th f'r cla:,ses , For f'x ample , in Fi g ur f' l. ti'3:! is approximatel~' 0, 5 , but
it wo uld bf' much la rge r if Cl werf' f mtll er to t lr f' If'ft , o r h a d a thinll e r ('o \'a ri a nce,
\\'e wi;-;h t o co mput f'
Blit l'f' nw mber th a t tl lf'
(/ 'j j ),1/
L: ~:~1 (/,h li h, \Vf'
rpquire:, th at
1'0 1'
ti'ij' S
a r f' d f' fin f'cl in t e rms of
(l ij'S,
t hus:
lI 'i)
pu t bo unds o n t.l lP (li j ' :, rela tive ly <-,a sil~' , [t sim ply
f'a ('h j \\,f' co mpllt f' l tllf' closf' ;-; t a nd fUl'tl lf'st po int fro m I',; within
((/11
I C o mpu t ing Ih bt' p oint:-. r equire,., non-t ri\'ia l co mpu t.a t ion a l geo Illetr,\ ' lwca u"e the co\'a ria lJ ce III a t rice:, are n ot llece""arily axis- a li gned , There i" n o space h ere fo r d el a iJ,."
A. W. Moore
546
Maximizer of a 2
.~
? ?
Figure 1: The rectangle denote" a h.\"lwrrectangle in the mrkd-tn'e. The !"mall
~2
"quares denote datapoint.s "owlled" h.\?
t lIe node.
Suppo:se t.here are ju:;( l \\'0
~Minimizer of a 1 claf-se!" , with the given means, and covaliances depicteel by the ellipse:;. Small
circles indicate the locations wit.hin the
rMinimizer of a 2 node for which (/) (i.e. P(.r I c))) would
-----"----------e>
NO .HYPERHECT,
be extremized.
using the Mahalanobis cli::otancp MHD(x, x') = (x-x/)T~.j I (x-x').
Call tllf':';P short.pst and furtllf'st squarpcl distancps illHDI11I11 and JIHD l11 ax . Then
(D)
is a lowpr bound for minx ,
nlln
x, E NO
END
(lij , with a similar dpfinition of aTflX . Thpn write
min (aij}Jj/L((ikPh'l
x, E NO
Wi'
)
k
>
ajlllnpj /(ClT ll1 pj
+
L ar
= x,min
(aij}Jj/(Clij}Jj + LouiN))
E NO
.
kt.l
1ax Pk)
=
W.Tll1
h?t j
wlwrp tl'T II1 is ulli' lo\\,pr bound. There is a similar definition for tl'.TflX. The iLlc'qualit.\' i;-, proved b)' elenH'ntary algebra, and requires that all qllantitips are positiw
(which thpy are). vVe can often tight.en thp bounds further using a procedure that
pxploits the fact. t.hat.
j Wij = 1, but space does not permit further discussion.
2::
tl'T
ax are close for all j. 'Vha t should be the criterion for
\ \,p will prune if wjllll1 anel
clospnpss? The first. idea that springs to mind is: Prune if Vj . (wj11aX - wj11lI1 < t).
But such a simplp critprioll is not suitable: some classps may be accumulating very
largp sums of weights, whilst others may bp accumulating vpry small Sllms. The
largp-sllll1-weight clasl>ps can t.olerate far looser bounds than the small-sum-weight
da.sses. Hprp, then, is a more satisfactory pruning critf'l'ion : Pnll1P ifVj . (wr ax Il',Tll1 < nC,;otal) where wjotal is the tot al weight. awarded to class j o\,pr tlw entire
dataset , and T is SOI1lP small constant. Sadl~' , w.ioTal is not. known ill advan('e, but
far + NO.NTTMPOINTS x
lI1 , where
happily we can find a lower bound on u,.~otal of
Lt'jofar is the total weight awarded to cla.ss j so fa.r during the sear('h over the kcl-trpp.
wr
wr
The algorithm as c1(>scribed so far performs c1ivide-and-conquer-\vith-cut.offs on the
spt of clatapoints. In addition, it is possiblp to achieve an extra ac(,pleration by
nwallS of diviclp and conquer on the class ('enters. Suppose there wpre N = 100
classps . Illstpad of considering all 100 classps at all Bodes, it is frequelltly possible
t.o clPlPrmine at SOI1W node that t.he maximum possi ble \\,pight. w,Tux for som e class j
is less thau a minisculp fraction of tllf' minimull1 pos:-;ible weight u't ln for sonlf' other
ax < Aut lll where /\
da:-,:-, "'. Thlb if we 0\'<"1' find that in some nocle
10 -..( .
tlLell class ('j is rel1lowc\ from ('onsicleration from all clescendpnt:-; ofthp Clll'l'pnt node.
FrpC[uPlltly this m ea llS that nea r tllf' tree's Ipa\'ps, only a tiny fraction of thp dassps
compete for o\\'nership of the datapoints, and thil> lea.ds to large time savings.
wr
=
Very Fast EM-Based Mixture Model Clusten'ng Using Multiresolution Kd-Trees
2
547
Results
\~'e
h a vp subj ed pd this a pp ro ach to llum prous i\ Iont.e-Ca rlo empirical tests . Her p
\VP report 0 11 on e ::::pt of Ruch tpsts . created with th e fo llowin g m eth od ology.
? We ra nd omly gP llerate a mi xt ure of Ga u::::sia ns in 1\J -dimensio ll a l : : pace (by
ciefa nlt .11 = 2 ). The numb er of G a ussians , N is , by d efa ult, :20. E ach
(~ a u ~ ~i a n h a ~ a m ean ly ing within the unit hypercub e, and a cova ria nce
m a tri x r and omly gen erat ed with dia gonal elem ents between 0 up to 40' :!
(by d efa ul t, 0' = 0.05) and rand om non-dia go nal elem ent.s t h a t ensure symm etric positive defini tene:-;s. T hus th e dist a nce from a G a ussia n cen ter t.o
it.s l -::::t.andard-elevi a tion contour is of th p order of magni t ude of 0'.
? \\lp r andomly generate a d ataset fro111 t he mixt ure m odel. The number of
point:::: , R , i~ (by defa ult) 160 ,000 . Figure :2 sh ow:::: a typical ge nerated set
of G a u::::~i a n s a nel clat apoinb.
? We then build an I/Irkd- t ree for th e d ataset. , a nd record th e m em ory requirPlllents a nd real time to build (on a Pent.ium :200Mhz, in seconds).
? We t hpn run Ei\I on the d at a. Ei\I begin:::: wit.h a n entirely different set of
(~ au ss i a n:-;, randomly ge ne ra ted using the sam e procedure.
? \Vp run 5 it era tions of the co nvent ional EM algori thm and the n ew mrkdt rpp-ba:: : pd algorithm. TllP n ew algorit.hm uses a defa ult value of 0 .1 for T .
\Vp record thp rpa l t ime (ill seconds ) for each itera tion of each a lgorithm,
a nd wp a lso reco rd t he m ean log-likelihood score (1/ R) L~= l log P(Xi I rl)
for t. he tth m od pl fo r both algo rithm:::: .
F igurf' :) :-;ho\\':-; t.h e n od es t.ha t arp visit.pd durin g It eration :2 of the Fast. EM wi t h
~y = (j cla::::ses . T a blp 1 shows t.h e d ptailecl resul ts a:::: the experim ental pa r am eters a re
varied. Speedups vary fro m 8-fold to 1000-fold . There a re 100-fold speedup:" even
wit.h very wiel e (no n-loca l) G a ussia ns. In oth pI' exp eriments, simil a r resul t s were also
obt ain f>c\ on l'ea l d ata ~ ets t ha t disobe.y t llP Gaussia n ass umption . Ther e too, we find
one- a.nd two-order- of-m agnitude computa tion al advantages with indis t. in guish able
::::tat.i :-; tical b ph a yi or (no bett.pr ancln o worse ) compared with conventi on al E i\ I.
R e al Data: Prelimin a ry experiments in a pplying t his to la rge d a tasets h ave bee n
encour ag ing. Fo r thrpe-dinlPnsi on al gala xy clust ering with 800 ,000 gala xies and
lUOO elust ns , tr adition al El\1 need ed :3?5 minutes per iteration, while t he mrkd-trees
rpquired only H SPcOl1(ls . With l. () millio n gala xies, t.radition al EM need ed 70
minut es a nd IIIrkd-trpPs required 14 seconds .
3
Conclusion
Th p use of vari a ble resolution struc t ures for clustering has been suggested in m a ny
pl aces (P.g . [7 ,8 , 4, !:l]). The BIRCH sys t em , in pa rt.icul a r , is popula r in the d a t.a b ase
co mmunity'. BIRCH is, howpver. Iln9blp to identify seconci-m Ol11Pllt feat ures of clust,pr:; (s uch as Il on-n xis-ali gned spread). Our co ntributions h ave been the use of a
ll1ulti-l'f'solut.io n a pproac h, with associa tf>d comput a tion a l benefi t s , a nd th e introducti on of a n pffi cient algo ri t hm t ha t leaves tllP sta tistica l aspects of mixture m od el
estil1l a tion uncil angpd. The growth of rpcpnt d a.t. a minin g algorihm s th a t are /l ot
based on st.a t istica l foundati ons has frec!pnt.ly been j ust.ified by the following statelllent: U:; illg st ate-of- t hp- Cl rt sta tisti cal techniques is too expensive because such
t pchniqu ps were not d psig npel to h andle la rgp da t asets and becom e intraet a bJe with
mi Ili on:'; of d a t a p oints . In earlier work we prO\ iclpd ev idence t.h a t t.hi s sta te m ent may
548
Effect of Number of Datapoints, R:
As R increases so Joes the computational aJvantage,
essentiall~' linearly. The tree-build time (11 seconds
at worst) is a tiny cost compared with even just one
iteration of Regular EM (2385 seconds, on the big
dataset.) FinalSlowSecs: 238.5. FinalFastSecs: 3.
A. W Moore
I"" ' ' ?
~""
,,,,,
','"
' ...
~
Num~
,
?.,
If'
"
,_
, '"
or pOinte (in thou ??nds)
300
Effect of Number of Dimensions, A/f:
As with many J.:d-tree algorithms. the benefits decline
as dimensionality increases, yet even in 6 dimensions,
there is an 8-fold advantage. FinalSlowSecs: 2742.
FinalFastSecs: 310.2.5.
EHect of N umber of Classes, N:
Conventional EM slows down linearly with the number of classes. Fast EM is clearly sublinear, with a
70-fold speedup even with 320 classes. Note how the
tree size grows. This is because more classes mean
a more uniform data distribution and fewer datapoints "sharing" tree leaves. FinalSlowSecs: 9278 .
FinalFastSecs: 143.:3.
Effect of Tau, T:
The larger T, the more willing we are to prune during
the tree search, anJ thus the faster we search, but the
less accurately we mirror EM's statistical behavior.
InJeeJ when T is large, the discrep<\llcy In the log
likelihood is relatively large. FinalSlowSecs: .584 . .5 .
FinalFastSecs: .)
Effect of Standard Deviation, 17:
Even with very wide Gaussians, with wide support. ,
we still get large savings . The nodes that are pruned
in these cases are rarely nodes with one class owning
all the probability, but instead are nodes where all
classes have non-zero, but little varying, probability.
FinalSlowSecs: 58.1. FinalFastSecs: 4.75.
1
2
3
4
5
6
Number of Inpuba
500
5
001 '
'0
20
003301
0025005
"u
01
40
80
03
09
02
04
160
320
Number of cent.,..
.Igma
Table 1: In all the above results all parameters were held at their default values except
for one, which varied as shown in the graphs. Each graph shows the factor by which
the new E1.'1 is faster than the conventional EM. Below each graph is the time to build
the mrkd-tree in seconds and the number of nodes in the t.ree. Note that although the
tree builJing cost is not included in the speedup calculation, it is negligibl~ in all cases,
especially considering that only one tree build is needed for all EM iterations. Does the
approximate nature of this process result in inferior clusters'? The answer is no : the
quality of clusters is indistinguishable between the slow and fast methods when measureJ
by log-likelihood and when viewed visually.
Very Fast EM-Based Mixture Model Clustering Using Multiresolution Kd-Trees
549
,[J
Figure 2: A typical set of Gaussians generated by our random procedure. They in t.urn generate the
datasets upon which we compare
the performance of the old and new
implementations of EM.
Figure 3: The ellipses show the model 8 t at the start
of an EM iteration. The rectangles depict the mrkdtree nodes that were pruned. Observe larger rectangles (and larger savings) in areas with less variation
in class probabilities. Note this is not merely able to
only prune where the data density is low.
not apply for locally weighted regression [.5] or Bayesian network learning [6], and
we hope this paper provides some evidence that it also needn't apply to clustering .
References
(1) P Cheeseman and R. Oldford. Se lectmg Models f,'om Data: Artljioal Int elligence and S tat/shes
IV. L ec tun No t es m S tattstt cs, vol. 89. Spl'inger Verlag, 1994
[:?) K Denl:) and A W Moore l\1ul t lresolutlOn Inst a nce-based Learning In Proceedl71gs of IJCAI-95.
Morgan Kaufmann , 1995.
[3) R O. Dud a and P E Hart
Pattern C la s,<ljicatlon alld Scen r AnalYSIS John Wil ey So: Sons,
[4) ]'v{ Ester , H P . Kriegel , and X . Xu A Database
In Proceeding' uf th e First
AAAI Press , 1~19 5.
1~173.
Int e rfa,~e
Int~rnatluna( Cunf~"eTlet
for Clustertllg In L a r ge Spati a l D atabas",s.
un I\'nowledge Discovery and D,lta .\Ilnlll~/?
[s] A . W M oo re , J Schneider. all d K DEcng EfficI",nt Lo cally \\'Ec lghted P olyno mi a l R Ecg ress lon Predi Ctions In D Fisher, editor, P"UCOedlTlYS ,'f the J 9Y7 ITlt~nllltlonal !\laehllLe Le(trnmg Cunlf 7'enCf.
1\10 rga n Kaufmann, 1~19 7
[6) Andrew \V M oo re and 1\t. S. Lee Cached SuffiCient Statistics for EffiCi e nt 1\1achm'" Learnmg With
Large Datasets Journal oj A,'tljicll1l Intelllyen ce Research , 8 , Marc h 1888.
[7) S M . Omo h u ndro . Efficien t Algortthms With Neural Network Be haVIOu r .
Systems, 1(2 ):2 73-34 7, 1987.
JOU7'T1l11
of Complex
[8) S. M Omohundro Burnptrees for Efficient FlmctlOn , Co nst ralllt , a nd C lassificaflon Learning. In R . P .
Lippmann, J E. Moody. and D S. Touretzky, editors, Advances tTl Neural Inform a tIOn ProCfssmg
S!jstElns '3 Morgan Kaufmann, 18~ll
[~7)
T. Zhang , R . Ramakrtshn a n, and M Llvny, BIRC H ' An Effici e nt Data Clustering Method for \'",ry
Large D a taba~,"s In Proceedwgs of th e FIfteenth AC'.\J ..;'JGACT-SIGMOD-8IGART ::"!jmpollum
0'1 Pnn np ies of Database Sys t ems: PODS 1991>. Ass n for Computing l\l achme ry, EI~II).
| 1490 |@word rising:1 nd:23 bf:1 willing:1 tat:3 covariance:1 cla:3 eld:1 tr:1 etric:1 score:1 imat:1 current:1 nt:9 od:6 si:1 yet:1 ust:1 must:1 tot:1 john:1 ints:1 depict:1 fewer:1 leaf:4 yr:1 ria:2 sys:2 ith:1 short:1 erat:1 record:2 lr:1 num:1 provides:1 node:17 location:1 ional:1 lx:2 zhang:1 ra:3 behavior:1 dist:2 ry:4 ol:1 sce:1 little:2 lll:3 considering:2 lib:1 itera:1 provided:1 xx:3 notation:1 begin:2 nea:1 taba:1 anj:1 ttl:1 fuzzy:1 developed:1 aximum:1 whilst:1 astronomy:1 ag:1 impl:1 computa:2 ti:2 ful:1 growth:1 estimat:1 ro:1 rm:1 bio:1 unit:1 ly:4 positive:1 io:3 rfa:1 era:1 ets:1 ak:1 id:1 ure:3 ree:2 andle:1 au:1 ecg:1 co:18 range:1 scribed:1 practice:1 tene:1 itlt:1 lf:6 ance:1 xr:1 procedure:4 area:1 empirical:1 eth:1 regular:1 atist:1 get:1 close:1 ga:1 cal:1 put:1 ilf:1 accumulating:2 mahala:1 conventional:2 center:1 go:1 l:2 pod:1 resolution:1 wit:6 splitting:5 rule:1 datapoints:6 his:1 lso:1 variation:1 pt:1 construction:1 suppose:1 gall:1 us:1 distinguishing:1 pa:3 expensive:1 ze:1 apo:1 cut:1 featur:1 database:2 tllp:4 ft:1 enters:1 worst:1 algorit:3 rig:1 pd:5 und:1 wil:1 icle:1 defini:1 inger:1 ov:1 tight:1 algo:2 ali:1 algebra:1 upon:1 po:3 easily:1 fast:7 kcl:1 minxi:1 hyper:1 ace:1 larger:3 whi:1 s:5 tlw:2 agnitude:1 statistic:1 cov:2 gp:1 ip:1 advantage:2 mb:2 gen:1 oint:1 multiresolution:5 achieve:1 ent:3 cli:1 parent:2 ijcai:1 p:5 ach:2 ecl:1 cluster:3 cached:1 tions:1 oo:3 andrew:2 ac:3 stat:1 ij:7 ard:1 sim:2 job:1 reco:1 c:3 nod:2 indicate:1 idp:1 owing:1 happily:1 nlt:1 alue:1 ja:1 strictly:1 pl:3 iny:1 visually:1 exp:1 nw:1 vith:2 cen:1 vary:1 numb:2 wre:1 estimation:1 applicable:1 ain:1 ound:1 tf:2 ere:2 weighted:1 hope:1 offs:1 clearly:1 gaussian:2 xact:1 pnt:2 spati:1 ck:2 aim:1 og:1 varying:1 ax:6 lon:1 likelihood:3 ave:2 centroid:2 wf:1 rpp:1 inst:2 bef:1 rgp:1 am:1 el:6 lj:1 entire:1 bt:1 her:1 ical:2 wij:2 rpa:1 umber:1 ill:3 loca:1 construct:1 saving:4 ng:1 ted:1 lit:1 np:1 others:1 report:1 lls:2 sta:3 randomly:2 ime:1 ve:2 baspd:2 ents:1 efa:2 mining:1 evaluation:1 mixture:9 sh:1 oth:1 llf:2 uill:1 tical:1 held:1 kt:1 xy:3 tree:22 impor:1 old:1 illg:1 divide:1 circle:1 iv:2 re:7 earlier:1 ar:2 shes:1 mhz:1 jib:1 cost:6 geo:1 deviation:1 subset:1 ory:1 uniform:1 uld:3 too:2 struc:1 answer:1 aw:1 st:7 ju:1 density:3 gaussia:1 lee:1 eel:1 picking:1 moody:1 aaai:1 algori:1 fir:1 ester:1 dr:1 worse:1 compu:1 depe:1 li:5 singleton:1 minin:1 needn:1 rsf:1 ology:1 int:8 vi:2 tli:1 root:3 tion:7 doing:2 defa:2 start:1 bayes:1 hf:2 odel:1 hpn:1 thp:8 ii1:1 alld:1 om:2 il:7 possi:1 kaufmann:3 identify:1 vp:4 bayesian:2 magni:1 eters:1 accurately:1 clust:3 lu:2 acc:1 datapoint:4 fo:10 inform:1 touretzky:1 ussia:2 ed:7 sharing:1 definition:1 esn:1 pp:1 mi:3 qualit:1 con:2 pst:1 dataset:3 popular:2 birch:2 wh:3 proved:1 rla:1 ut:1 dimensionality:1 cj:6 spt:1 ea:5 ta:2 y7:1 rand:1 box:2 xa:1 just:1 wlj:1 d:1 ntj:1 ei:3 su:1 maximizer:1 jll:1 quality:1 grows:1 effect:4 contain:1 lgorithm:1 mrkd:3 dud:1 moore:5 satisfactory:1 wp:1 mahalanobis:1 indistinguishable:1 ll:8 during:2 inferior:1 criterion:1 tt:1 omohundro:1 vo:2 tn:1 performs:1 pro:3 hin:2 meaning:1 fi:3 mixt:2 rl:1 tilt:3 jl:1 he:10 elem:2 nst:1 rd:1 hp:1 hus:2 access:1 clll:1 pu:1 awarded:2 verlag:1 nln:1 binary:1 yi:1 morgan:2 aut:1 care:1 schneider:1 ey:1 prune:4 clt:1 ii:4 thell:1 needing:1 mber:1 ing:5 faster:2 calculation:1 hart:1 e1:2 ase:1 functio:1 ellipsis:1 visit:2 y:1 das:2 va:2 regression:1 expectation:1 symm:1 lly:1 iteration:7 fifteenth:1 pnn:1 ion:3 c1:1 rea:1 addition:1 robotics:1 ures:2 lea:4 aws:1 ot:4 resul:2 extra:1 sr:1 tri:1 blp:1 ample:1 mod:1 pach:1 call:1 ter:1 iii:2 li1:1 ering:2 reduce:1 idea:1 frec:1 decline:1 ti1:1 rms:2 ul:2 wo:3 jj:4 dramatically:1 se:4 obt:1 locally:1 ph:1 nel:1 clusten:1 tth:1 bett:1 generate:2 nsf:1 ipa:1 wr:5 suppo:1 per:1 pace:1 carnegie:1 write:2 vol:1 threshold:1 enormous:1 kes:1 pj:1 ce:2 nal:2 ht:2 rectangle:4 graph:3 merely:1 fraction:2 nce:6 sum:2 compete:1 run:2 throughout:1 spl:1 looser:1 ble:3 vf:3 entirely:1 fl:1 hi:3 bound:5 fold:5 subj:1 idf:1 bp:1 ri:4 aspect:1 llin:1 min:2 ible:1 ults:1 urn:2 c8:1 spring:1 pruned:2 relatively:1 speedup:4 uf:1 combination:1 kd:4 associa:1 ate:2 llum:1 em:22 idence:1 ur:1 son:1 sam:1 tw:1 lp:2 qu:1 wi:3 pr:4 ln:1 turn:1 ffi:1 lta:1 needed:1 mind:1 ge:4 end:3 dia:2 gaussians:2 operation:1 permit:1 experim:1 apply:2 observe:1 cia:1 rac:1 ho:1 hat:2 rp:1 rlo:1 cent:2 top:1 res:1 cf:1 ensure:1 clustering:6 eriments:1 sw:1 cally:1 sigmod:1 especially:1 build:5 widest:2 ellipse:1 dat:2 conquer:2 bl:1 mpu:3 advan:1 ude:1 unsplit:1 fa:1 rt:2 ulli:1 aussian:1 minx:1 win:1 ow:1 predi:1 ause:1 gned:2 sia:1 elligence:1 ified:1 nc:1 equivalently:3 ncl:1 slows:1 uch:1 ba:2 implementation:1 l11:2 datasets:2 sm:1 fin:2 coincident:1 t:5 varied:2 thm:1 introduced:1 edi:1 pair:1 required:1 tive:1 ium:1 ther:1 able:2 suggested:1 pent:1 kriegel:1 llp:1 ev:1 below:4 pattern:1 solut:1 dismissed:1 nly:1 built:1 oj:1 memory:1 tau:1 tun:1 mall:1 rf:3 suitable:1 eration:1 including:1 ia:2 turning:1 ll1:2 ipoi:1 cheeseman:1 ne:1 axis:1 created:1 fro:2 hm:4 rela:1 categorical:1 ss:1 lij:2 discovery:1 hen:1 bee:1 fully:1 mixed:1 sublinear:1 lv:1 sil:1 scen:1 anel:3 sufficient:1 editor:2 tiny:2 pi:1 ata:3 lo:5 soil:1 aij:2 wide:2 benefit:1 dimension:7 default:1 vari:1 contour:1 ental:1 ig:1 ec:3 far:3 approximate:1 lippmann:1 pruning:1 feat:1 ons:1 pittsburgh:1 owns:1 xi:12 ctions:1 un:1 gala:3 nef:1 search:2 table:1 mtll:1 learn:1 nature:1 ca:8 ean:2 as:2 whe:1 complex:1 cl:8 som:1 rue:1 vj:1 marc:1 sp:1 pk:1 da:4 spread:1 linearly:2 bounding:3 big:1 child:3 xu:1 vve:1 cient:1 tl:11 en:3 tlll:1 rithm:1 owning:2 slow:1 sto:1 n:7 ny:1 sf:2 xl:1 lie:1 comput:2 ply:2 lw:1 tin:1 rk:1 minute:1 down:2 xt:1 jt:2 er:8 extremized:1 evidence:1 ih:1 mirror:1 aria:1 nat:1 te:3 tio:1 werf:1 lt:3 simply:1 ijj:1 bo:2 trf:3 monotonic:1 ch:1 minimizer:1 owned:4 tle:1 rice:1 coil:1 viewed:1 ownership:1 fisher:1 included:1 typical:2 except:1 rge:2 total:1 pas:1 arp:1 e:7 la:4 ili:1 est:1 ew:2 rarely:1 otal:2 illc:1 incl:1 support:1 mput:2 ult:3 tllf:9 reg:1 |
537 | 1,491 | Kernel peA and De-Noising in Feature Spaces
Sebastian Mika, Bernhard Scholkopf, Alex Smola
Klaus-Robert Muller, Matthias Scholz, Gunnar Riitsch
GMD FIRST, Rudower Chaussee 5, 12489 Berlin, Germany
{mika, bs, smola, klaus, scholz, raetsch} @first.gmd.de
Abstract
Kernel PCA as a nonlinear feature extractor has proven powerful as a
preprocessing step for classification algorithms. But it can also be considered as a natural generalization of linear principal component analysis. This gives rise to the question how to use nonlinear features for
data compression, reconstruction, and de-noising, applications common
in linear PCA. This is a nontrivial task, as the results provided by kernel PCA live in some high dimensional feature space and need not have
pre-images in input space. This work presents ideas for finding approximate pre-images, focusing on Gaussian kernels, and shows experimental
results using these pre-images in data reconstruction and de-noising on
toy examples as well as on real world data.
1 peA and Feature Spaces
Principal Component Analysis (PC A) (e.g. [3]) is an orthogonal basis transformation.
The new basis is found by diagonalizing the centered covariance matrix of a data set
{Xk E RNlk = 1, ... ,f}, defined by C = ((Xi - (Xk))(Xi - (Xk))T). The coordinates in the Eigenvector basis are called principal components. The size of an Eigenvalue
>. corresponding to an Eigenvector v of C equals the amount of variance in the direction
of v. Furthermore, the directions of the first n Eigenvectors corresponding to the biggest
n Eigenvalues cover as much variance as possible by n orthogonal directions. In many applications they contain the most interesting information: for instance, in data compression,
where we project onto the directions with biggest variance to retain as much information
as possible, or in de-noising, where we deliberately drop directions with small variance.
Clearly, one cannot assert that linear PCA will always detect all structure in a given data set.
By the use of suitable nonlinear features, one can extract more information. Kernel PCA
is very well suited to extract interesting nonlinear structures in the data [9]. The purpose of
this work is therefore (i) to consider nonlinear de-noising based on Kernel PCA and (ii) to
clarify the connection between feature space expansions and meaningful patterns in input
space. Kernel PCA first maps the data into some feature space F via a (usually nonlinear)
function <II and then performs linear PCA on the mapped data. As the feature space F
might be very high dimensional (e.g. when mapping into the space of all possible d-th
order monomials of input space), kernel PCA employs Mercer kernels instead of carrying
537
Kernel peA and De-Noising in Feature Spaces
out the mapping <I> explicitly. A Mercer kernel is a function k(x, y) which for all data
sets {Xi} gives rise to a positive matrix Kij = k(Xi' Xj) [6]. One can show that using
k instead of a dot product in input space corresponds to mapping the data with some <I>
to a feature space F [1], i.e. k(x,y) = (<I>(x) . <I>(y)). Kernels that have proven useful
include Gaussian kernels k(x, y) = exp( -llx - Yll2 Ie) and polynomial kernels k(x, y) =
(x?y)d. Clearly, all algorithms that can be formulated in terms of dot products, e.g. Support
Vector Machines [1], can be carried out in some feature space F without mapping the data
explicitly. All these algorithms construct their solutions as expansions in the potentially
infinite-dimensional feature space.
The paper is organized as follows: in the next section, we briefly describe the kernel PCA
algorithm. In section 3, we present an algorithm for finding approximate pre-images of
expansions in feature space. Experimental results on toy and real world data are given in
section 4, followed by a discussion of our findings (section 5).
2
Kernel peA and Reconstruction
To perform PCA in feature space, we need to find Eigenvalues A > 0 and Eigenvectors
V E F\{O} satisfying AV = GV with G = (<I>(Xk)<I>(Xk)T).1 Substituting G into the
Eigenvector equation, we note that all solutions V must lie in the span of <I>-images of the
training data. This implies that we can consider the equivalent system
A( <I>(Xk) . V)
= (<I>(Xk) . GV) for all k = 1, ... ,f
and that there exist coefficients Q1 , ...
V =
,Ql
L
(1)
such that
l
i =l
Qi<l>(Xi)
(2)
Substituting C and (2) into (1), and defining an f x f matrix K by Kij := (<I>(Xi)? <I>(Xj)) =
k( Xi, Xj), we arrive at a problem which is cast in terms of dot products: solve
fAa = Ko.
(3)
where 0. = (Q1, ... ,Ql)T (for details see [7]). Normalizing the solutions V k , i.e. (V k .
Vk) = 1, translates into Ak(o.k .o.k) = 1. To extract nonlinear principal components
for the <I>-image of a test point X we compute the projection onto the k-th component by
13k := (V k . <I> (X)) = 2:f=l Q~ k(x, Xi). Forfeature extraction, we thus have to evaluate f
kernel functions instead of a dot product in F, which is expensive if F is high-dimensional
(or, as for Gaussian kernels, infinite-dimensional). To reconstruct the <I>-image of a vector
X from its projections 13k onto the first n principal components in F (assuming that the
Eigenvectors are ordered by decreasing Eigenvalue size), we define a projection operator
P n by
(4)
k=l
If n is large enough to take into account all directions belonging to Eigenvectors with nonzero Eigenvalue, we have Pn<l>(Xi) = <I>(Xi). Otherwise (kernel) PCA still satisfies (i) that
the overall squared reconstruction error 2: i II Pn<l>(Xi) - <I>(xdll 2 is minimal and (ii) the
retained variance is maximal among all projections onto orthogonal directions in F. In
common applications, however, we are interested in a reconstruction in input space rather
than in F. The present work attempts to achieve this by computing a vector z satisfying
<I>(z) = Pn<l>(x). The hope is that for the kernel used, such a z will be a good approximation of X in input space. However. (i) such a z will not always exist and (ii) if it exists,
I For simplicity, we assume that the mapped data are centered in F. Otherwise, we have to go
through the same algebra using ~(x) := <I>(x) - (<I>(x;?).
S. Mika et al.
538
it need be not unique. 2 As an example for (i), we consider a possible representation of F.
One can show [7] that <I> can be thought of as a map <I> (x) = k( x, .) into a Hilbert space 1{k
of functions Li ai k( Xi, .) with a dot product satisfying (k( x, .) . k(y, .)) = k( x, y). Then
1{k is caHed reproducing kernel Hilbert space (e.g. [6]). Now, for a Gaussian kernel, 1{k
contains aHlinear superpositions of Gaussian bumps on RN (plus limit points), whereas
by definition of <I> only single bumps k(x,.) have pre-images under <1>. When the vector
P n<l>( x) has no pre-image z we try to approximate it by minimizing
p(z) = 1I<I>(z) - Pn<l>(x) 112
(5)
This is a special case of the reduced set method [2]. Replacing terms independent of z by
0, we obtain
p(z) = 1I<I>(z)112 - 2(<I>(z) . Pn<l>(x))
+0
(6)
Substituting (4) and (2) into (6), we arrive at an expression which is written in terms of
dot products. Consequently, we can introduce a kernel to obtain a formula for p (and thus
V' % p) which does not rely on carrying out <I> explicitly
p(z) = k(z, z) - 2
3
n
l
k=l
i=l
L f3k L a~ k(z, Xi) + 0
(7)
Pre-Images for Gaussian Kernels
To optimize (7) we employed standard gradient descent methods. If we restrict our attention
to kernels of the form k(x, y) = k(llx - Y1l2) (and thus satisfying k(x, x) == const. for all
x), an optimal z can be determined as foHows (cf. [8]): we deduce from (6) that we have
to maximize
l
p(z)
= (<I>(z) . Pn<l>(x)) + 0' = L
Ii k(z, Xi)
+ 0'
(8)
i=l
where we set Ii = L~=l f3ka: (for some 0' independent of z). For an extremum, the
gradient with respect to z has to vanish: V' %p(z) = L~=l lik'(llz - xi112)(Z - Xi) = O.
This leads to a necessary condition for the extremum: z = Li tJixd Lj tJj, with tJi
,ik'(llz - xiIl 2). For a Gaussian kernel k(x, y) = exp( -lix - Yll2 jc) we get
z=
L~=l Ii exp( -liz - xil1 2 jC)Xi
l
Li=l liexp(-llz - xil1 2jc)
~
We note that the denominator equals (<I>(z) . P n<l>(X)) (cf. (8?. Making the assumption that
Pn<l>(x) i= 0, we have (<I>(x) . Pn<l>(x)) = (Pn<l>(x) . Pn<l>(x)) > O. As k is smooth, we
conclude that there exists a neighborhood of the extremum of (8) in which the denominator
of (9) is i= O. Thus we can devise an iteration scheme for z by
Zt+l =
L~=l Ii exp( -llzt - xill 2 jC)Xi
l
Li=l li exp(-lI z t
-
xil1 2jc)
(10)
Numerical instabilities related to (<I>(z) . Pn<l>(x)) being smaH can be dealt with by restarting the iteration with a different starting value. Furthermore we note that any fixed-point
of (10) will be a linear combination of the kernel PCA training data Xi. If we regard (10)
in the context of clustering we see that it resembles an iteration step for the estimation of
2If the kernel allows reconstruction of the dot-product in input space, and under the assumption
that a pre-image exists, it is possible to construct it explicitly (cf. [7]). But clearly, these conditions
do not hold true in general.
Kernel peA and De-Noising in Feature Spaces
539
the center of a single Gaussian cluster. The weights or 'probabilities' T'i reflect the (anti-)
correlation between the amount of cP (x) in Eigenvector direction Vk and the contribution
of CP(Xi) to this Eigenvector. So the 'cluster center' z is drawn towards training patterns
with positive T'i and pushed away from those with negative T'i, i.e. for a fixed-point Zoo the
influence of training patterns with smaner distance to x wi11 tend to be bigger.
4 Experiments
To test the feasibility of the proposed algorithm, we run several toy and real world experiments. They were performed using (10) and Gaussian kernels of the form k(x, y) =
exp( -(llx - YI12)/(nc)) where n equals the dimension of input space. We mainly focused
on the application of de-noising, which differs from reconstruction by the fact that we are
allowed to make use of the original test data as starting points in the iteration.
Toy examples: In the first experiment (table 1), we generated a data set from eleven Gaussians in RIO with zero mean and variance u 2 in each component, by selecting from each
source 100 points as a training set and 33 points for a test set (centers of the Gaussians randomly chosen in [-1, 1]10). Then we applied kernel peA to the training set and computed
the projections 13k of the points in the test set. With these, we carried out de-noising, yielding an approximate pre-image in RIO for each test point. This procedure was repeated for
different numbers of components in reconstruction, and for different values of u. For the
kernel, we used c = 2u 2 ? We compared the results provided by our algorithm to those of
linear peA via the mean squared distance of an de-noised test points to their corresponding center. Table 1 shows the ratio of these values; here and below, ratios larger than one
indicate that kernel peA performed better than linear peA. For almost every choice of
nand u, kernel PeA did better. Note that using alllO components, linear peA is just a
basis transformation and hence cannot de-noise. The extreme superiority of kernel peA
for small u is due to the fact that all test points are in this case located close to the eleven
spots in input space, and linear PeA has to cover them with less than ten directions. Kernel PeA moves each point to the correct source even when using only a sman number of
components.
n=1
2
3
4
7
8
5
6
9
0.05 2058.42 1238.36 846.14 565.41 309.64 170.36 125.97 104.40 92.23
0.1
10.22
31.32 21.51
29.24 27.66 2:i.5:i 29.64 40.07 63.41
1.12
1.18
0.2
0.99
1.50
2.11
2.73
3.72
5.09 6.32
1.07
1.44
0.4
1.26
1.64
1.91
2.08
2.34 2.47
2.22
0.8
1.39
1.54
2.25 2.39
1.23
1.7U
1.8U
1.96
2.10
Table 1: De-noising Gaussians in RIO (see text). Performance ratios larger than one indicate how much better kernel PeA did, compared to linear PeA, for different choices of the
Gaussians' std. dev. u, and different numbers of components used in reconstruction.
To get some intuitive understanding in a low-dimensional case, figure 1 depicts the results
of de-noising a half circle and a square in the plane, using kernel peA, a nonlinear autoencoder, principal curves, and linear PeA. The principal curves algorithm [4] iteratively
estimates a curve capturing the structure of the data. The data are projected to the closest
point on a curve which the algorithm tries to construct such that each point is the average
of all data points projecting onto it. It can be shown that the only straight lines satisfying
the latter are principal components, so principal curves are a generalization of the latter.
The algorithm uses a smoothing parameter whic:h is annealed during the iteration. In the
nonlinear autoencoder algorithm, a 'bottleneck' 5-layer network is trained to reproduce the
input values as outputs (i.e. it is used in autoassociative mode). The hidden unit activations
in the third layer form a lower-dimensional representation of the data, closely related to
540
S. Mika et al.
PCA (see for instance [3]). Training is done by conjugate gradient descent. In all algorithms, parameter values were selected such that the best possible de-noising result was
obtained. The figure shows that on the closed square problem, kernel PeA does (subjectively) best, followed by principal curves and the nonlinear autoencoder; linear PeA fails
completely. However, note that all algorithms except for kernel PCA actually provide an
explicit one-dimensional parameterization of the data, whereas kernel PCA only provides
us with a means of mapping points to their de-noised versions (in this case, we used four
kernel PCA features, and hence obtain a four-dimensional parameterization).
kernel PCA
nonlinear autoencoder
Principal Curves
linear PCA
~It%i 1J;:qf~~
I::f;~.;~,: '~ll
Figure 1: De-noising in 2-d (see text). Depicted are the data set (small points) and its
de-noised version (big points, joining up to solid lines). For linear PCA, we used one
component for reconstruction, as using two components, reconstruction is perfect and thus
does not de-noise. Note that all algorithms except for our approach have problems in
capturing the circular structure in the bottom example.
USPS example: To test our approach on real-world data, we also applied the algorithm
to the USPS database of 256-dimensional handwritten digits. For each of the ten digits,
we randomly chose 300 examples from the training set and 50 examples from the test
set. We used (10) and Gaussian kernels with c = 0.50, equaling twice the average of
the data's variance in each dimensions. In figure 4, we give two possible depictions of
1111 II iI 111111 fill
0"(1' . ___ ? ?
11m 111111111111
Figure 2: Visualization of Eigenvectors (see
text). Depicted are the 2?, ... , 28 -th Eigenvector (from left to right). First row: linear
PeA, second and third row: different visualizations for kernel PCA.
the Eigenvectors found by kernel PCA, compared to those found by linear PCA for the
USPS set. The second row shows the approximate pre-images of the Eigenvectors V k ,
k = 2?, ... ,2 8 , found by our algorithm. In the third row each image is computed as
follows: Pixel i is the projection of the <II-image of the i-th canonical basis vector in input
space onto the corresponding Eigenvector in features space (upper left <II(el) . V k , lower
right <II (e256) . Vk). In the linear case, both methods would simply yield the Eigenvectors
oflinear PCA depicted in the first row; in this sense, they may be considered as generalized
Eigenvectors in input space. We see that the first Eigenvectors are almost identical (except
for signs). But we also see, that Eigenvectors in linear PeA start to concentrate on highfrequency structures already at smaller Eigenvalue size. To understand this, note that in
linear PCA we only have a maximum number of 256 Eigenvectors, contrary to kernel PCA
which gives us the number of training examples (here 3000) possible Eigenvectors. This
541
Kernel peA and De-Noising in Feature Spaces
? ? ? & ? ? ? ? ? ? ? ? 3.55
? ? ? &3
o.n
1.02
1.02
1.01
0.113
I.CI'
0.111
0.118
0.118
1.01
0.60
0.78
0.76
0.52
0.73
0.7(
0.80
0.7(
0.7(
0.72
? ? ? ? ? ? ? ? ? ? ? ? ? 3S ? ? ? ? S!
Figure 3: Reconstruction of USPS data. Depicted are the reconstructions of the first digit
in the test set (original in last column) from the first n = 1, ... ,20 components for linear
peA (first row) and kernel peA (second row) case. The numbers in between denote the
fraction of squared distance measured towards the original example. For a small number
of components both algorithms do nearly the same. For more components, we see that
linear peA yields a result resembling the original digit, whereas kernel peA gives a result
resembling a more prototypical 'three'
also explains some of the results we found when working with the USPS set. In these
experiments, linear and kernel peA were trained with the original data. Then we added (i)
additive Gaussian noise with zero mean and standard deviation u = 0.5 or (ii) 'speckle'
noise with probability p = 0.4 (i.e. each pixel flips to black or white with probability
p/2) to the test set. For the noisy test sets we computed the projections onto the first n
linear and nonlinear components, and carried out reconstruction for each case. The results
were compared by taking the mean squared distance of each reconstructed digit from the
noisy test set to its original counterpart. As a third experiment we did the same for the
original test set (hence doing reconstruction, not de-noising). In the latter case, where
the task is to reconstruct a given example as exactly as possible, linear peA did better,
at least when using more than about 10 components (figure 3). This is due to the fact
that linear peA starts earlier to account for fine structures, but at the same time it starts
to reconstruct noise, as we will see in figure 4. Kernel PCA, on the other hand, yields
recognizable results even for a small number of components, representing a prototype of
the desired example. This is one reason why our approach did better than linear peA for the
de-noising example (figure 4). Taking the mean squared distance measured over the whole
test set for the optimal number of components in linear and kernel PCA, our approach did
better by a factor of 1.6 for the Gaussian noise, and 1.2 times better for the 'speckle' noise
(the optimal number of components were 32 in linear peA, and 512 and 256 in kernel
PCA, respectively). Taking identical numbers of components in both algorithms, kernel
peA becomes up to 8 (!) times better than linear peA. However, note that kernel PCA
comes with a higher computational complexity.
5 Discussion
We have studied the problem of finding approximate pre-images of vectors in feature space,
and proposed an algorithm to solve it. The algorithm can be applied to both reconstruction
and de-noising. In the former case, results were comparable to linear peA, while in the
latter case, we obtained significantly better results. Our interpretation of this finding is as
follows. Linear peA can extract at most N components, where N is the dimensionality of
the data. Being a basis transform, all N components together fully describe the data. If the
data are noisy, this implies that a certain fraction of the components will be devoted to the
extraction of noise. Kernel peA, on the other hand, allows the extraction of up to f features,
where f is the number of training examples. Accordingly, kernel peA can provide a larger
number of features carrying information about the structure in the data (in our experiments,
we had f > N). In addition, if the structure to be extracted is nonlinear, then linear peA
must necessarily fail , as we have illustrated with toy examples.
These methods, along with depictions of pre-images of vectors in feature space, provide
some understanding of kernel methods which have recently attracted increasing attention.
Open questions include (i) what kind of results kernels other than Gaussians will provide,
542
S. Mika et al.
Figure 4: De-Noising of USPS data (see text). The left half shows: top: the first occurrence
of each digit in the test set, second row: the upper digit with additive Gaussian noise (0' =
0.5), following five rows: the reconstruction for linear PCA using n = 1,4,16,64,256
components, and, last five rows: the results of our approach using the same number of
components. In the right half we show the same but for 'speckle' noise with probability
p = 0.4.
(ii) whether there is a more efficient way to solve either (6) or (8), and (iii) the comparison
(and connection) to alternative nonlinear de-noising methods (cf. [5]).
References
[1] B. Boser, I. Guyon, and V.N. Vapnik. A training algorithm for optimal margin classifiers. In D. Haussler, editor, Proc. COLT, pages 144-152, Pittsburgh, 1992. ACM
Press.
[2] C.J.c. Burges. Simplified support vector decision rules. In L. Saitta, editor, Prooceedings, 13th ICML, pages 71-77, San Mateo, CA, 1996.
[3] K.I. Diamantaras and S.Y. Kung. Principal Component Neural Networks. Wiley, New
York,1996.
[4] T. Hastie and W. Stuetzle. Principal curves. JASA, 84:502-516,1989.
[5] S. Mallat and Z. Zhang. Matching Pursuits with time-frequency dictionaries. IEEE
Transactions on Signal Processing, 41(12):3397-3415, December 1993.
[6] S. Saitoh. Theory of Reproducing Kernels and its Applications. Longman Scientific &
Technical, Harlow, England, 1988.
[7] B. Scholkopf. Support vector learning. Oldenbourg Verlag, Munich, 1997.
[8] B. Scholkopf, P. Knirsch, A. Smola, and C. Burges. Fast approximation of support vector kernel expansions, and an interpretation of clustering as approximation in feature
spaces. In P. Levi et. a1., editor, DAGM'98, pages 124 - 132, Berlin, 1998. Springer.
[9] B. Scholkopf, A.J. Smola, and K.-R. Muller. Nonlinear component analysis as a kernel
eigenvalue problem. Neural Computation, 10:1299-1319,1998.
| 1491 |@word version:2 briefly:1 polynomial:1 compression:2 open:1 covariance:1 q1:2 solid:1 contains:1 selecting:1 riitsch:1 activation:1 must:2 written:1 attracted:1 oldenbourg:1 additive:2 numerical:1 eleven:2 gv:2 drop:1 half:3 selected:1 parameterization:2 accordingly:1 plane:1 xk:7 provides:1 zhang:1 five:2 along:1 ik:1 scholkopf:4 recognizable:1 introduce:1 decreasing:1 increasing:1 becomes:1 provided:2 project:1 what:1 kind:1 eigenvector:7 finding:5 transformation:2 extremum:3 assert:1 every:1 exactly:1 classifier:1 unit:1 diamantaras:1 superiority:1 positive:2 limit:1 ak:1 joining:1 mika:5 might:1 plus:1 chose:1 resembles:1 twice:1 black:1 studied:1 mateo:1 scholz:2 unique:1 differs:1 digit:7 procedure:1 stuetzle:1 spot:1 thought:1 significantly:1 projection:7 matching:1 pre:12 get:2 onto:7 cannot:2 close:1 operator:1 noising:19 context:1 live:1 instability:1 influence:1 optimize:1 equivalent:1 map:2 center:4 annealed:1 go:1 attention:2 starting:2 resembling:2 focused:1 simplicity:1 rule:1 haussler:1 fill:1 coordinate:1 mallat:1 us:1 satisfying:5 expensive:1 located:1 std:1 database:1 bottom:1 lix:1 equaling:1 noised:3 complexity:1 rudower:1 carrying:3 trained:2 algebra:1 basis:6 completely:1 usps:6 fast:1 describe:2 klaus:2 neighborhood:1 larger:3 solve:3 reconstruct:3 otherwise:2 transform:1 noisy:3 eigenvalue:7 matthias:1 reconstruction:17 product:7 maximal:1 achieve:1 intuitive:1 cluster:2 perfect:1 measured:2 implies:2 indicate:2 come:1 direction:9 concentrate:1 closely:1 correct:1 tji:1 pea:39 centered:2 explains:1 generalization:2 clarify:1 hold:1 considered:2 liz:1 exp:6 mapping:5 bump:2 substituting:3 dictionary:1 purpose:1 estimation:1 proc:1 superposition:1 hope:1 clearly:3 gaussian:13 always:2 rather:1 pn:11 vk:3 mainly:1 detect:1 sense:1 rio:3 saitta:1 el:1 dagm:1 lj:1 nand:1 hidden:1 reproduce:1 interested:1 germany:1 pixel:2 overall:1 classification:1 among:1 colt:1 smoothing:1 special:1 equal:3 construct:3 extraction:3 identical:2 icml:1 nearly:1 employ:1 randomly:2 attempt:1 circular:1 extreme:1 yielding:1 pc:1 devoted:1 necessary:1 orthogonal:3 circle:1 desired:1 minimal:1 instance:2 kij:2 column:1 earlier:1 dev:1 cover:2 deviation:1 monomials:1 ie:1 retain:1 together:1 wi11:1 squared:5 reflect:1 knirsch:1 li:6 toy:5 account:2 de:25 coefficient:1 jc:5 explicitly:4 performed:2 try:2 closed:1 doing:1 start:3 contribution:1 square:2 variance:7 yield:3 dealt:1 handwritten:1 zoo:1 straight:1 sebastian:1 definition:1 frequency:1 f3k:1 dimensionality:1 organized:1 hilbert:2 actually:1 focusing:1 higher:1 done:1 furthermore:2 just:1 smola:4 correlation:1 working:1 hand:2 replacing:1 nonlinear:15 mode:1 scientific:1 contain:1 true:1 deliberately:1 former:1 hence:3 counterpart:1 nonzero:1 iteratively:1 illustrated:1 white:1 ll:1 during:1 faa:1 generalized:1 performs:1 cp:2 image:17 recently:1 common:2 interpretation:2 raetsch:1 llx:3 ai:1 had:1 dot:7 depiction:2 deduce:1 subjectively:1 saitoh:1 closest:1 certain:1 verlag:1 muller:2 devise:1 employed:1 maximize:1 signal:1 ii:16 lik:1 smooth:1 technical:1 england:1 bigger:1 a1:1 feasibility:1 qi:1 ko:1 denominator:2 iteration:5 kernel:63 whereas:3 addition:1 chaussee:1 fine:1 source:2 tend:1 december:1 contrary:1 iii:1 enough:1 xj:3 hastie:1 restrict:1 idea:1 prototype:1 translates:1 bottleneck:1 whether:1 expression:1 pca:31 york:1 autoassociative:1 useful:1 harlow:1 eigenvectors:13 amount:2 ten:2 gmd:2 reduced:1 exist:2 canonical:1 sign:1 gunnar:1 four:2 levi:1 drawn:1 longman:1 fraction:2 yll2:2 run:1 powerful:1 arrive:2 almost:2 guyon:1 decision:1 comparable:1 pushed:1 capturing:2 layer:2 followed:2 nontrivial:1 alex:1 whic:1 span:1 munich:1 combination:1 belonging:1 conjugate:1 smaller:1 b:1 making:1 projecting:1 equation:1 visualization:2 fail:1 flip:1 pursuit:1 gaussians:5 away:1 occurrence:1 alternative:1 original:7 top:1 clustering:2 include:2 cf:4 const:1 move:1 question:2 already:1 added:1 highfrequency:1 gradient:3 distance:5 mapped:2 berlin:2 reason:1 assuming:1 retained:1 ratio:3 minimizing:1 nc:1 ql:2 robert:1 potentially:1 negative:1 rise:2 zt:1 perform:1 upper:2 av:1 descent:2 anti:1 speckle:3 defining:1 rn:1 reproducing:2 cast:1 connection:2 boser:1 usually:1 pattern:3 xill:1 below:1 yi12:1 suitable:1 natural:1 rely:1 diagonalizing:1 representing:1 scheme:1 carried:3 extract:4 autoencoder:4 text:4 understanding:2 fully:1 interesting:2 tjj:1 prototypical:1 proven:2 jasa:1 mercer:2 llz:3 editor:3 row:10 qf:1 last:2 understand:1 burges:2 taking:3 regard:1 curve:8 dimension:2 world:4 xil1:3 preprocessing:1 projected:1 simplified:1 san:1 transaction:1 reconstructed:1 approximate:6 restarting:1 bernhard:1 pittsburgh:1 conclude:1 xi:19 why:1 table:3 ca:1 expansion:4 necessarily:1 did:6 big:1 noise:10 whole:1 allowed:1 repeated:1 biggest:2 depicts:1 oflinear:1 wiley:1 fails:1 explicit:1 lie:1 vanish:1 third:4 extractor:1 formula:1 normalizing:1 exists:3 vapnik:1 ci:1 margin:1 suited:1 depicted:4 simply:1 ordered:1 springer:1 corresponds:1 satisfies:1 extracted:1 acm:1 formulated:1 consequently:1 towards:2 infinite:2 determined:1 except:3 principal:13 called:1 experimental:2 meaningful:1 support:4 latter:4 kung:1 evaluate:1 |
538 | 1,492 | Viewing Classifier Systems
as Model Free Learning in POMDPs
Akira Hayashi and Nobuo Suematsu
Faculty of Information Sciences
Hiroshima City University
3-4-1 Ozuka-higashi, Asaminami-ku, Hiroshima, 731-3194 Japan
{akira,suematsu }@im.hiroshima-cu.ac.jp
Abstract
Classifier systems are now viewed disappointing because of their problems such as the rule strength vs rule set performance problem and the
credit assignment problem. In order to solve the problems, we have developed a hybrid classifier system: GLS (Generalization Learning System). In designing GLS, we view CSs as model free learning in POMDPs
and take a hybrid approach to finding the best generalization, given the
total number of rules. GLS uses the policy improvement procedure by
Jaakkola et al. for an locally optimal stochastic policy when a set of
rule conditions is given. GLS uses GA to search for the best set of rule
conditions.
1
INTRODUCTION
Classifier systems (CSs) (Holland 1986) have been among the most used in reinforcement
learning. Some of the advantages of CSs are (1) they have a built-in feature (the use of
don't care symbols "#") for input generalization, and (2) the complexity of pOlicies can
be controlled by restricting the number of rules. In spite of these attractive features, CSs
are now viewed somewhat disappointing because of their problems (Wilson and Goldberg
1989; Westerdale 1997). Among them are the rule strength vs rule set performance problem, the definition of the rule strength parameter, and the credit assignment (BBA vs PSP)
problem.
In order to solve the problems, we have developed a hybrid classifier system: GLS (Generalization Learning System). GLS is based on the recent progress ofRL research in partially
observable Markov decision processes (POMDPs). In POMDPs, the environments are really Markovian, but the agent cannot identify the state from the current observation. It may
be due to noisy sensing or perceptual aliasing. Perceptual aliasing occurs when the sensor
returns the same observation in multiple states. Note that even for a completely observable
A. Hayashi and N. Suematsu
990
MDP, the use of don't care symbols for input generalization will make the process as if it
were partially observable.
In designing GLS, we view CSs as RL in POMDPs and take a hybrid approach to finding
the best generalization, given the total number of rules. GLS uses the policy improvement
procedure in Jaakkola et a!. (1994) for an locally optimal stochastic policy when a set of
rule conditions is given. GLS uses GA to search for the best set of rule conditions.
The paper is organized as follows. Since CS problems are easier to understand from GLS
perspective, we introduce Jaakkola et a!. (1994), propose GLS, and then discuss CS problems.
2
LEARNING IN POMDPS
Jaakkola et a1. (1994) consider POMDPs with perceptual aliasing and memoryless stochastic policies. Following the authors, let us call the observations messages. Therefore, a
policy is a mapping from messages to probability distributions (PDs) over the actions.
Given a policy 7r, the value of a state s, V7!' (s), is defined for POMDPs just as for MDPs.
Then, the value of a message m under policy 7r, V7!' (m ), can be defined as follows:
V7!'(m) = LP7!'(slm)V7!'(s)
(1)
sES
where P7!' (slm) is the probability that the state is s when the message is m under the policy
7r.
Then, the following holds.
N
lim'"' E{R(st, at) -R
N-+(X)~
t=l
E{V(s) Is
--t
lSI
= s}
m}
(2)
(3)
where St and at refer to the state and the action taken at the tth step respectively, R( St, at)
is the immediate reward at the tth step, R is the (unknown) gain (Le. the average reward
per step). s --t m refers to all the instances where m is observed in sand E{? I s --t m} is
a Monte-Carlo expectation.
In order to compute E{V(s)
Is
--t
m}, Jaakkola et a1. showed a Monte-Carlo procedure:
1
vt(m) = 'k{
Rtl
+rl,IRtl+l + rl,2Rtl+2 + ... + rl ,t-tIRt
+
Rt2
+r2,IRt2 +l + r2,2Rt2+2 + ... + r2,t-t2Rt
(4)
+ Rtk +rk ,IRtdl + ... + rk ,t-tkRtl
where tk denotes. the time step corresponding to the kth occurrence of the message m,
R t = R(st, at) - R for every t, rk,T indicates the discounting at the Tth step in the kth
sequence. By estimating R and by suitably setting rk ,T, Vt(m) converges to V7!'(m).
Q7!' (m, a), Q-value of the message m for the action a under the policy 7r, is also defined
and computed in the same way.
Jaakkola et a1. have developed a policy improvement method:
Step 1 Evaluate the current policy 7r by computing V7!' (m) and Q7!' (m, a) for each m and
a.
Viewing Classifier Systems as Model Free Learning in POMDPs
991
Step 2 Test for any m whether max a Q1r (m, a) > V 1r (m) holds. If 110t, then return 7r.
Step 3 For each m and a, define 7r 1 (alm) as follows:
7r 1 (aim) = 1.0 when a = argmaxaQ1r(m, a),
7r 1 (aim) = 0.0 otherwise.
f
f
Then, define 7r as 7r (aim) = (1 - ? )7r( aim) + ?7r 1 (aim)
Step 4 Set the new policy as 7r = 7r f , and goto Stepl.
3
GLS
Each rule in GLS consists of a condition part, an action part, and an evaluation
part: Rule = (Condit'i on, Action, Evaluation). The condition part is a string c
over the alphabet {O, 1, #}, and is compared with a binary sensor message. # is a
don't care symbol, and matches 0 and 1. When the condition c matches the message, the action is randomly selected using the PD in the action part: Action =
(p(allc),p(a21c), ... ,p(a IA!lc)), I:j'!\ p(ajlc) = 1.0 where IAI is the total number ofactions. The evaluation part records the value of the condition V (c) and the Q-values of the
condition action pairs Q(c, a): Evaluation = (V(c), Q(c, ad , Q(c, a2), ... ,Q(c, a lAI))'
Each rule set consists of N rules, {Rulel, Rule2,"" RuleN}. N, the total number of
rules in a rule set, is a design parameter to control the complexity of policies. All the rules
except the last one are called standard rules. The last rule Rule N is a special rule which is
called the default rule. The condition part of the default rule is a string of # 's and matches
any message.
Learning in GLS proceeds as follows: (1 )Initialization: randomly generate an initial population of M rule sets, (2)Policy Evaluation and Improvement: for each rule set, repeat a
policy evaluation and improvement cycle for a suboptimal policy, then, record the gain of
the policy for each rule set, (3)Genetic Algorithm: use the gain of each rule set as its fitness measure and produce a new generation of rule sets, (4) Repeat: repeat from the policy
evaluation and improvement step with the new generation of rule sets.
In (2)Policy Evaluation and Improvement, GLS repeats the following cycle for each rule
set.
Step 1 Set ? sufficiently small. Set t max sufficiently large.
Step 2 Repeat for 1 :::; t :::; t max ?
1. Make an observation of the environment and receive a message mt from the
sensor.
2. From all the rules whose condition matches the message mt, find the rule
whose condition is the most specific l . Let us call the rule the active rule.
3. Select the next action at randomly according to the PD in the action part of
the active rule, execute the action, and receive the reward R( St, at) from the
environment. (The state St is not observable.)
4. Update the current estimate of the gain R from its previous estimate and
R( St, ad . Let R t = R( St , ad - R. For each rule, consider its condition Ci
as (a generalization of) a message, and update its evaluation part V (Ci ) and
Q(c;, aHa E A) using Eq.(4).
Step 3 Check whether the following holds. If not, exit.
3i (1 :::; i :::; N), max a Q (Ci , a) > V (cd
Step 4 Improve the current policy according to the method in the previous section, and
update the action part of the corresponding rules and goto Step 2.
IThe most specific rule has the least number of #'s. This is intended only for saving the number
of rules.
A. Hayashi and N. Suematsu
992
GLS extracts the condition parts of all the rules in a rule set and concatenates them to
form a string. The string will be an individual to be manipulated by the genetic algorithm
(GA). The genetic algorithm used in GLS is a fairly standard one. GLS combines the SGA
(the simple genetic algorithm) (Goldberg 1989) with the elitist keeping strategy. The SGA
is composed of three genetic operators: selection, crossover, and mutation. The fitness
proportional selection and the single-point crossover are used. The three operators are
applied to an entire population at each generation. Since the original SGA does not consider
#'s in the rule conditions, we modified SGA as follows. When GLS randomly generates
an initial population of rule sets, it generates # at each allele position in rule conditions
according to the probability P#.
4
CS PROBLEMS AND GLS
In the history of classifier systems, there were two quite different approaches: the Michigan
approach (Holland and Reitman 1978), and the Pittsburgh (Pitt) approach (Dejong 1988).
In the Michigan approach, each rule is considered as an individual and the rule set as the
population in GA. Each rule has its strength parameter, which is based on its future payoff
and is used as the fitness measure in GA. These aspects of the approach cause many problems. One is the rule strength vs rule set performance problem. Can we collect only strong
rules and get the best rule set performance? Not necessarily. A strong rule may cooperate
with weak rules to increase its payoff. Then, how can we define and compute the strength
parameter for the best rule set performance? In spite of its problems, this approach is now
so much more popular than the other, that when people simply say classifier systems, they
refer to Michigan type classifier systems. In the Pitt approach, the problems of the Michigan approach are avoided by requiring GA to evaluate a whole rule set. In the approach, a
rule set is considered as an individual and multiple rule sets are kept as the population. The
problem of the Pitt approach is its computational difficulties.
GLS can be considered as a combination of the Michigan and Pitt approaches. GA in GLS
works as that in the Pitt approach. It evaluates a total rule set, and completely avoids the
rule strength vs rule set performance problem in the Michigan approach. As the Michigan
type CSs, GLS evaluates each rule to improve the policy. This alleviates the computational
burden in the Pitt approach. Moreover, GLS evaluates each rule in a more formal and sound
way than the Michigan approach. The values, V(c), and Q(c, a), are defined on the basis
of POMDPs, and the policy improvement procedure using the values is guaranteed to find
a local maximum.
Westerdale (1997) has recently made an excellent analysis of problematic behaviors of
Michigan type CSs. Two popular methods for credit assignment in CSs are the bucket
brigade algorithm (BBA) (Holland 1986) and the profit sharing plan (PSP) (Grefenstette
1988). Westerdale shows that BBA does not work in POMDPs. He insists that PSP with
infinite time span is necessary for the right credit assignment, although he does not show
how to carry out the computation. GLS does not use BBA or PSP. GLS uses the Monte
Carlo procedure, Eq.(4), to compute the value of each condition action pair. The series
in Eq.(4) is slow to converge. But, this is the cost we have to pay for the right credit
assignment in POMDPs. Westerdale points out another CS problem. He claims that a
distinction must be made between the availability and the payoff of rules. We agree with
him. As he says, if the expected payoff of Rule 1 is twice as much as Rule 2, then we
want to a/ways choose Rule 1. GLS makes the distinction. The probability of a stochastic
policy 71'(alc) in GLS corresponds to the availability, and the value of a condition action
pair Q ( c, a) corresponds to the payoff.
Samuel System (Grefenstette et a1. 1990) can also be considered as a combination of the
Michigan and Pitt approaches. Samuel is a highly sophisticated system which has lots of
features. We conjecture, however, that Samuel is not free from the CS problems which
Viewing Classifier Systems as Model Free Learning in POMDPs
993
Westerdale has analyzed. This is because Samuel uses PSP for credit assignment, and
Samuel uses the payoff of each rule for action selection, and does not make a distinction
between the availability and the payoff of rules.
xes (Wilson 1995) seems to be an exceptionally reliable Michigan?type es. In xes, each
rule's fitness is based not on its future payoff but on the prediction accuracy of its future
payoff (XeS uses BBA for credit assignment). Wilson reports that xes's population tends
to form a complete and accurate mapping from sensor messages and actions to payoff
predictions. We conjecture that xes tries to build the most general Markovian model of
the environment. Therefore, it will be difficult to apply xes when the environment is not
Markovian, or when we cannot afford the number of rules enough to build a Markovian
model of the environment, even if the environment itself is Markovian. As we will see in
the next section, GLS is intended exactly for these situations.
Kaelbling et a1. (19%) surveys methods for input generalization when reward is delayed.
The methods use a function approximator to represent the value function by mapping a state
description to a value . Since they use value iteration or Q?leaming anyway, it is difficult to
apply the methods when the generalization violates the Markov assumption and induces a
POMDP.
5
EXPERIMENTS
We have tested GLS with some of the representative problems in es literature. Fig. 1 shows
Grefl world (Grefenstette 1987). In Grefl world, we used GLS to find the smallest rule set
which is necessary for the optimal performance. Since this is not a POMDP but an MDP, the
optimal policy can easily be learned when we have a corresponding rule for each of the 16
states. However, when the total number of rules is less than that of states, the environment
looks like a POMDP to the learning agent, even if the environment itself is an MDP. The
graph shows how the gain of the best rule set in the population changes with the generation.
We can see from the figure that four rules are enough for the optimal performance. Also
note that the saving of the rules is achieved by selecting the most specific matching rule
as an active rule. The rule set with this rule selection is called the defallit hierarchy in es
literature.
payoff
150
200
'i
~~~--------~
ISO ....................................... .. ..........~ ................. .
100
N: . l N=3 N=2
50
M ? ?? ? ? _ _
O L-~~__~~~~
~L~~
?I ~
.. -.~
. ~.-~
-
o
10
15
10
15
30
35
40
g.!ner.a tioruJ
Figure 1: LEFT: GREF1 World. States {O, 1,2, 3} are the start states and states {12.13, 14, 15 }
are the end states. In each state, the agent gets the state number (4 bits) as a message, and chooses
an action a,b,c, or d. When the agent reaches the end states, he receives reward 1000 in state 13, but
reward 0 in other states. Then the agent is put in one of the start states with equal probability. We
added 10% action errors to make the process ergodic. When an action error occurs, the agent moves
to one of the 16 states with equal probability.
RIGHT: Gain of the best rule set. Parameters: tma ;r =: 10000. ? =: 0.10. M =: 10. N =:
2,3 , 4, P# =: 0.33. For N =: 4, the best rule set at the 40 th generation was { if 0101 (State 5)
then a 1.0, if 1010 (State 10) then c 1.0, if ##11 (States 3,7,11,15) then d 1.0, if #### (Default
Rule) then b 1.0}.
A. Hayashi and N. Suematsu
994
oo~~~~~~~~~--~
80
70
60
BlillaD
a
II
II II
a
a
30
N06NoS-
20
10 L -......~~~oo4---'---'!'Pz.:tim:=o?.:...1-_""....
-.-J.
o W 20 30 ~ ~ 60 m 80 00 ~
aenerations
Figure 2: LEFf: McCallum's Maze. We show the state numbers in the left, and the messages in the
right. States 8 and 9 are the start states, and state G is the goal state. In each state, the agent receives
a sensor message which is 4 bit long, Each bit in the message tells whether a wall exists in each of
the four directions. From each state, the agent moves to one of the adjacent states. When the agent
reaches the goal state, he receives reward 1000. The agent is then put in one of the start states with
equal probability.
RIGHT: Gain of the best rule set. Parameters: t mBX = 50000, ~ = 0.10, M = 10, N = 5,6, P# =
0.33.
Fig. 2 is a POMDP known as as McCallum's Maze (McCallum 1993). Thanks to the use
of stochastic policies, GLS achieves near optimal gain for memoryless poliCies. Note that
no memoryless deterministic policy can take the agent to the goal for this problem.
We have seen GLS's generalization capability for an MDP in Grefl World, the advantage
of stochastic policies for a POMDP in McCallum's maze. In Woods7 (Wilson 1994), we
attempt to test GLS's generalization capability for a POMDP. See Fig. 3. Since each sensor
message is 16 bit long, and the conditions of GLS rules can have either O,l,or # for each of
the 16 bits, there are 3 16 possible conditions in total. When we notice that there are only
92 different actual sensor messages in the environment, it seems quite difficult to discover
them only by using GA. In fact, when we ran GLS for the first time, the standard rules
very rarely matched the messages and the default rule took over most of the time. In order
to avoid the no matching rule problem, we made the number of rules in a rule set large
(N = 100), increased P# from 0.33 in the previous problems to 0.70.
The problem was independently attacked by other methods. Wilson applied his ZCS, zeroth
level classifier system, to Woods7 (Wilson 1994). The gain was 0.20. ZCS has a special
covering procedure to tum around the no matching rule problem. The covering procedure
generates a rule which matches a message when none of the current rules matches the
message. We expect further improvement on the gain, if we equip GLS with some covering
procedure.
6
SUMMARY
In order to solve the CS problems such as the rule strength vs rule set performance problem
and the credit assignment problem, we have developed a hybrid classifier system: GLS.
We notice that generalization often leads to state aliasing. Therefore, in designing GLS,
we view CSs as model free learning in POMDPs and take a hybrid approach to finding
the best generalization, given the total number of rules. GLS uses the policy improvement
procedure by Jaakkola et a1. for an locally optimal stochastic policy when a set of rule
conditions is given. GLS uses GA to search for the best set of rule conditions.
995
Viewing Classifier Systems as Model Free Learning in POMDPs
0.24
????? , ????? 0 . ? ??????...? .. ???????.??? 00 ? ? , ????? ,0 ?? . , . "
??
. or o . . . . . F . ???? .. . F .? ? .???. ,D . ????? . F ?? , ? . ? . ? . r o., ..... .
.. . . ?? 0 . ? . ?? . . . 00 . . . ???. F . ? ?.?.? ?. ????? ..
? . . . .? . ? ?? . ?? ?? ? . .???? .? ??? 0 . . . . . . 0 . . . . . . . .. F ?????? 0 ?? ???
. . . r ' " .. oro ..... .. . oro .. _.. . .... F . ?.???.?? 00 ???? . ? r ... .
? . ? 00 .. . . ?? . .???. . . ? ? ? . . ? ? ? ? . .. .??. 0 . . . ?..?? . ???? ? ??. . 0 ?? .
:~r~:.:: : :~~::::: : :~r: ::: : ::~::: :::: :: : ~t~ : :: :::: :~:: ::::: 'i
? ???????? ? ???.??.????????????.? ? ?.? . ?? ? ? . .??.????. 0 . ?.?? . .
. . ? 00 . . .. ?. , 0 . . . . . .
. . . F . . ..... r ......
. .... . . . . ~
. . . . . _ ? . 0 .?? . . ..? 00 ? . .. ? ?. 0 . "
. ? ?? .
. o.. . . .. . . Fo . .. . . ... F... ... .. or . . . . oro.
.. .
~
.. .. . . . ... . . . . ... ... .. . .. .. . . . .
? . 0 , , , ?? . ? . . . . . ? ? . .??.?? . 0 ? . ..? . ??.??? 0 ???? . ? 0 ?????? . 0 ????
. . r . .. . ... F . ?? . ? .. ?? . . ?.. ? F . . ? ????. . Fo . ? . ... . r .? ?. ?. oF . . .
. . 0 , .. _ . . ? 00 ?. . ? .. .? . . . . ? ?? 0 .?? ??.? , .
... ... .. . . . . . . . . . . .. . 0 ? .? ?? ? ? . ?.? "
.. . F .. .. . ? ? ? oro ...
.. . r . . . .. . . "
. . . 00 . ? . ? . . ??.?? ?. ?? . 0 . . . . . .
.??? ? ? ? ? 0 ? ?.? .??? ..
....... 0 . . ... . ....... .
F . . ? .?.?? r .
... . 0 F .. ????.?
. . . 00 . . . . ??? 0 ? .? ?? . 0 ..? ?.?? .
.----~~~~~~~.....,_.__,_"
0.23
02 2
021
0.2
0.19
0.18
0.17
0.16
0.15
0.14
L.-~~~~~~~_---.J
o
W W
~
~
~
~
~
~
~
~
geoentionJ
Figure 3: LEFT: Woods7.Each cell is either empty".", contains a stone "0", or contains food "F'.
The cells which contain a stone are not passable, and the cells which contain food are goals. In each
cell, the agent receives a 2 * 8 = 16 bit long sensor message, which tells the contents of the eight
adjacent cells. From each cell, the agent can move to one of the eight adjacent cells. When the agent
reaches a cell which contains food, he receives reward 1. The agent is then put in one of the empty
cells with equal probability.
RlGHT:Gain of the best rule set. Parameters: t ma x = 10000, to = 0.10, M = 10, N = 100, P#
0.70.
=
References
Dejong, K. A. (1988). Learning with genetic algorithms: An overview. Machine Learning, 3:121-138.
Goldberg, D. E. (1989). Genetic Algorithms in Search, Optimization, and Machine
Learning. Addison-Wesley.
Grefenstette, J. J. (1987). Multilevel credit assignment in a genetic learning system. In
Proc. Second Int. Con! on Genetic Algorithms, pp. 202-209.
Grefenstette, J. J. (1988). Credit assignment in rule discovery systems based on genetic
algorithms. Machine Learning, 3:225-245.
Grefenstette, J. J., C. L. Ramsey, and A. C. Schultz (1990). Learning sequential decision
rules using simulation and competition. Machine Learning, 5:355-381.
Holland, J. H. (1986). Escaping brittleness: the possibilities of general purpose learning
algorithms applied to parallel rule-based systems. In Machine Learning II, pp. 593623. Morgan Kaufmann.
Holland, J. H. and J. S. Reitman (1978). Cognitive systems based on adaptive algorithms. In D. A. Waterman and F. Hayes-Roth (Eds.), Pattern-directed inference
systems. Academic Press.
Jaakkola, T., S. P. Singh, and M. I. Jordan (1994). Reinforcement learning algorithm for
partially observable markov decision problems. In Advances of Neural Information
Processing Systems 7, pp. 345-352.
Kaelbling, L. P., M. L. Littman, and A. W. Moore (1996). Reinforcement learning: A
survey. Journal of Artificial Intelligence Research, 4:237-285.
McCallum, R. A. (1993). Overcoming incomplete perception with utile distinction
memory. In Proc . the Tenth Int. Con! on Machine Learning, pp. 190-196.
Westerdale, T. H. (1997). Classifier systems - no wonder they don't work. In Proc. Second Annual Genetic Programming Conference, pp. 529-537.
Wilson, S. W. (1994). Zcs: A zeroth order classifier system. Evolutionary Computation ,
2(1): 1-18.
Wilson, S. W. (1995). Classifier fitness based on accuracy. Evolutionary Computation ,
3(2): 149-175.
Wilson, S. W. and D. E. Goldberg (1989). A critical review of classifier systems. In Proc .
Third Int . Con! on Genetic Algorithms, pp. 244-255.
| 1492 |@word cu:1 faculty:1 seems:2 suitably:1 simulation:1 profit:1 carry:1 initial:2 series:1 contains:3 selecting:1 genetic:12 ramsey:1 current:5 must:1 update:3 v:6 intelligence:1 selected:1 p7:1 mccallum:5 iso:1 record:2 utile:1 consists:2 combine:1 introduce:1 alm:1 expected:1 behavior:1 aliasing:4 food:3 actual:1 estimating:1 moreover:1 discover:1 matched:1 string:4 developed:4 dejong:2 finding:3 every:1 exactly:1 classifier:17 slm:2 control:1 ner:1 local:1 tends:1 rule2:1 zeroth:2 twice:1 initialization:1 collect:1 directed:1 procedure:9 crossover:2 sga:4 matching:3 refers:1 spite:2 get:2 cannot:2 ga:9 selection:4 operator:2 put:3 gls:41 deterministic:1 roth:1 independently:1 survey:2 pomdp:6 ergodic:1 rule:104 brittleness:1 his:1 population:7 anyway:1 cs:9 hierarchy:1 programming:1 us:10 designing:3 goldberg:4 q7:2 observed:1 higashi:1 cycle:2 ran:1 environment:10 pd:3 complexity:2 reward:8 littman:1 singh:1 ithe:1 exit:1 completely:2 basis:1 easily:1 hiroshima:3 alphabet:1 monte:3 artificial:1 tell:2 whose:2 quite:2 solve:3 say:2 s:1 otherwise:1 noisy:1 itself:2 advantage:2 sequence:1 took:1 propose:1 alleviates:1 description:1 competition:1 empty:2 produce:1 bba:5 converges:1 tk:1 tim:1 oo:1 ac:1 rt2:2 progress:1 eq:3 strong:2 c:6 direction:1 stochastic:7 allele:1 viewing:4 violates:1 sand:1 multilevel:1 generalization:13 really:1 wall:1 im:1 hold:3 sufficiently:2 credit:10 considered:4 around:1 mapping:3 pitt:7 claim:1 oro:4 achieves:1 a2:1 smallest:1 purpose:1 proc:4 him:1 city:1 sensor:8 aim:5 modified:1 avoid:1 jaakkola:8 wilson:9 improvement:10 indicates:1 check:1 inference:1 entire:1 among:2 plan:1 special:2 fairly:1 equal:4 saving:2 look:1 future:3 report:1 randomly:4 manipulated:1 composed:1 individual:3 delayed:1 fitness:5 intended:2 attempt:1 message:23 highly:1 possibility:1 evaluation:9 analyzed:1 accurate:1 necessary:2 incomplete:1 aha:1 instance:1 increased:1 markovian:5 assignment:10 cost:1 kaelbling:2 wonder:1 chooses:1 st:8 thanks:1 choose:1 v7:6 cognitive:1 return:2 japan:1 availability:3 int:3 ad:3 view:3 lot:1 try:1 start:4 capability:2 parallel:1 mutation:1 irtl:1 accuracy:2 kaufmann:1 identify:1 weak:1 none:1 carlo:3 pomdps:15 history:1 reach:3 fo:2 sharing:1 ed:1 definition:1 evaluates:3 pp:6 con:3 gain:11 popular:2 lim:1 organized:1 sophisticated:1 wesley:1 tum:1 insists:1 iai:1 execute:1 just:1 receives:5 mdp:4 requiring:1 contain:2 discounting:1 memoryless:3 moore:1 attractive:1 adjacent:3 covering:3 samuel:5 stone:2 complete:1 cooperate:1 recently:1 mt:2 rl:4 overview:1 brigade:1 jp:1 he:7 refer:2 leff:1 zcs:3 recent:1 showed:1 perspective:1 disappointing:2 binary:1 vt:2 suematsu:5 seen:1 morgan:1 akira:2 care:3 somewhat:1 q1r:1 converge:1 ii:4 multiple:2 sound:1 match:6 academic:1 long:3 lai:1 a1:6 controlled:1 prediction:2 expectation:1 iteration:1 represent:1 achieved:1 cell:9 receive:2 want:1 mbx:1 goto:2 jordan:1 call:2 near:1 enough:2 suboptimal:1 escaping:1 whether:3 cause:1 afford:1 action:20 tma:1 locally:3 induces:1 tth:3 generate:1 lsi:1 problematic:1 notice:2 rtk:1 per:1 four:2 tenth:1 kept:1 graph:1 decision:3 bit:6 pay:1 guaranteed:1 annual:1 strength:8 generates:3 aspect:1 span:1 conjecture:2 according:3 combination:2 psp:5 bucket:1 taken:1 agree:1 discus:1 addison:1 end:2 apply:2 eight:2 occurrence:1 original:1 denotes:1 build:2 move:3 added:1 occurs:2 strategy:1 evolutionary:2 kth:2 equip:1 condit:1 difficult:3 design:1 policy:32 unknown:1 observation:4 markov:3 waterman:1 attacked:1 immediate:1 payoff:11 situation:1 overcoming:1 pair:3 distinction:4 learned:1 proceeds:1 pattern:1 perception:1 built:1 max:4 reliable:1 memory:1 ia:1 critical:1 difficulty:1 hybrid:6 improve:2 mdps:1 elitist:1 extract:1 review:1 literature:2 discovery:1 expect:1 generation:5 proportional:1 approximator:1 agent:15 cd:1 summary:1 repeat:5 last:2 free:7 keeping:1 formal:1 understand:1 default:4 world:4 avoids:1 maze:3 author:1 made:3 reinforcement:3 adaptive:1 avoided:1 schultz:1 observable:5 rtl:2 active:3 hayes:1 pittsburgh:1 don:4 search:4 ku:1 concatenates:1 excellent:1 necessarily:1 whole:1 fig:3 representative:1 slow:1 lc:1 position:1 perceptual:3 third:1 rk:4 specific:3 symbol:3 sensing:1 r2:3 x:6 pz:1 burden:1 exists:1 restricting:1 sequential:1 alc:1 ci:3 easier:1 michigan:11 simply:1 partially:3 hayashi:4 holland:5 corresponds:2 ma:1 grefenstette:6 viewed:2 goal:4 leaming:1 passable:1 exceptionally:1 change:1 content:1 infinite:1 except:1 total:8 called:3 e:3 rarely:1 select:1 people:1 evaluate:2 tested:1 |
539 | 1,493 | Multi-electrode spike sorting
by clustering transfer functions
Dmitry Rinberg
Hanan Davidowitz
N aftali Tishby*
NEe Research Institute
4 Independence Way
Princeton, N J 08540
E-mail: {dima,hanan, tishby }<Dreseareh. nj . nee . com
Categories: spike sorting, population coding, signal processing.
Abstract
A new paradigm is proposed for sorting spikes in multi-electrode
data using ratios of transfer functions between cells and electrodes.
It is assumed that for every cell and electrode there is a stable
linear relation. These are dictated by the properties of the tissue,
the electrodes and their relative geometries. The main advantage
of the method is that it is insensitive to variations in the shape and
amplitude of a spike. Spike sorting is carried out in two separate
steps. First, templates describing the statistics of each spike type
are generated by clustering transfer function ratios then spikes are
detected in the data using the spike statistics. These techniques
were applied to data generated in the escape response system of
the cockroach.
1
Introduction
Simultaneous recording of activity from many neurons can greatly expand our understanding of how information is coded in neural systems[l]. 11ultiple electrodes
are often used to measure the activity in neural tissue and have become a standard
tool in neurophysiology [2 , 3,4]. Since every electrode is in a different position it will
measure a different contribution from each of the different neurons . Simply stated,
the problem is this: how can these complex signals be untangled to determine when
each individual cell fired? This problem is difficult because, a) the objects being
classified are very similar and often noisy, b) spikes coming from the same cell can
?Permanent address: Institute of Computer Science and Center for Neural Computation, The Hebrew University, Jerusalem, Israel. Email: tishby<Des.huji.ae.il
147
Transfer Function Spike Sorting
vary in both shape and amplitude, depending on the previous activity of the cell
and c) spikes can overlap in time, resulting in even more complex temporal patterns.
Current approaches to spike sorting are based primarily on the presumed consistency of the spike shape and amplitude for a given cell [5, 6]. This is clearly the
only possible basis for sorting using a single electrode. Multiple electrodes, however,
provide additional independent information through the differences in the way the
same neuron is detected by the different electrodes. The same spike measured on
different electrodes can differ in amplitude, shape and its relative timing. These
differences can depend on the specific cell, the electrode and the media between
them. They can be characterized by linear transfer functions that are invariant to
changes in the overall spike waveform. In this paper the importance of this information is highlighted by using only the differences in how signals are measured on
different electrodes. It is then shown that clusters of similar differences correspond
to the same neuron. It should be emphasized that in a full treatment this transfer
function information will be combined with other cues to sort spikes.
2
Spikes, spectra and noise
The basic assumption behind the spike sorting approach described here is that the
medium between each neuron-electrode pair can be characterized by a linear system
that remains fixed during the course of an experiment. This assumption is justified
by the approximately linear dielectric properties of the electrode and its surrounding
nerve tissues.
Linear systems are described by their phase and amplitude response to pure frequencies , namely, by their complex transfer function H(w) = O(w)j I(w), where
I(w) and O(w) are the complex spectra (Le. Fourier transform, henceforth called
spectrum) of the input and output of the system, respectively. In the experiments
described here the input signal is the spectrum of the action potential generated
by cell j, denoted by Sj(w) and the output signal is the spectrum of the voltage
measured at electrode Il, denoted by VJ.L(w). The transfer function of the system
that links Sj(w) and VJ.L(w) is then defined as Hf(w) = VJ.L(w)jSj(w).
If the transfer functions are fixed in time, the ratio between the complex spectrum
of any spike from cell j as detected by electrodes Il and v , VJ.L (w) and V II (w ), is
given by,
Hf(w)
Hj(w) ,
(1)
which is independent of the cell action potential spectrum Sj(w), provided that the
spike was detected by both electrodes.
Thus, even if a spike varies in shape and amplitude, Tjll (w) will remain a fixed
complex function of frequency. This ratio is also invariant with respect to time
translations of the spikes. In addition, the frequency components are asymptotically un correlated for stationary processes, which justifies treating the frequency
components as statistically independent[7] . The idea behind the approach described
here is shown in Figure 1.
In real experiments, however, noise can corrupt the invariance of Tjll. There are
several possible sources of noise in experiments of this kind: a) fluctuations in the
transfer function, b) changes in the spike shape, <;j and c) electrical and electrochemical noise, nJ.L.
148
D. Rinberg, H. Davidowitz and N Tishby
cell-1
cell-1
cell-2
time
time
time
time
time
time
I~
~
frequency
frequency
Q)
u
::l
a.
E
ctl
Q)
u
::l
:::
a.
E
ctl
c::
0
u
-- c::
::::I
0
~
ctl
~
Q)
( /)
c::
ctl
~
frequency
Figure 1: The idea behind spike sorting by clustering of transfer function ratios.
Two spikes from the same cell (cell-I) may vary in shape/ amplitude during bursting
activity, for example. Although the spike shapes may differ, the transfer functions
relating them to the electrodes do not change so the transfer function ratios are
similar (two left columns). A different cell (cell-2) has a different transfer function
ratio even though the spikes shapes themselves are similar to those of cell-l (right
column).
Hf
If
varies slowly with time, the transfer function noise is small relative to
and n Y ?
can then be expanded to first order in <;j, nJ.l. and n Y as
Try
<;j,
nY
(2)
which is independent of <;j. Since the noise, nJ.l., is un correlated with the spike signal,
5 j , the variance at each frequency component can be considered to be Gaussian
with equal variances on the real and imaginary axes. Thus the mean of
will be
independent of 5 j , <;j and nJ.l. while its variance will be inversely proportional to 5 j .
Try
3
A model system: the escape response of the cockroach
These techniques were tested on a relatively simple neural system - the escape
response system of the American cockroach. The escape behaviour, which has been
studied extensively [9, 10, 11], is activated when the insect detects air currents
149
Transfer Function Spike Sorting
it
"
!~
f?- ,"-,
(
-
I
: 10 ms
. t \\ ..
~
1. lt \'\
10.5mv
50 ms
Figure 2: A schematic representation of the experiment. Typical raw data measured on two electrodes is shown at right. Relative time delays are evident in the
inset, but are not a necessary condition for the sorting techniques described here.
Abbreviations are: p-puffers, cg-circal ganglion, c-cerci.
produced by the movements of a predator. The insect detects the approach of a
predator, determines the direction of approach and starts running in an appropriate
direction. The cockroach does this by detecting the movement of several hundred
fine hairs located on two appendages, called cerci, protruding from the posterior
end of the animal. Each of these hairs is connected to a single neuron. Axons
from these cells converge on a dense neuropil called the cercal ganglion (cg), where
directional information is extracted and conveyed to the rest of the body by axons
in the abdominal nerve. This is shown schematically in Figure 2.
This system proved to be well suited as a first test of the sorting technique. The
system is simple enough so that it is not overwhelming (since only 7 neurons are
known to contribute to the code) but complex enough to really test the approach .
In addition, the nerve cords are linear in geometry, easily accessible and very stable.
Male cockroaches (Periplaneta americana) were dissected from the dorsal side to
expose the nerve cord. The left and right cords were gently separated and two tungsten wire electrodes were hooked onto the connective about 2 mm apart, separated
by abdominal ganglia. The stimulus was presented by two loudspeakers driving
two miniature wind tunnels pointed at the cerci, at 90 degrees from one another
as shown in Figure 2. Recordings typically lasted for several hours. Data were
collected with a sampling frequency of 2 . 104 Sis which was sufficient to preserve
the high frequency components of the spikes.
D. Rinberg, H. Davidowitz and N. Tishby
150
0.5
1
-1.5
-1
Re(T)
1
Figure 3: Real and imaginary parts of Tr v a single w. The circles have centers
(radii) equal to the average (variance) of Tr v at w = 248.7 rad S-l. Note that
while some clusters seem to overlap at this frequency they may be well seperated
at others. Cluster-l is dispersed throughout the complex plane and its variance is
well beyond the range of this plot.
4
Clustering and the detection of spikes
The spike sorting algorithm described here is done is two separate stages. First,
a statistical model of the individual spike types is built from "clean" examples
found in the data. Only then are occurrences of these spikes detected in the multielectrode data. This two-step arrangement allows a great deal of flexibility by
disconnecting the clustering from the detection. For example, here the clustering
was done on transfer function ratios while the detection was done on full complex
spectra. These stages are described below in more detail.
4.1
The clustering phase
First, the multi-electrode recording is chopped into 3 ms long frames using a sliding
window. Frames that have either too low total energy or too high energy at the
window edges are discarded. This leaves frames that are energetic in their central
2 ms and are assumed to carry one spike. No attempt is made to find all spikes in
the data. Instead, the idea is to generate a set of candidate spike types from clean
frames.
Once a large collection of candidate spikes is found, TrV(w) is calculated for every
151
Transfer Function Spike Sorting
Hook #1
cluster
Hook #2
frames
yCluster
6
324
~
5
d
4
rvv
V-
518
'-"
748
&
3
cs
2
~
1
ca
&
720
-'lr
-'\I
z
7E
2
R &
729
?
..-
~6
I!!!I!
R
g
.... --
4V
E2?@?C9
500~V
~.I",
I
ru@
1 w?
Figure 4: Results of clustering spikes using transfer function ratios. Note that
although cluster-5 and cluster-6 are similarly shaped on hook-l they are time shifted
on hook-3. Cluster-l is made up of overlaps which are dealt with in the detection
phase.
Ttl!
(w).
spike. These are then grouped together into clusters containing similar
Results of the clustering are shown in Figure 3 while the corresponding waveforms
are shown in Figure 4. Full complex spectra are then used to build a statistical model of the different spike types, {Vj(w), af(w)}, which represent each cell's
action potential as it appears on each of the electrodes.
4.2
The detection phase
Once the cluster statistics are determined , an independent detection algorithm is
used. The data is again broken into short frames but now the idea is to find which
of the spike types (represented by the different clusters found in the previous steps)
best represents the data in that frame. Each frame can contain either noise, a spike
or an overlap of 2 spikes (overlaps of more than 2 spikes are not dealt with). This
part is not done on transfer function ratios because dealing with overlaps is more
difficult.
5
Concl usion
A new method of spike sorting using transfer function ratios has been presented.
In effect the sorting is done on the properties of the tissue between the neuron and
152
D. Rinberg, H Davidowitz and N. Tishby
the electrode and individual spike shapes become less important. This method may
be useful when dealing with bursting cells where the transfer function ratios should
remain constant even though the spike amplitude can change significantly. This
technique may prove to be a useful tool for analysing multi-electrode data.
Acknowledgments
We are grateful to Bill Bialek for numerous enlightening discussions and many useful
suggestions.
References
[1] M. Barinaga. Listening in on the brain. Science 280, 376-378 (1998).
[2] M. Abeles. Coriiconics, (Cambridge University Press, Cambridge, 1991)
[3] B.L. McNaughton, J. O 'Keefe and C.A. Barnes. The stereotrode: a new technique for simultaneous isolation of several single units in the central nervous
system from multiple unit records. Journal of Neuroscience Methods, 8, 391-7
(1983).
[4] M.L. Reece and J. O'Keefe. The tetrode: a new technique for multi-unit extracellular recording. Society of Neuroscience Abstracts 15, 1250 (1989).
[5] M.S. Fee, P. P. Mitra and D. Kleinfeld. Automatic sorting of multiple unit
neuronal signals in the presence of anisotropic and non-Gaussian variability.
Journal of Neuroscience Methods 69, 175-188 (1996).
[6] M.S. Lewicki. A review of methods for spike sorting: the detection and classification of neural potentials. Network: Compututational Neural Systems 9 ,
R53-R78 (1998).
[7] A. Papoulis. Probability, random variables and stochastic processes, (McGrawHill, New-York, 1965).
[8] M. Abeles and G.L. Gerstein. Detecting spatio-temporal firing patterns among
simultaneously recorded single neurons. Journal of Neurophysiology 60(3) ,
909-924 (1988).
[9] J .M. Camhi and A. Levy. The code for stimulus direction in a cell assembly in
the cockroach. Journal of Comparative Physiology A 165 , 83-97 (1989).
[10] L. Kolton and J.M . Camhi. Cartesian representation of stimulus direction:
parallel processing by two sets of giant interneurons in the cockroach. Journal
of Comparative Physiology A 176, 691-702 (1995).
[11] J. Westin, J.J. Langberg and J.M. Camhi. Responses of Giant Interneurons of
the cockroach Periplaneta americana to wind puffs of different directions and
velocities . Journal of Comparative Physiology A 121,307-324 (1977) .
| 1493 |@word neurophysiology:2 tr:2 papoulis:1 carry:1 imaginary:2 current:2 com:1 si:1 shape:10 treating:1 plot:1 stationary:1 cue:1 leaf:1 nervous:1 plane:1 short:1 record:1 lr:1 detecting:2 contribute:1 become:2 prove:1 presumed:1 themselves:1 multi:5 brain:1 detects:2 overwhelming:1 window:2 trv:1 provided:1 medium:2 israel:1 ttl:1 kind:1 connective:1 giant:2 nj:5 temporal:2 every:3 dima:1 unit:4 timing:1 mitra:1 fluctuation:1 firing:1 approximately:1 studied:1 bursting:2 tungsten:1 range:1 statistically:1 acknowledgment:1 significantly:1 physiology:3 onto:1 disconnecting:1 r78:1 davidowitz:4 bill:1 center:2 jerusalem:1 pure:1 periplaneta:2 population:1 variation:1 mcnaughton:1 velocity:1 located:1 electrical:1 cord:3 connected:1 movement:2 rinberg:4 broken:1 depend:1 grateful:1 basis:1 easily:1 represented:1 hooked:1 surrounding:1 separated:2 reece:1 detected:5 dielectric:1 statistic:3 highlighted:1 noisy:1 transform:1 advantage:1 coming:1 fired:1 flexibility:1 untangled:1 electrode:25 cluster:10 comparative:3 object:1 depending:1 measured:4 c:1 differ:2 direction:5 waveform:2 radius:1 stochastic:1 dissected:1 behaviour:1 really:1 mm:1 considered:1 great:1 miniature:1 driving:1 vary:2 expose:1 grouped:1 ctl:4 tool:2 clearly:1 gaussian:2 hj:1 voltage:1 ax:1 lasted:1 greatly:1 cerci:3 cg:2 camhi:3 typically:1 relation:1 expand:1 overall:1 hanan:2 classification:1 among:1 denoted:2 insect:2 animal:1 equal:2 once:2 shaped:1 sampling:1 represents:1 others:1 stimulus:3 escape:4 primarily:1 preserve:1 simultaneously:1 individual:3 geometry:2 phase:4 attempt:1 detection:7 abdominal:2 interneurons:2 male:1 activated:1 behind:3 usion:1 edge:1 necessary:1 re:1 circle:1 column:2 hundred:1 delay:1 tishby:6 too:2 varies:2 abele:2 combined:1 huji:1 accessible:1 together:1 again:1 central:2 recorded:1 containing:1 slowly:1 henceforth:1 american:1 potential:4 de:1 coding:1 permanent:1 mv:1 try:2 wind:2 mcgrawhill:1 start:1 sort:1 hf:3 parallel:1 predator:2 contribution:1 jsj:1 air:1 il:3 variance:5 correspond:1 directional:1 dealt:2 raw:1 produced:1 tissue:4 classified:1 simultaneous:2 email:1 energy:2 frequency:11 e2:1 proved:1 treatment:1 amplitude:8 nerve:4 appears:1 response:5 done:5 though:2 stage:2 kleinfeld:1 r53:1 effect:1 contain:1 deal:1 during:2 m:4 evident:1 insensitive:1 gently:1 anisotropic:1 relating:1 cambridge:2 automatic:1 consistency:1 similarly:1 pointed:1 concl:1 stable:2 posterior:1 dictated:1 apart:1 additional:1 determine:1 paradigm:1 converge:1 signal:7 ii:1 sliding:1 multiple:3 full:3 characterized:2 af:1 long:1 coded:1 schematic:1 basic:1 hair:2 ae:1 represent:1 cell:22 justified:1 addition:2 schematically:1 fine:1 chopped:1 source:1 rest:1 recording:4 seem:1 presence:1 enough:2 independence:1 isolation:1 idea:4 listening:1 energetic:1 york:1 action:3 tunnel:1 useful:3 extensively:1 category:1 generate:1 shifted:1 neuroscience:3 cercal:1 clean:2 appendage:1 asymptotically:1 throughout:1 gerstein:1 fee:1 multielectrode:1 barnes:1 activity:4 fourier:1 expanded:1 relatively:1 extracellular:1 loudspeaker:1 remain:2 invariant:2 remains:1 describing:1 end:1 appropriate:1 occurrence:1 clustering:9 running:1 assembly:1 build:1 society:1 arrangement:1 spike:50 bialek:1 separate:2 link:1 mail:1 collected:1 ru:1 code:2 ratio:12 hebrew:1 difficult:2 stated:1 neuron:9 wire:1 discarded:1 variability:1 frame:8 pair:1 namely:1 rad:1 nee:2 hour:1 address:1 beyond:1 below:1 pattern:2 built:1 enlightening:1 overlap:6 cockroach:8 inversely:1 numerous:1 hook:4 carried:1 review:1 understanding:1 relative:4 suggestion:1 proportional:1 degree:1 conveyed:1 sufficient:1 corrupt:1 translation:1 course:1 side:1 institute:2 template:1 protruding:1 calculated:1 made:2 collection:1 sj:3 dmitry:1 dealing:2 langberg:1 assumed:2 spatio:1 spectrum:9 un:2 seperated:1 transfer:22 ca:1 neuropil:1 complex:10 vj:5 main:1 dense:1 noise:7 body:1 neuronal:1 ny:1 axon:2 position:1 candidate:2 levy:1 specific:1 emphasized:1 inset:1 americana:2 tetrode:1 keefe:2 importance:1 justifies:1 cartesian:1 sorting:18 westin:1 suited:1 lt:1 simply:1 ganglion:3 lewicki:1 determines:1 dispersed:1 extracted:1 abbreviation:1 change:4 analysing:1 typical:1 determined:1 called:3 total:1 invariance:1 puff:1 dorsal:1 c9:1 princeton:1 tested:1 correlated:2 |
540 | 1,494 | Where does the population vector of motor
cortical cells point during reaching movements?
Pierre Baraduc*
pbaraduc@snv.jussieu.fr
Emmanuel Guigon
guigon@ccr.jussieu.fr
Yves Burnod
ybteam@ccr.jussieu.fr
INSERM U483, Universite Pierre et Marie Curie
9 quai St Bernard, 75252 Paris cedex 05, France
Abstract
Visually-guided arm reaching movements are produced by distributed
neural networks within parietal and frontal regions of the cerebral cortex.
Experimental data indicate that (I) single neurons in these regions are
broadly tuned to parameters of movement; (2) appropriate commands are
elaborated by populations of neurons; (3) the coordinated action of neurons can be visualized using a neuronal population vector (NPV). However, the NPV provides only a rough estimate of movement parameters
(direction, velocity) and may even fail to reflect the parameters of movement when arm posture is changed. We designed a model of the cortical
motor command to investigate the relation between the desired direction
of the movement, the actual direction of movement and the direction of
the NPV in motor cortex. The model is a two-layer self-organizing neural
network which combines broadly-tuned (muscular) proprioceptive and
(cartesian) visual information to calculate (angular) motor commands for
the initial part of the movement of a two-link arm. The network was
trained by motor babbling in 5 positions. Simulations showed that (1)
the network produced appropriate movement direction over a large part
of the workspace; (2) small deviations of the actual trajectory from the
desired trajectory existed at the extremities of the workspace; (3) these
deviations were accompanied by large deviations of the NPV from both
trajectories. These results suggest the NPV does not give a faithful image
of cortical processing during arm reaching movements.
? to whom correspondence should be addressed
84
P. Baraduc, E. Guigon and Y. Burnod
1 INTRODUCTION
When reaching to an object, our brain transforms a visual stimulus on the retina into a
finely coordinated motor act. This complex process is subserved in part by distributed
neuronal populations within parietal and frontal regions of the cerebral cortex (Kalaska
and Crammond 1992). Neurons in these areas contribute to coordinate transformations
by encoding target position and kinematic parameters of reaching movements in multiple frames of reference and to the elaboration of motor commands by sending directional
and positional signals to the spinal cord (Georgopoulos 1996). An ubiquitous feature of
cortical populations is that most neurons are broadly tuned to a preferred attribute (e.g.
direction) and that tuning curves are uniformly (or regularly) distributed in the attribute
space (Georgopoulos 1996). Accordingly, a powerful tool to analyse cortical populations
is the NPV which describes the behavior of a whole population by a single vector (Georgopoulos 1996). Georgopoulos et al. (1986) have shown that the NPV calculated on a set
of directionally tuned neurons in motor cortex points approximately (error
15?) in the
direction of movement. However, the NPV may fail to indicate the correct direction of
movement when the arm is in a particular posture (Scott and Kalaska 1995). These data
raise two important questions: (1) how populations of broadly tuned neurons learn to compute a correct sensorimotor transformation? Previous models (Burnod et al. 1992; Bullock
et al. 1993; Salinas and Abbott 1995) provided partial solutions to this problem but we still
lack a model which closely matches physiological and psychophysical data on reaching
movements; (2) Are cortical processes involved in the visual guidance of arm movements
readable with the NPV tool? This article provides answers to these questions through a
physiologically inspired model of sensorimotor transformations.
f"o.J
2 MODEL OF THE VISUAL-TO-MOTOR TRANSFORMATION
2.1
ARM GEOMETRY
The arm model has voluntarily been chosen simple. It is a planar, two-link arm, with
limited (160 degrees) joint excursion at shoulder and elbow. An agonist/antagonist pair is
attached at each joint.
2.2
INPUT AND OUTPUT eODINGS
No cell is finely tuned to a specific input or output value to mimic the broad tunings or
monotonic firing characteristics found in cortical visuomotor areas.
2.2.1
Arm position
By analogy with the role of muscle spindles, proprioceptive sensors are assumed to code
muscle length. Arm position is thus represented by the population activity of NT = 20
neurons coding for the length of each agonist or antagonist. The activity of a sensor neuron
k is defined by:
Tk
where
LIl ( k)
=
adLn(k))
is the length of muscle number n(k) , and
ak
a piecewise linear sigmoid:
L ~ Ak
Ak < L
< Ak
L? Ak
Sensibility thresholds Ak are uniformly distributed in [L min , L max], and the dynamic range
is Ak - Ak is taken constant, equal to Lmax - L min .
85
Population Coding ofReaching Movements
2.2.2
Desired direction
The direction V of the desired movement in visual space is coded by a population of
N x = 50 neurons with cosine tuning in cartesian space. Each visual neuron j thus fires as:
Xj
= V? Vj
being the preferred direction of the cell. These 50 preferred directions are chosen
uniformly distributed in 2-D space.
Vj
2.2.3
Motor Command
In attempt to model the existence of muscular synergies (Lemon 1988), we identified motor command with joint movement rather than with muscle contraction . A motor neuron
i among Nt = 50 contributes to the effective movement M by its action on a synergy
(direction in joint space) Mi. This collective effect is formally expressed by:
M .= LtiMi
where ti is the activity of motor neuron i. The 50 directions of action Mi are supposed
uniformly distributed in joint space.
3
3.1
NETWORK STRUCTURE AND LEARNING
STRUCTURE OF THE NETWORK
Information concerning the position of the arm and the desired direction in cartesian space
desired
=. . . .~(visual)
0-
~,
cY
direction
~~~~----~
.....
Cf
t;
~~~~t???
motor synergy
Figure 1: Network Architecture
is combined asymmetrically (Fig. I). First, an intermediate (somatic) layer of neurons
P. Baraduc. E. Guigon and Y. Bumod
86
forms an internal representation of the arm position by a combination of the input from the
NT muscle sensors and the lateral interactions inside the population. Activity in this layer
is expressed by:
(1)
Sij =
Wijk Tk +
ljp Sip
L
L
k
p
where the lateral connections are:
ljp
= cos (27r(j -
p)/NT
)
Equation 1 is self-referent; so calculation is done in two steps. The feed-forward input
first arrives at time zero when there is no activity in the layer; iterated action of the lateral
connections comes into play when this feed-forward input vanishes.
The activity in the somatic layer is then combined with the visual directional information
by the output sigma-pi neurons as follows:
ti =
L
Xj Sij
j
3.2
WEIGHTS AND LEARNING
The only adjustable weights are the Wijk linking the proprioceptive layer to the somatic
layer. Connectivity is random and not complete: only 15% of the somatic neurons receive
information on arm position. The visuomotor mapping is learnt by modifying the internal
representation of the arm.
Motor commands issued by the network are correlated with the visual effect of the movement ("motor babbling"). More precisely, the learning algorithm is a repetition of the
following cycle:
1. choice of an arm position among 5 positions (stars on Fig. 2)
2. random emission of a motor command (ti)
3. corresponding visual reafference (Xj)
4. weight modification according to a variant of the delta rule:
c:'Wijk
oc
(tiXj - Sij) Tk
The random commands are gaussian distributions of activity over the output layer. 5000
learning epochs are sufficient to obtain a stabilized performance. It must be noted that
the error between the ideal response of the network and the actual performance never decreases completely to zero, as the constraints of the visuomotor transformation vary over
the workspace.
4 RESULTS
4.1
NETWORK PERFORMANCE
Correct learning of the mapping was tested in 21 positions in the workspace in a pointing
task toward 16 uniformly distributed directions in cartesian space. Movement directions
generated by the network are shown in Fig. 2 (desired direction 0 degree is shown bold).
Norm of movement vectors depends on the global activity in the network which varies with
arm position and movement direction.
Performance of the network is maximal near the learning positions. However, a good generalization is obtained (directional error 0.3?, SD 12.1?); a bias toward the shoulder can be
observed in extreme right or left positions. A similar effect was observed in psychophysical
experiments (Ghilardi et a1. 1995).
87
Population Coding of Reaching Movements
90
180.0
270
Figure 2: Performance in a pointing task
4.2
4.2.1
PREFERRED DIRECTIONS AND POPULATION VECTOR
Behavior of the population vector
Preferred directions (PO) of output units were computed using a multilinear regression;
a perfect cosine tuning was found, which is a consequence of the exact multiplication in
sigma-pi neurons. Then, the population vector, the effective movement vector, and the
desired movement were compared (Fig. 3) for two different arm configurations A and B
marked on Fig. 2. The movement generated by the network (dashed arrow) is close to the
~,
~1'
, ,
deSired direction
contnbution of one neuron
population vector ~
actual movement ......... ;:.,
Figure 3: Actual movement and population vector in two arm positions
desired one (dotted rays) for both arm configurations. However, the population vector (solid
arrow) is not always aligned with the movement. The discrepancy between movement and
population vector depends both on the direction and the position of the arm: it is maximal
88
P Baraduc, E. Guigon and Y Burnod
for positions near the borders of the workspace as position B. Fig. 3 (position B) shows that
the deviations of the population vector are due to the anisotropic distribution of the PDs in
cartesian space for given positions.
4.2.2
Difference between direction of action and preferred direction
Marked anisotropy in the distribution of PDs is compatible with accurate performance. To
see why, let us call "direction of action" (DA) the motor cell's contribution to the movement. The distribution of DAs presents an anisotropy due to the geometry of the arm. This
anisotropy is canceled by the distribution of PDs. Mathematically, if U is a N x 2 matrix of
uniformly distributed 2D vectors, the PD matrix is UJ- 1 whereas the DA matrix is UJ T ,
J being the jacobian of the angular-to-cartesian mapping. Difference between DA and PD
has been plotted with concentric arcs for four representative neurons at 21 arm positions
in Fig. 4. Sign and magnitude of the difference vary continuously over the workspace and
neuron number
/
4
Vi:
DA,
_
clockwise
=
counterclockwise
" ..
Figure 4: Difference between direction of action and preferred direction for four units.
often exceed 45 degrees. It can also be noted that preferred directions rotate with the arm
as was experimentally noted by (Caminiti et a1. 1991).
5 DISCUSSION
We first asked how a network of broadly tuned neurons could produce visually guided
arm movements. The model proposed here produces a correct behavior over the entire
workspace. Biases were observed at the extreme right and left which closely resemble experimental data in humans (Ghilardi et a1. 1995). Single cells in the output layer behave as
motor cortical cells do and the NPV of these cells correctly indicated the direction of movement for hand positions in the central region of the workspace (see Caminiti et al. 1991).
Models of sensorimotor transformations have already been proposed. However they either
considered motor synergies in cartesian coordinates (Burnod et a1. 1992), or used sharply
Population Coding ofReaching Movements
89
tuned units (Bullock et al. 1993), or motor effects independent of arm position (Salinas
and Abbott 1995). Next, the use of the NPV to describe cortical activity was questioned.
A fundamental assumption in the calculation of the NPV is that the PD of a neuron is the
direction in which the arm would move if the neuron were stimulated. The model shows
that the two directions DA and PD do not necessarily coincide, which is probably the case
in motor cortex (Scott and Kalaska 1995). It follows that the NPV often points neither
in the actual movement direction nor in the desired movement direction (target direction),
especially for unusual arm configurations. A maximum-likelihood estimator does not have
these flaws; it would however accurately predict the desired movement out of the output
unit activities, even for a wrong actual movement. In conclusion: (l) the NPV does not
provide a faithful image of cortical visuomotor processes; (2) a correct NPV should be
based on the DAs, which cannot easily be determined experimentally; (3) planning of trajectories in space cannot be realized by the successive recruitment of motor neurons whose
PDs sequentially describe the movement.
References
Bullock, D., S. Grossberg, and F. Guenther (1993). A self-organizing neural model of
motor equivalent reaching and tool use by a multijoint arm. J Cogn Neurosci 5(4), 408435.
Burnod, Y., P. Grandguillaume, I. Otto, S. Ferraina, P. Johnson, and R Caminiti (1992).
Visuomotor transformations underlying arm movements toward visual targets: a neural
network model of cerebral cortical operations. J Neurosci 12(4), 1435-53.
Caminiti, R, P. Johnson, C. Galli, S. Ferraina, and Y. Burnod (1991). Making arm movements within different parts of space: the premotor and motor cortical representation of a
coordinate system for reaching to visual targets. J Neurosci 11(5), 1182-97.
Georgopoulos, A (1996). On the translation of directional motor cortical commands to
activation of muscles via spinal interneuronal systems. Brain Res Cogn Brain Res 3(2),
151-5.
Georgopoulos, A, A Schwartz, and R Kettner (1986). Neuronal population coding of
movement direction . Science 233(4771), 1416-9.
Ghilardi, M. , J. Gordon, and C. Ghez (1995). Learning a visuomotor transformation in a
local area of work space pr oduces directional biases in other areas. J NeurophysioI73(6),
2535-9.
Kalaska, J. and D. Crammond (1992). Cerebral cortical mechanisms of reaching movements. Science 255(5051),1517-23.
Lemon, R (1988). The output map of the primate motor cortex. Trends Neurosci 11 (II),
501-6.
Salinas, E. and L. Abbott (1995). Transfer of coded information from sensory to motor
networks. J Neurosci 15(10),6461-74.
Scott, S. and J. Kalaska (1995). Changes in motor cortex activity during reaching movements with similar hand paths but different arm postures. J Neurophysioi 73(6), 2563-7.
| 1494 |@word norm:1 simulation:1 contraction:1 solid:1 initial:1 configuration:3 tuned:8 nt:4 activation:1 must:1 motor:29 designed:1 accordingly:1 provides:2 contribute:1 successive:1 combine:1 ray:1 inside:1 guenther:1 behavior:3 nor:1 planning:1 brain:3 inspired:1 anisotropy:3 actual:7 elbow:1 provided:1 underlying:1 transformation:8 sensibility:1 act:1 ti:3 sip:1 wrong:1 schwartz:1 interneuronal:1 unit:4 local:1 sd:1 consequence:1 encoding:1 ak:8 extremity:1 firing:1 path:1 approximately:1 co:1 limited:1 range:1 grossberg:1 faithful:2 cogn:2 area:4 suggest:1 caminiti:4 close:1 cannot:2 equivalent:1 map:1 rule:1 estimator:1 population:23 coordinate:3 target:4 play:1 exact:1 velocity:1 trend:1 observed:3 role:1 calculate:1 cy:1 region:4 cord:1 cycle:1 movement:45 decrease:1 voluntarily:1 vanishes:1 pd:8 asked:1 dynamic:1 trained:1 raise:1 completely:1 po:1 joint:5 easily:1 represented:1 effective:2 describe:2 visuomotor:6 salina:3 whose:1 premotor:1 otto:1 analyse:1 directionally:1 interaction:1 maximal:2 fr:3 aligned:1 organizing:2 supposed:1 produce:2 perfect:1 object:1 tk:3 resemble:1 indicate:2 come:1 direction:37 guided:2 closely:2 correct:5 attribute:2 modifying:1 human:1 generalization:1 multilinear:1 mathematically:1 considered:1 visually:2 mapping:3 predict:1 pointing:2 vary:2 multijoint:1 repetition:1 tool:3 rough:1 sensor:3 gaussian:1 always:1 reaching:11 rather:1 command:10 emission:1 referent:1 likelihood:1 flaw:1 entire:1 relation:1 france:1 canceled:1 among:2 equal:1 never:1 broad:1 mimic:1 discrepancy:1 stimulus:1 piecewise:1 gordon:1 retina:1 geometry:2 fire:1 attempt:1 investigate:1 kinematic:1 wijk:3 arrives:1 extreme:2 accurate:1 partial:1 desired:12 plotted:1 re:2 guidance:1 deviation:4 johnson:2 answer:1 varies:1 learnt:1 combined:2 st:1 spindle:1 fundamental:1 workspace:8 continuously:1 ghez:1 connectivity:1 reflect:1 central:1 accompanied:1 star:1 coding:5 bold:1 coordinated:2 depends:2 vi:1 curie:1 elaborated:1 contribution:1 yves:1 characteristic:1 directional:5 iterated:1 accurately:1 produced:2 agonist:2 trajectory:4 sensorimotor:3 involved:1 universite:1 mi:2 ubiquitous:1 feed:2 planar:1 burnod:7 response:1 done:1 angular:2 hand:2 lack:1 indicated:1 effect:4 proprioceptive:3 during:3 self:3 noted:3 cosine:2 oc:1 antagonist:2 complete:1 image:2 snv:1 sigmoid:1 spinal:2 attached:1 cerebral:4 anisotropic:1 linking:1 tuning:4 cortex:7 showed:1 issued:1 muscle:6 babbling:2 signal:1 dashed:1 multiple:1 clockwise:1 ii:1 match:1 calculation:2 kalaska:5 elaboration:1 concerning:1 coded:2 a1:4 variant:1 regression:1 cell:7 receive:1 whereas:1 addressed:1 finely:2 probably:1 cedex:1 counterclockwise:1 regularly:1 call:1 near:2 ideal:1 intermediate:1 exceed:1 xj:3 architecture:1 identified:1 questioned:1 action:7 ghilardi:3 transforms:1 visualized:1 stabilized:1 dotted:1 sign:1 delta:1 correctly:1 ccr:2 broadly:5 four:2 threshold:1 marie:1 neither:1 abbott:3 recruitment:1 powerful:1 excursion:1 layer:9 correspondence:1 existed:1 activity:11 lemon:2 precisely:1 constraint:1 sharply:1 georgopoulos:6 min:2 according:1 combination:1 describes:1 bullock:3 modification:1 making:1 primate:1 sij:3 pr:1 taken:1 equation:1 fail:2 mechanism:1 sending:1 unusual:1 operation:1 appropriate:2 pierre:2 existence:1 cf:1 readable:1 emmanuel:1 uj:2 especially:1 psychophysical:2 move:1 question:2 realized:1 posture:3 already:1 link:2 lateral:3 whom:1 toward:3 code:1 length:3 sigma:2 lil:1 collective:1 adjustable:1 neuron:24 arc:1 behave:1 parietal:2 shoulder:2 frame:1 somatic:4 concentric:1 pair:1 paris:1 connection:2 scott:3 max:1 arm:32 epoch:1 multiplication:1 analogy:1 degree:3 sufficient:1 article:1 pi:2 translation:1 compatible:1 changed:1 lmax:1 bias:3 distributed:8 curve:1 calculated:1 cortical:14 sensory:1 forward:2 coincide:1 preferred:8 synergy:4 inserm:1 global:1 sequentially:1 assumed:1 physiologically:1 why:1 subserved:1 stimulated:1 learn:1 kettner:1 transfer:1 contributes:1 complex:1 necessarily:1 vj:2 da:7 neurosci:5 arrow:2 whole:1 border:1 reafference:1 neuronal:3 fig:7 representative:1 position:22 jussieu:3 jacobian:1 specific:1 physiological:1 magnitude:1 cartesian:7 visual:12 positional:1 expressed:2 monotonic:1 neurophysioi:1 marked:2 ljp:2 experimentally:2 change:1 muscular:2 determined:1 uniformly:6 crammond:2 bernard:1 asymmetrically:1 experimental:2 formally:1 internal:2 npv:15 rotate:1 frontal:2 tested:1 correlated:1 |
541 | 1,495 | A High Performance k-NN Classifier Using a
Binary Correlation Matrix Memory
Ping Zhou
zhoup@cs.york.ac.uk
Jim Austin
austin@cs.york.ac.uk
John Kennedy
johnk@cs.york.ac.uk
Advanced Computer Architecture Group
Department of Computer Science
University of York, York YOW 500, UK
Abstract
This paper presents a novel and fast k-NN classifier that is based on a
binary CMM (Correlation Matrix Memory) neural network. A robust
encoding method is developed to meet CMM input requirements . A
hardware implementation of the CMM is described, which gives over 200
times the speed of a current mid-range workstation, and is scaleable to
very large problems. When tested on several benchmarks and compared
with a simple k-NN method, the CMM classifier gave less than I % lower
accuracy and over 4 and 12 times speed-up in software and hardware
respectively.
1 INTRODUCTION
Pattern classification is one of most fundamental and important tasks, and a k-NN rule is
applicable to a wide range of classification problems. As this method is too slow for many
applications with large amounts of data, a great deal of effort has been put into speeding it
up via complex pre-processing of training data, such as reducing training data (Dasarathy
1994) and improving computational efficiency (Grother & Candela 1997). This work
investigates a novel k-NN classification method that uses a binary correlation matrix
memory (CMM) neural network as a pattern store and match engine. Whereas most neural
networks need a long iterative training time, a CMM is simple and quick to train. It
requires only one-shot storage mechanism and simple binary operations (Willshaw &
Buneman 1969), and it has highly flexible and fast pattern search ability. Therefore, the
combination of CMM and k-NN techniques is likely to result in a generic and fast
classifier. For most classification problems, patterns are in the form of multi-dimensional
real numbers, and appropriate quantisation and encoding are needed to convert them into
binary inputs to a CMM. A robust quantisation and encoding method is developed to meet
requirements for CMM input codes , and to overcome the common problem of identical
data points in many applications, e .g. background of images or normal features in a
diagnostic problem.
Many research projects have applied the CMM successfully to commercial problems, e.g.
symbolic reasoning in the AURA (Advanced Uncertain Reasoning Architecture) approach
P. Zhou. J. Austin and J. Kennedy
714
(Austin 1996), chemical structure matching and post code matching. The execution of the
CMM has been identified as the bottleneck. Motivated by the needs of these applications
for a further high speed processing, the CMM has been implemented in dedicated
hardware, i.e. the PRESENCE architecture. The primary aim is to improve the execution
speed over conventional workstations in a cost-effective way.
The following sections discuss the CMM for pattern classification, describe the
PRESENCE architecture (the hardware implementation of CMM), and present
experimental results on several benchmarks.
2 BINARY CMM k-NN CLASSIFIER
The key idea (Figure I) is to use a CMM to pre-select a smaIl sub-set of training patterns
from a large number of training data, and then to apply the k-NN rule to the sub-set. The
CMM is fast but produces spurious errors as a side effect (Turner & Austin 1997); these
are removed through the application of the k-NN rule. The architecture of the CMM
classifier (Figure I) includes an encoder (detailed in 2.2) for quantising numerical inputs
and generating binary codes, a CMM pattern store and match engine and a conventional kNN module as detailed below .
Training patterns
stored in CMM
Patterns preselected by CMM
?
?
?
k-NN patterns
?
?
?
?
?
I
I
~B~r:~~k
classification
Figure 1: Architecture of the binary CMM k-NN classifier
2.1 PATTERN MATCH AND CLASSIFICATION WITH CMM
A correlation matrix memory is basically a single layer network with binary weights M. In
the training process a unique binary vector or separator s, is generated to label an unseen
input binary vector P,; the CMM learns their association by performing the following
logical ORing operation:
M=VS,TPi
(1)
i
In a recall process, for a given test input vector
Vk=MPJ=(ysTpl,J
Pk' the CMM performs :
(2)
followed by thresholding v k and recovering individual separators. For speed, it is
appropriate to use a fixed thresholding method and the threshold is set to a level
proportional to the number of 'I' bits in the input pattern to allow an exact or partial
match. To understand the recall properties of the CMM , consider the case where a known
pattern Pk is represented, then Equation 2 can be written as the following when two
different patterns are orthogonal to each other:
(3)
where np is a scalar, i.e. the number of 'I' bits in P k ' and P,P: =0 for i:;; k . Hence a
perfect recall of Sk can be obtained by thresholding v, at the level n" . In practice
'partially orthogonal' codes may be used to increase the storage capacity of the CMM and
the recall noise can be removed via appropriately thresholding vk (as p,p[ ~ n p for i :;; k )
RNNs Can Learn Symbol-Sensitive Counting
715
and post-processing (e.g. applying k-NN rule). Sparse codes are usually used, i.e. only a
few bits in SA and P, being set to 'I' , as this maximises the number of codes and
minimises the computation time (Turner & Austin 1997). These requirements for input
codes are often met by an encoder as detailed below.
The CMM exhibits an interesting 'partial match' property when the data dimensionality d
is larger than one and input vector p; consists of d concatenated components. If two
different patterns have some common components, v k also contains separators for
partially matched patterns, which can be obtained at lower threshold levels. This partial or
near match property is useful for pattern classification as it allows the retrieval of stored
patterns that are close to the test pattern in Hamming distance.
From those training patterns matched by the CMM engine, a test pattern is classified using
the k-NN rule. Distances are computed in the original input space to minimise the
information loss due to quantisation and noise in the above match process. As the number
of matches returned by the CMM is much smaller than the number of training data, the
distance computation and comparison are dramatically reduced compared with the simple
k-NN method. Therefore, the speed of the classifier benefits from fast training and
matching of the CMM, and the accuracy gains from the application of the k-NN rule for
reducing information loss and noise in the encoding and match processes.
2.2 ROBUST UNIFORM ENCODING
Figure 2 shows three stages of the encoding process. d-dimensional real numbers, xi' are
quantised as y; ; sparse and orthogonal binary vectors, Ci ' are generated and concatenated
to form a CMM input vector.
Yd
(~,
Figure 2: Quantisation, code generation and concatenation
CMM input codes should be distributed as uniformly as possible in order to avoid some
parts of the CMM being used heavily while others are rarely used. The code uniformity is
met at the quantisation stage. For a given set of N training samples in some dimension (or
axis), it is required to divide the axis into Nb small intervals, called bins, such that they
contain uniform numbers of data points. As the data often have a non-uniform distribution,
the sizes of these bins should be different. It is also quite common for real world problems
that many data points are identical. For instance, there are 11 %-99.9% identical data in
benchmarks used in this work. Our robust quantisation method described below is
designed to cope with the above problems and to achieve a maximal uniformity.
In our method data points are first sOfted in ascending order, N , identical points are then
identified, and the number of non-identical data points in each bin is estimated as
N p = (N - N, )/ Nb . B in boundaries or partitions are determined as follows. The right
boundary of a bin is initially set to the next N I' -th data point in the ordered data sequence;
the number of identical points on both sides of the boundary is identified; these are either
included in the current or next bin. If the number of non-identical data points in the last
bin is N , and N,~(Np +Nb)' Np may be increased by (N, -Np)/Nb and the above partition
process may be repeated to increase the uniformity. Boundaries of bins obtained become
parameters of the encoder in Figure 2. In general it is appropriate to choose Nh such that
each bin contains a number of samples, which is larger than k nearest neighbours for the
optimal classification.
P. Zhou, J. Austin and J. Kennedy
716
3 THE PRESENCE ARCHITECTURE
The pattern match and store engine of the CMM k-NN classifier has been implemented
using a novel hardware based CMM architecture. i.e. the PRESENCE.
3.1 ARCHITECTURE DESIGN
Important design decisions include the use of cheap memory, and not embedding both the
weight storage and the training and testing in hardware (VLSI). This arises because the
applications commonly use CMMs with over 100Mb of weight memory. which would be
difficult and expensive to implement in custom silicon. VME and PCI are chosen to host
on industry standard buses and to allow widespread application.
The PRESENCE architecture implements the control logic and accumulators, i.e. the core
of the CMM. As shown in Figure 3a binary input selects rows from the CMM that are
added, thresholded using L-max (Austin & Stonham 1987) or fixed global thresholding,
and then returned to the host for further processing. The PRESENCE architecture shown
in Figure 3b consists of a bus interface, a buffer memory which allows interleaving of
memory transfer and operation of the PRESENCE system, a SATCON and SA TSUM
combination that accumulates and thresholds the weights. The data bus connects to a pair
of memory spaces, each of which contains a control block, an input block and an output
block. Thus the PRESENCE card is a memory mapping device, that uses interrupts to
confirm the completion of each operation. For efficiency, two memory input/output areas
are provided to be acted on from the external bus and used by the card. The control
memory input block feeds to the control unit, which is a FPGA device. The input data are
fed to the weights and the memory area read is then passed to a block of accumulators. In
our current implementation the data width of each FPGA device is 32 bits, which allows
us to add a 32 bit row from the weights memory in one cycle per device
Input (sparse codes)
wei hts (-)
p
Data bus
??
Sumv
Separator output s
Figure 3: (a) correlation matrix memory. and (b) overall architecture of PRESENCE
Currently we have 16Mb of 25ns static memory implemented on the VME card, and 128
Mb of dynamic (60ns) memory on the PCI card. The accumulators are implemented along
with the thresholding logic on another FPGA device (SATSUM). To enable the SA TSUM
processors to operate faster, a 5 stage pipeline architecture was used, and the data
accumulation time is reduced from 175ns to 50ns. All PRESENCE operations are
supported by a C++ library that is used in all AURA applications. The design of the
SATCON allows many SATSUM devices to be used in parallel in a SIMD configuration.
The VME implementation uses 4 devices per board giving a 128 bit wide data path. In
addition the PCI version allows daisy chaining of cards allowing a 4 card set for a 512 bit
wide data path. The complete VME card assembly is shown in Figure 4. The SATCON
and SATSUM devices are mounted on a daughter board for simple upgrading and
alteration. The weights memory, buffer memory and VME interface are held on the
mother board.
717
RNNs Can Learn Symbol-Sensitive Counting
Figure 4: The VME based PRESENCE card (a) motherboard, and (b) daughterboard
3.2 PERFORMANCE
By an analysis of the state machines used in the SATCON device the time complexity of
the approach can be calculated. Equation 4 is used to calculate the processing time, T, in
seconds to recall the data with N index values, a separator size of S, R 32 bit SATSUM
devices, and the clock period of C.
T
= C[23+(s-l)/32R+I)(N +38+2R)]
(4)
A comparison with a Silicon Graphics 133MHz R4600SC Indy in Table
shows the
speed up of the matrix operation (Equation 2) for our VME implementation (128 bits
wide) using a fixed threshold. The values for processing rate are given in millions of
binary weight additions per-second (MW/s). The system cycle time needed to sum a row
of weights into the counters (i.e. time to accumulate one line) is SOns for the VME version
and lOOns for the PCI version. In the PCI form, we will use 4 closely coupled cards,
which result in a speed-up of 432. The build cost of the VME card was half the cost of the
baseline SGI Indy machine, when using 4Mb of 20ns static RAM. In the PCI version the
cost is greatly reduced through the use of dynamic RAM devices allowing a 128Mb
memory to be used for the same cost. allowing only a 2x slower system with 32x as much
memory per card (note that 4 cards used in Table I hold 512Mb of memory).
Table I : Relative speed-up of the PRESENCE architecture
Platform
Workstation
I Card VME implementation
Four card PCI system (estimate)
Processing_ Rate
11.8 MW/s
2557MW/s
17,114MW/s
I
Relative Speed
I
216
432
-
The training and recogmtlon speed of the system are approximately equal. This is
particularly useful in on-line applications, where the system must learn to solve the
problem incrementally as it is presented. In particular, the use of the system for high speed
reasoning allows the rules in the system to be altered without the long training times of
other systems. Furthermore our use of the system for a k-NN classifier also allows high
speed operation compared with a conventional implementation of the classifier, while still
allowing very fast training times.
4 RESULTS ON BENCHMARKS
Performance of the robust quantisation method and the CMM classifier have been
evaluated on four benchmarks consisting of large sets of real world problems from the
Statlog project (Michie & Spiegelhalter 1994), including a satellite image database, letter
image recognition database. shuttle data set and image segmentation data set. To visualise
the result of quantisation, Figure Sa shows the distribution of numbers of data points of
the 8th feature of the image segment data for equal-size bins. The distribution represents
P. Zhou, J. Austin and J. Kennedy
718
the inherent characteristics of the data. Figure 5b shows our robust quantisation (RQ) has
resulted in the uniform distribution desired.
400~~~--
________
--~
40~
350
~ 300
"
30
;""2SO
'"
25
20
~
~ 200
~
.i1
.i1
15
g
10
E
g
~
____
~~
;
150
E
100
~
_ _ _ _- -_ _
35
1 111111
o
5
10
15
20
values o f)ll;
25
30
o
35
o
10
15
20
values of x
25
30
3S
Figure 5: Distributions of the image segment data for (a) equal bins, (b) RQ bins
We compared the CMM classifier with the simple k-NN method, multi-layer perceptron
(MLP) and radial basis function (RBF) networks (Zhou and Austin 1997). In the
evaluation we used the CMM software libraries developed in the project AURA at the
University of York. Between 1 and 3 '1' bits are set in input vectors and separators.
Experiments were conducted to study influences of a CMM's size on classification rate (crate) on test data sets and speed-up measured against the k-NN method (as shown in
Figure 6). The speed-up of the CMM classifier includes the encoding, training and test
time. The effects of the number of bins N b on the performance were also studied.
~
0.89
i'!
g
0 .88
e 0.87
'"~ 0.86
0 0.85
0 .84
0.5
I
1.5
2
2.5
3
CMM Si7.e (MBytes)
15
4
I
1.5
2
2.5
3
CMM size (MBytes)
3. 5
4
Figure 6: Effects of the CMM size on (a) c-rate and (b) speed-up on the satellite image data
Choices of the CMM size and the number of bins may be application dependent, for
instance, in favour of the speed or accuracy. In the experiment it was required that the
speed-up is not 4 times less and c-rate is not 1% lower than that of the k-NN method.
Table 2 contains the speed-up of MLP and RBF networks and the CMM on the four
benchmarks. It is interesting to note that the k-NN method needed no training. The recall
of MLP and RBF networks was very faster but their training was much slower than that of
the CMM classifier. The recall speed-up of the CMM was 6-23 times, and the overall
speed-up (including training and recall time) was 4-15x. When using the PRESENCE, i.e.
the dedicated CMM hardware, the speed of the CMM was further increased over 3 times.
This is much less than the speed-up of 216 given in Table 1 because of recovering
separators and k-NN classification are performed in software.
Table 2: Speed-up of MLP, RBF and CMM relative to the simple k-NN method
Image segment
method
training
0.04
MLPN
RBFN
0.09
simplek-NN
CMM
18
test
18
9
I
9
Satellite image
training
0.2
0.07
-
15.8
Test
28.4
20.3
1
5.7
Letter
training
0.2
0.3
-
24.6
test
96.5
66.4
1
6.8
Shuttle
training
4.2
1.8
43
test
587.2
469.7
I
23
The classification rates by the four methods are given in Table 3, which shows the CMM
classifier performed only 0-1% less accurate than the k-NN method.
719
RNNs Can Learn Symbol-Sensitive Counting
Table 3: Classification rates of four methods on four benchmarks
MLPN
RBFN
simple k-NN
CMM
Image segment
0.950
0.939
0.956
0.948
Satellite image
0.914
0.914
0.906
0.901
Letter
0.923
0.941
0.954
0.945
Shuttle
0.998
0.997
0.999
0.999
5 CONCLUSIONS
A novel classifier is presented, which uses a binary CMM for storing and matching a large
amount of patterns efficiently, and the k-NN rule for classification . The RU encoder
converts numerical inputs into binary ones with the maximally achievable uniformity to
meet requirements of the CMM. Experimental results on the four benchmarks show that
the CMM classifier, compared with the simple k-NN method , gave slightly lower
classification accuracy (less than 1% lower) and over 4 times speed in software and 12
times speed in hardware. Therefore our method has resulted in a generic and fast
classifier.
This paper has also described a hardware implementation of a FPGA based chip set and a
processor card that will support the execution of binary CMM . It has shown the viability
of using a simple binary neural network to achieve high processing rates. The approach
allows both recognition and training to be achieved at speeds well above two orders of
magnitude faster than conventional workstations at a much lower cost than the
workstation. The system is scaleable to very large problems with very large weight arrays.
Current research is aimed at showing that the system is scaleable, evaluating methods for
the acceleration of the pre- and post processing tasks and considering greater integration
of the elements of the processor through VLSI. For more details of the AURA project and
the hardware described in this paper see http://www.cs.york.ac.uk/arch/nnJaura.html.
Acknowledgements
We acknowledge British Aerospace and the Engineering and Physical Sciences Research
Council (grant no. GRiK 41090 and GR/L 74651) for sponsoring the research. Our thanks
are given to R Pack, A Moulds, Z Ulanowski. R Jennison and K Lees for their support.
References
Willshaw, 0.1., Buneman, O.P. & Longuet-Higgins, H .C. (1969) Non-holographic
associative memory. Nature, Vol. 222, p960-962.
Austin, J. (1996) AURA, A distributed associative memory for high speed symbolic
reasoning. In: Ron Sun (ed), Connectionist Symbolic Integration. Kluwer.
Turner, M. & Austin, J. (1997) Matching performance of binary correlation matrix
memories. Neural Networks; 10:1637-1648.
Dasarathy, B.V. (1994) Minimal consistent set (MCS) identification for optimal nearest
neighbor decision system design. IEEE Trans. Systems Man Cybernet; 24:511-517.
Grother, P.l., Candela, G.T. & Blue, J.L. (1997) Fast implementations of nearest neighbor
classifiers. Pattern Recognition ; 30:459-465.
Austin, J., Stonham, T.J. (1987) An associative memory for use in image recognition and
occlusion analysis. Image and Vision Computing; 5:251-261.
Michie, D., Spiegelhalter, 0.1. & Taylor, c.c. (1994) Machine learning, neural and
statistical classification (Chapter 9). New York, Ellis Horwood.
Zhou, P. & Austin J. (1998) Learning criteria for training neural network classifiers .
Neural Computing and Applications Forum; 7:334-342.
PART VI
SPEECH, HANDWRITING AND SIGNAL
PROCESSING
| 1495 |@word version:4 achievable:1 shot:1 configuration:1 contains:4 current:4 written:1 must:1 john:1 numerical:2 partition:2 cheap:1 designed:1 hts:1 v:1 half:1 device:11 core:1 ron:1 along:1 become:1 consists:2 multi:2 considering:1 mpj:1 project:4 provided:1 matched:2 developed:3 willshaw:2 classifier:21 uk:5 control:4 unit:1 grant:1 engineering:1 encoding:7 accumulates:1 dasarathy:2 meet:3 path:2 yd:1 approximately:1 rnns:3 studied:1 range:2 unique:1 accumulator:3 testing:1 practice:1 block:5 implement:2 area:2 matching:5 pre:3 radial:1 symbolic:3 close:1 put:1 storage:3 applying:1 nb:4 influence:1 accumulation:1 conventional:4 www:1 quick:1 smail:1 rule:8 higgins:1 array:1 embedding:1 commercial:1 heavily:1 exact:1 us:4 element:1 expensive:1 particularly:1 recognition:4 michie:2 database:2 module:1 calculate:1 cycle:2 sun:1 counter:1 removed:2 rq:2 complexity:1 dynamic:2 uniformity:4 segment:4 mbytes:2 efficiency:2 basis:1 chip:1 represented:1 chapter:1 train:1 fast:8 effective:1 describe:1 pci:7 quite:1 larger:2 solve:1 encoder:4 ability:1 knn:1 unseen:1 associative:3 sequence:1 tpi:1 maximal:1 mb:6 achieve:2 sgi:1 requirement:4 satellite:4 produce:1 generating:1 perfect:1 ac:4 completion:1 minimises:1 measured:1 nearest:3 sa:4 indy:2 implemented:4 c:4 recovering:2 met:2 closely:1 enable:1 bin:13 statlog:1 hold:1 normal:1 great:1 mapping:1 applicable:1 label:1 currently:1 sensitive:3 council:1 successfully:1 aim:1 zhou:6 avoid:1 shuttle:3 interrupt:1 vk:2 greatly:1 baseline:1 dependent:1 nn:28 cmm:60 initially:1 spurious:1 vlsi:2 selects:1 i1:2 overall:2 classification:16 flexible:1 aura:5 html:1 platform:1 integration:2 equal:3 simd:1 identical:7 represents:1 np:4 others:1 loon:1 inherent:1 few:1 connectionist:1 neighbour:1 resulted:2 individual:1 connects:1 consisting:1 occlusion:1 scaleable:3 mlp:4 highly:1 custom:1 evaluation:1 held:1 accurate:1 partial:3 orthogonal:3 divide:1 taylor:1 desired:1 minimal:1 uncertain:1 instance:2 increased:2 industry:1 elli:1 mhz:1 cost:6 uniform:4 fpga:4 holographic:1 conducted:1 gr:1 too:1 graphic:1 stored:2 thanks:1 fundamental:1 lee:1 choose:1 external:1 alteration:1 includes:2 vi:1 performed:2 candela:2 parallel:1 daisy:1 accuracy:4 characteristic:1 efficiently:1 identification:1 basically:1 vme:10 mc:1 kennedy:4 processor:3 classified:1 ping:1 ed:1 against:1 workstation:5 hamming:1 gain:1 static:2 handwriting:1 logical:1 recall:8 dimensionality:1 segmentation:1 feed:1 wei:1 maximally:1 evaluated:1 furthermore:1 stage:3 arch:1 correlation:6 clock:1 incrementally:1 widespread:1 effect:3 contain:1 hence:1 chemical:1 read:1 deal:1 ll:1 width:1 sponsoring:1 chaining:1 criterion:1 complete:1 performs:1 dedicated:2 interface:2 reasoning:4 image:13 novel:4 common:3 physical:1 nh:1 million:1 association:1 visualise:1 kluwer:1 accumulate:1 silicon:2 mother:1 quantisation:9 add:1 store:3 buffer:2 binary:19 greater:1 period:1 signal:1 match:10 faster:3 long:2 retrieval:1 host:2 post:3 buneman:2 vision:1 achieved:1 whereas:1 background:1 addition:2 crate:1 interval:1 appropriately:1 operate:1 near:1 presence:13 counting:3 mw:4 viability:1 gave:2 architecture:14 identified:3 idea:1 minimise:1 bottleneck:1 favour:1 motivated:1 passed:1 effort:1 returned:2 speech:1 york:8 dramatically:1 useful:2 detailed:3 aimed:1 amount:2 mid:1 hardware:10 reduced:3 http:1 diagnostic:1 estimated:1 per:4 blue:1 vol:1 group:1 key:1 four:7 threshold:4 thresholded:1 ram:2 convert:2 sum:1 grother:2 quantised:1 letter:3 decision:2 investigates:1 oring:1 bit:10 layer:2 followed:1 software:4 speed:28 performing:1 acted:1 department:1 upgrading:1 combination:2 smaller:1 slightly:1 son:1 pipeline:1 equation:3 bus:5 discus:1 mechanism:1 needed:3 horwood:1 ascending:1 fed:1 operation:7 apply:1 generic:2 appropriate:3 mlpn:2 slower:2 original:1 include:1 assembly:1 giving:1 concatenated:2 build:1 forum:1 added:1 primary:1 exhibit:1 distance:3 card:15 capacity:1 concatenation:1 ru:1 code:11 index:1 difficult:1 daughter:1 implementation:9 design:4 maximises:1 allowing:4 benchmark:8 acknowledge:1 jim:1 pair:1 required:2 aerospace:1 engine:4 trans:1 below:3 pattern:24 usually:1 preselected:1 max:1 memory:26 including:2 advanced:2 turner:3 improve:1 altered:1 spiegelhalter:2 library:2 axis:2 coupled:1 speeding:1 acknowledgement:1 relative:3 loss:2 interesting:2 generation:1 proportional:1 mounted:1 consistent:1 thresholding:6 storing:1 austin:14 row:3 supported:1 last:1 side:2 allow:2 understand:1 perceptron:1 wide:4 neighbor:2 sparse:3 benefit:1 distributed:2 overcome:1 dimension:1 boundary:4 world:2 calculated:1 evaluating:1 commonly:1 cope:1 logic:2 confirm:1 global:1 xi:1 search:1 iterative:1 sk:1 table:8 nature:1 learn:4 transfer:1 robust:6 pack:1 longuet:1 improving:1 complex:1 separator:7 pk:2 noise:3 repeated:1 board:3 slow:1 n:5 sub:2 learns:1 interleaving:1 british:1 showing:1 symbol:3 rbfn:2 ci:1 magnitude:1 execution:3 likely:1 ordered:1 partially:2 scalar:1 acceleration:1 rbf:4 man:1 included:1 determined:1 reducing:2 uniformly:1 called:1 experimental:2 rarely:1 select:1 support:2 arises:1 tested:1 |
542 | 1,496 | Coordinate Transformation Learning of
Hand Position Feedback Controller by
U sing Change of Position Error Norm
Eimei Oyama*
Mechanical Eng. Lab.
Namiki 1-2, Tsukuba Science City
Ibaraki 305-8564 Japan
Susumu Tachi
The University of Tokyo
Hongo 7-3-1, Bunkyo-ku
Tokyo 113-0033 Japan
Abstract
In order to grasp an object, we need to solve the inverse kinematics problem, i.e., the coordinate transformation from the visual
coordinates to the joint angle vector coordinates of the arm. Although several models of coordinate transformation learning have
been proposed, they suffer from a number of drawbacks. In human
motion control, the learning of the hand position error feedback
controller in the inverse kinematics solver is important. This paper
proposes a novel model of the coordinate transformation learning
of the human visual feedback controller that uses the change of
the joint angle vector and the corresponding change of the square
of the hand position error norm. The feasibility of the proposed
model is illustrated using numerical simulations.
1
INTRODUCTION
The task of calculating every joint angle that would result in a specific hand position
is called the inverse kinematics problem. An important topic in neuroscience is the
study of the learning mechanisms involved in the human inverse kinematics solver.
We questioned five pediatricians about the motor function of infants suffering from
serious upper limb disabilities. The doctors stated that the infants still were able
to touch and stroke an object without hindrance. In one case, an infant without
a thumb had a major kinematically influential surgical operation, transplanting an
index finger as a thumb. After the operation, the child was able to learn how to
use the index finger like a thumb [1]. In order to explain the human motor learning
? Phone:+81-298-58-7298, Fax:+81-298-58-7201, e-mail:eimei@mel.go.jp
Coordinate Transformation Learning ofFeedback Controller
1039
capability, we believe that the coordinate transformation learning of the feedback
controller is a necessary component.
Although a number of learning models of the inverse kinematics solver have been
proposed, a definitive learning model has not yet been obtained. This is from the
point of view of the structural complexity of the learning model and the biological plausibility of employed hypothesis. The Direct Inverse Modeling employed
by many researchers [2] requires the complex switching of the input signal of the
inverse model. When the hand position control is performed, the input of the inverse model is the desired hand position, velocity, or acceleration. When the inverse
model learning is performed, the input is the observed hand position, velocity, or
acceleration. Although the desired signal and the observed signal could coincide,
the characteristics of the two signals are very different. Currently, no research has
succeesfully modeled the switching system. Furthermore, that learning model is not
"goal-directed"; i.e., there is no direct way to find an action that corresponds to a
particular desired result. The Forward and Inverse Modeling proposed by Jordan
[3] requires the back-propagation signal, a technique does not have a biological basis. That model also requires the complex switching of the desired output signal
for the forward model. When the forward model learning is performed, the desired
output is the observed hand position. When the inverse kinematics solver learning is performed, the desired output is the desired hand position. The Feedback
Error Learning proposed by Kawato [4] requires a pre-existing accurate feedback
controller.
It is necessary to obtain a learning model that possesses a number of characteristics:
(1) it can explain the human learning function; (2) it has a simple structure; and
(3) it is biologically plausible. This paper presents a learning model of coordinate
transformation function of the hand position feedback controller. This model uses
the joint angle vector change and the corresponding change of square of the hand
position error norm.
2 BACKGROUND
2.1 Discrete Time First Order Model of Hand Position Controller
Let 8 E Rm be the joint angle vector and x ERn be the hand position/orientation
vector given by the vision system. The relationship between x and 8 is expressed
as x = /(8) where / is a C 1 class function. The Jacobian of the hand position
vector is expressed as J(8) = 8/(8)/88. Let Xd be the desired hand position and
e = Xd - X = Xd - /(8) be the hand position error vector. In this paper, an
inverse kinematics problem is assumed to be a least squares minimization problem
that calculates 8 in order to minimize the square of the hand position error norm
S(xd,8)
= le1 2 /2 = IXd -
/(8)1 2 /2.
First, the feed-forward controller in the human inverse kinematics solver is disregarded and the following first order control system, consisting of a learning feedback
controller, is considered:
8(k
Position
+
Xti
Desired
Hand
Position
Error
e(k)
+ 1)
Feedback
?
=
8(k)
+ A8(k)
Di~rbance
d(k)
NOIse
+
tPp/.8,
e)
Joint Angle
Vector
~k~
(+-~
(1)
(f?-'f(8)
H
uman
Arm
Hand
Position
X(!i...
8(k)
Figure 1: Configuration of 1-st Order Model of Hand Position Controller
E. Oyama and S. Tachi
1040
a8(k) = ~fb(8(k), e(k)) + d(k)
(2)
e(k) = Xd - f(8(k))
(3)
where d(k) is assumed to be a disturbance noise from all components except the
hand position control system. Figure 1 shows the configuration of the control system. In this figure, Z-l is the operator that indicates a delay in the discrete time
signal by a sampling interval of tl.t. Although the human hand position control system includes higher order complex dynamics terms which are ignored in Equation
(2), McRuer's experimental model of human compensation control suggests that the
term that converts the hand position error to the hand velocity is a major term in
the human control system [5]. We consider Equation (2) to be a good approximate
model for the analysis of human coordinates transformation learning.
The learner ~ fb (8, e) E R m, which provides the hand position error feedback, is
modeled using the artificial neural network. In this paper, the hand position error
feedback controller learning by observing output x(k) is considered without any
prior knowledge of the function f (8).
2.2
Learning Model of the Neural Network
Let ~fb(8, e) be the desired output of the learner ~fb(8, e). ~fb(8, e) functions as
a teacher for ~fb(8,e). Let ~jb(8 , e) be the updated output of ~fb(8,e) by the
learning. Let E[t(8, e)18, e] be the expected value of a scalar, a vector, or a matrix
function t(8,e) when the input vector (8 , e) is given. We assume that ~fb(8 , e) is
an ideal learner which is capable of realizing the mean of the desired output signal,
completely. ~+ fb(8, e) can be expressed as follows:
~jb(8, e)
~
E[~fb(8 , e)18, e]
a~fb(8 , e)
When the expected value of
= ~fb(8 , e) + E[a~fb(8, e)18, e]
= ~fb(8 , e) -
a~fb(8,
~fb(8, e)
(4)
(5)
e) is expressed as:
E[a~fb(8,e)18,e] ~ Gfbe - Rfb~fb(8 , e) ,
(6)
Rfb E Rm xm is a positive definite matrix, and the inequality
I8~jb(8, e) I = I8(G fb e -
(Rfb - I)~fb(8, e?
e)
8~ fb(8 , e)
is satisfied, the final learning result can be expressed as:
~fb(8 , e) ~ Rjb1Gfbe
I< 1
8~fb(8,
by the iteration of the update of
3
3.1
~fb(8 ,
(7)
(8)
e) expressed in Equation (4).
USE OF CHANGE OF POSITION ERROR NORM
A Novel Learning Model of Feedback Controller
The change of the square of the hand position error norm tl.S = S(Xd , 8 + a8) S(Xd , 8) reflects whether or not the change of the joint angle vector A8 is in proper
direction. The propose novel learning model can be expressed as follows:
~fb(8, e) = -atl.Sa8
(9)
where a is a small positive real number. We now consider a large number of trials
of Equation (2) with a large variety of initial status 8(0) with learnings conducted
at the point of the input space of the feedback controller (8, e) = (8(k -1), e(k -1?
at time k. tl.S and a8 can be calculated as follows.
tl.S
a8
S(k) - S(k - 1) =
=
a8(k - 1)
~(le(kW -Ie(k -
1W)
(10)
(11)
1041
Coordinate Transformation Learning ofFeedback Controller
.---------, Change of Square of
Hand Position Error Norm
e(k)
Hand
Position
Change of
Joint Angle
Vector
Input for
Learning
Error
e(k-l) ---
e(k)
d8(k)
Error Signal for k ' / '---,..--T----'
Feedback
Input for
Controller
Learning
Input for
Control
(J(k-l)
d(k)
Dist""'-z
NoiJe
8(k)
Figure 2: Configuration of Learning Model of Feedback Controller
Figure 2 shows the conceptual diagram of the proposed learning model.
Let p(qI8, e) be the probability density function of a vector q at at the point (8, e)
in the input space of ~fb(8, e). In order to simplify the analysis of the proposed
learning model, d(k) is assumed to satisfy the following equation:
p(dI8, e) = p( -dI8, e)
(12)
When d8 is small enough, the result of the learning using Equation (9) can be
expressed as:
(13)
~fb(8, e) ~ a(~R9JT (8)J(8) + 1)-1 R9JT (8)e
R9
=
E[d8d8T I8, e]
(14)
where JT (8)e is a vector in the steepest descent direction of S(Xd, 8). When d(k) is
a non-zero vector, R9 is a positive definite symmetric matrix and (~R9JT J + 1)-1
is a positive definite matrix. When a is appropriate, ~ fb(8, e) as expressed in
Equation (13) can provide appropriate output error feedback control. The derivation of the above result will be illustrated in Section 3.2. A partially modified
steepest descent direction can be obtained without using the forward model or the
back-propagation signal, as Jordan's forward modeling [3].
Let Rd be the covariance matrix of the disturbance noise d(k). When a is infinitesimal, R9 ~ Rd is established and an approximate solution ~fb(8,e) ~ aRdJ T (8)e
is obtained.
3.2 Derivation of Learning Result
The change of the square of the hand position error norm llS(Xd, 8) by d8 can be
determined as:
llS(xd, 8) =
8S~;, 8) d8 + ~d8T H(Xd, 8)d8 + O(d83 )
= -eT (J(8) + ~ 8~~8)
i&l d8)d8
(15)
+ ~d8T J T (8)J(8)d8 + O(d83 )
where i&l is a 2-operand operator that indicates the Croneker's product. H(Xd,8) E
is the Hessian of S(Xd, 8). O(d8 3 ) is the sum of third and higher order
terms of d8 in each equation. When d8 is small enough, the following approximate
equations are obtained:
Rmxm
18J(8)
1
dx ~ J(8)d8 ~ J(8 + 2"d8)d8 ~ (J(8) + 2 88 i&l d8)d8
Therefore, llS can be approximated as follows:
1
llS ~ _eT J(8)d8 + 21dXI2
(16)
(17)
E. Oyama and S. Tachi
1042
Since e T J AOAO = AOAO T JT e and IAxI2 AO = AOAOT JT J AO are determined, tl.S AO can be approximated as:
(18)
Considering AO njb defined as AO njb = AO - .jb(O,e), the expected value of the
product of AO and tl.S at the point (O,e) in the input space of .jb(O,e) can be
approximated as follows:
E[tl.SAOIO, e]
TIT
+ 2ReJ
~
-ReJ e
+
1
T T
2E[AOAO J J AOnjblO, e]
J.jb(O,e)
(19)
When the arm is controlled according to Equation (2), AO njb is the disturbance
noise d(k). Since d(k) satisfies Equation (12), the following equation is established.
E[AOAO T JT JAOnjbIO,e] = 0
(20)
Therefore, the expected value of A.jb(O, e) can be expressed as;
TaT
E[A.jb(O, e)IO, e] ~ aReJ e - (2ReJ J
+ I).jb(O, e)
(21)
When a is small enough, the condition described in Equation (7) is established.
The learning result expressed as Equation (13) is obtained as described in Section
2.2.
It should be noted that the learning algorithm expressed in Equation (9) is applicable not only to S(Xd,O), but also to general penalty functions of hand po-
sition error norm lei. The proposed learning model synthesizes a direction that
decreases S(Xd,O) by summing after weighting AO based on the increase or decrease of S(Xd, 0).
The feedback controller defined in Equation (13) requires a number of iterations
to find a correct inverse kinematics solution, as the coordinates transformation
function of the controller is incomplete. However, by using Kawato's feedback
error learning [4], the second feedback controller; the feed-forward controller; or the
inverse kinematics model that has a complete coordinate transformation function
can be obtained as shown in Section 4.
4
TRACKING CONTROL SYSTEM LEARNING
In this section, we will consider the case where Xd changes as xd(k)(k
1,2, ... ). The hybrid controller that includes the learning feed-forward controller
.ff(O(k), AXd(k)) E Rm that transforms the change of the desired hand position
AXd(k) = xd(k + 1) - xd(k) to the joint angle vector space is considered:
AO(k) = .ff(O(k), AXd(k)) + .,b(O(k),e(k))
e(k)
xd(k) - x(k)
+ d(k)
=
(22)
(23)
The configuration of the hybrid controller is illustrated in Figure 3.
By using the modified change of the square of the error norm expressed as:
1
2
2
tl.S = 2(lxd(k - 1) - x(k)1 - le(k - 1)1 )
(24)
and AO(k) as defined in Equation (22), the feedback controller learning rule defined
in Equation (9) is useful for the tracking control system. A sample holder for
memorizing xd(k -1) is necessary for the calculation of tl.S. When the distribution
1043
Coordinate Transformation Learning ofFeedback Controller
Error Signal for ~:,..-:--::----:-"'
Fccdforword
Controller
Fccdforwonl
CoDtrolJer
iixd(k)
~
~
4},(8,Lix" (k?)- J*(8)Lh" (k)
i----r---,
:I
J(8)J*(8)=1
./i(8'. aj(8)
I
arr-
HWIIIJIArm
Position
I!m>r
+
~(k)
Desired
Hand
Position
e(k)
-
~
Feedback
?
8(k)
Hand
Position
x(k)
X=f(O)
8(k)
Figure 3: Configuration of Hybrid Controller
of 4Xd(k) satisfies Equation (20), Equation (13) still holds. When 4Xd(k) has no
correlation with d(k) and 4Xd(k) satisfies p(4XdI8, e) p( -4XdI8, e), Equation
(20) is approximately established after the feed-forward controller learning.
=
Using 48(k) defined in Equation (2) and e(k) defined in Equation (23), tl.S defined in Equation (10) can be useful for the calculation of ~fb(8, e). Although the
learning calculation becomes simpler, the learning speed becomes much lower.
Let~' ff(8(k),
4Xd(k)) be the desired output of ~,,(8(k), 4Xd(k)). According to
Kawato's feedback error learning [4], we use ~',,(8(k), 4Xd(k)) expressed as:
~',,(8(k),
4Xd(k)) = (1 - >..)~,,(8(k), 4Xd(k))
+ ~fb(8(k + 1), e(k + 1))
(25)
where >.. is a small, positive, real number for stabilizing the learning process and
ensuring that equation ~,,(8,O) ~ 0 holds . If >.. is small enough, the learning
feed-forward controller will fulfill the equation:
J~,,(8,
5
4Xd) ~ 4Xd
(26)
NUMERICAL SIMULATION
Numerical simulation experiments were performed in order to evaluate the performance of the proposed model. The inverse kinematics of a 3 DOF arm moving on
a 2 DOF plane were considered. The relationship between the joint angle vector
8 = (8 1 '(h, ( 3 ) T and the hand position vector x = (x, y) T was defined as:
x
= Xo + Ll cos(8t} + L2 cos(81 + (2 ) + L3 cos(81 + 82 + ( 3 )
y = Yo
+ Ll sin(81) + L2 sin(81 + (2 ) + L3 sin(81 + 82 + ( 3 )
(27)
(28)
The range for 81 was (-30 0 ,1200 ); the range for 82 was (0 0 ,120 0 ); and the range
for 83 was (_75 0 ,75 0 ). Ll was 0.30 m, L2 was 0.25 m and L3 was 0.15 m.
Random straight lines were generated as desired trajectories for the hand. The
tracking control trials expressed as Equation (22) with the learning of the feedback
controller and the feed-forward controller were performed. The standard deviation
of each component of d was 0.01. Learnings based on Equations (9), (22) , (24),
and (25) were conducted 20 times in one tracking trial. 1,000 tracking trials were
conducted to estimate the RMS(Root Mean Square) of e(k) .
In order to accelerate the learning, a in Equation (9) was modified as a =
0.5/(Itl.xI 2 + 0.11tl.(12). >.. in Equation (25) was set to O.OOL
Two neural networks with 4 layers were used for the simulation. The first layer had
5 neurons and the forth layer had 3 neurons. The other layers had 15 neurons each.
The first layer and the forth layer consisted of linear neurons. The initial values of
weights of the neural networks were generated by using uniform random numbers.
The back-propagation method without optimized learning coefficients was utilized
for the learning.
E. Oyama and S. Tachi
1044
El00~------------------~
.......
y
0.5
...
g
w 10.2 +-r--___--.........__~
___.............-r.........f
10?101102103104105106107
o
0.5 x
of Trials
Figure 4: Learning Process of Controller Figure 5: One Example of Tracking Control
CE:
Number
Figure 4 shows the progress of the proposed learning model. It can be seen that the
RMS error decreases and the precision of the solver becomes higher as the number
of trials increases. The RMS error became 9.31 x 1O- 3 m after 2 x 107 learning trials .
Figure 5 illustrates the hand position control by the inverse kinematics solver after
2 x 107 learning trials. The number near the end point of the arm indicates the
value of k. The center of the small circle in Figure 5 indicates the desired hand
position. The center of the large circle indicates the final desired hand position.
Through learning, a precise inverse kinematics solver can be obtained. However,
for RMS error to fall below 0.02, trials must be repeated more than 106 times. In
such cases, more efficient learner or a learning rule is necessary.
6
CONCLUSION
A learning model of coordinate transformation of the hand position feedback controller was proposed in this paper. Although the proposed learning model may
take a long time to learn, it is capable of learning a correct inverse kinematics
solver without using a forward model, a back-propagation signal, or a pre-existing
feedback controller.
We believe that the slow learning speed can be improved by using neural networks
that have a structure suitable for the coordinate transformation. A major limitation
of the proposed model is the structure of the learning rule, since the learning rule
requires the calculation of the product of the change of the error penalty function
and the change of the joint angle vector. However, the existence of such structure in
the nervous system is unknown. An advanced learning model which can be directly
compared with the physiological and psychological experimental results is necessary.
References
[1] T. Ogino and S. Ishii, "Long-term Results after Pollicization for Congenital
Hand Deformities," Hand Surgery, 2, 2,pp.79-85,1997
[2] F. H. Guenther and D. M. Barreca," Neural models for flexible control of redundant systems," in P. Morasso and V. Sanguineti (Eds.), Self-organization, Computational Maps, and Motor Control. Amsterdam: Elsevier, pp.383-421 ,1997
[3J M.1. Jordan, "Supervised Learning and Systems with Excess Degrees of Freedom," COINS Technical Report,88-27,pp.1-41 ,1988
[4J M. Kawato, K. Furukawa and R. Suzuki, "A Hierarchical Neural-network
Model for Control and Learning of Voluntary Movement," Biological Cybernetics, 57, pp.169-185, 1987
[5] D.T. McRuer and H. R. Jex, "A Review of Quasi-Linear Pilot Models," IEEE
Trans. on Human Factors in Electronics, HFE-8, 3, pp.38-51, 1963
| 1496 |@word trial:9 norm:10 tat:1 simulation:4 eng:1 covariance:1 electronics:1 configuration:5 initial:2 existing:2 yet:1 dx:1 must:1 numerical:3 motor:3 update:1 infant:3 nervous:1 plane:1 steepest:2 realizing:1 provides:1 simpler:1 five:1 direct:2 guenther:1 expected:4 dist:1 xti:1 solver:9 considering:1 becomes:3 transformation:14 every:1 xd:32 rm:3 control:18 positive:5 io:1 switching:3 sanguineti:1 approximately:1 suggests:1 co:3 range:3 directed:1 definite:3 pre:2 operator:2 map:1 center:2 go:1 stabilizing:1 ixd:1 rule:4 atl:1 coordinate:16 updated:1 us:2 hypothesis:1 velocity:3 approximated:3 utilized:1 observed:3 lix:1 decrease:3 movement:1 complexity:1 dynamic:1 tit:1 learner:4 basis:1 completely:1 po:1 joint:11 accelerate:1 finger:2 derivation:2 artificial:1 dof:2 solve:1 plausible:1 final:2 propose:1 product:3 arr:1 forth:2 object:2 progress:1 direction:4 drawback:1 tokyo:2 correct:2 human:11 ao:11 biological:3 hold:2 rfb:3 considered:4 major:3 applicable:1 currently:1 city:1 reflects:1 minimization:1 modified:3 fulfill:1 yo:1 deformity:1 indicates:5 ishii:1 elsevier:1 quasi:1 orientation:1 flexible:1 proposes:1 sampling:1 kw:1 jb:9 report:1 simplify:1 serious:1 lls:4 njb:3 consisting:1 freedom:1 organization:1 grasp:1 accurate:1 capable:2 necessary:5 lh:1 incomplete:1 desired:17 circle:2 psychological:1 modeling:3 deviation:1 uniform:1 delay:1 conducted:3 teacher:1 st:1 density:1 ie:1 satisfied:1 kinematically:1 d8:17 japan:2 includes:2 coefficient:1 satisfy:1 performed:6 view:1 root:1 lab:1 observing:1 doctor:1 capability:1 minimize:1 square:9 holder:1 became:1 characteristic:2 surgical:1 thumb:3 trajectory:1 researcher:1 cybernetics:1 straight:1 stroke:1 explain:2 ed:1 infinitesimal:1 pp:5 involved:1 di:1 pilot:1 knowledge:1 oyama:4 back:4 feed:6 higher:3 supervised:1 improved:1 furthermore:1 correlation:1 hand:42 hindrance:1 touch:1 propagation:4 aj:1 lei:1 believe:2 consisted:1 symmetric:1 illustrated:3 ll:3 sin:3 self:1 noted:1 mel:1 complete:1 motion:1 novel:3 kawato:4 operand:1 jp:1 rd:2 axd:3 had:4 l3:3 moving:1 phone:1 inequality:1 furukawa:1 seen:1 employed:2 redundant:1 signal:12 technical:1 plausibility:1 calculation:4 long:2 feasibility:1 calculates:1 controlled:1 ensuring:1 controller:35 vision:1 sition:1 iteration:2 background:1 interval:1 diagram:1 posse:1 jordan:3 structural:1 near:1 ideal:1 enough:4 variety:1 ibaraki:1 tpp:1 whether:1 rms:4 penalty:2 ool:1 suffer:1 bunkyo:1 questioned:1 hessian:1 action:1 ignored:1 useful:2 transforms:1 neuroscience:1 discrete:2 susumu:1 ce:1 convert:1 sum:1 inverse:19 angle:11 layer:6 tsukuba:1 speed:2 jex:1 ern:1 influential:1 according:2 biologically:1 memorizing:1 xo:1 equation:30 kinematics:14 mechanism:1 end:1 operation:2 limb:1 hierarchical:1 appropriate:2 coin:1 rmxm:1 rej:3 existence:1 calculating:1 surgery:1 disability:1 topic:1 mail:1 index:2 modeled:2 relationship:2 stated:1 proper:1 unknown:1 upper:1 neuron:4 sing:1 compensation:1 descent:2 voluntary:1 precise:1 lxd:1 mechanical:1 optimized:1 established:4 trans:1 able:2 below:1 xm:1 suitable:1 hybrid:3 disturbance:3 advanced:1 arm:5 morasso:1 fax:1 prior:1 review:1 l2:3 limitation:1 degree:1 i8:3 fall:1 feedback:25 calculated:1 fb:31 forward:12 suzuki:1 coincide:1 excess:1 approximate:3 status:1 hongo:1 conceptual:1 summing:1 assumed:3 xi:1 ku:1 learn:2 synthesizes:1 complex:3 definitive:1 noise:4 suffering:1 child:1 repeated:1 tl:11 ff:3 slow:1 precision:1 position:41 congenital:1 jacobian:1 third:1 weighting:1 specific:1 jt:4 uman:1 physiological:1 illustrates:1 disregarded:1 visual:2 expressed:15 amsterdam:1 tracking:6 partially:1 scalar:1 corresponds:1 a8:7 satisfies:3 itl:1 goal:1 acceleration:2 change:16 determined:2 except:1 called:1 r9:3 experimental:2 evaluate:1 |
543 | 1,497 | Finite-dimensional approximation of
Gaussian processes
Giancarlo Ferrari Trecate
Dipartimento di Informatica e Sistemistica, Universita di Pavia,
Via Ferrata 1, 27100 Pavia, Italy
ferrari@conpro.unipv.it
Christopher K. I. Williams
Department of Artificial Intelligence, University of Edinburgh,
5 Forrest Hill, Edinburgh EH1 2QL,
ckiw@dai.ed.ac.uk.
Manfred Opper
Neural Computing Research Group
Division of Electronic Engineering and Computer Science
Aston University, Birmingham, B4 7ET, UK
m.opper@aston.ac.uk
Abstract
Gaussian process (GP) prediction suffers from O(n3) scaling with the
data set size n. By using a finite-dimensional basis to approximate the
GP predictor, the computational complexity can be reduced. We derive optimal finite-dimensional predictors under a number of assumptions, and show the superiority of these predictors over the Projected
Bayes Regression method (which is asymptotically optimal). We also
show how to calculate the minimal model size for a given n. The
calculations are backed up by numerical experiments.
1
Introduction
Over the last decade there has been a growing interest in the Bayesian approach to
regression problems, using both neural networks and Gaussian process (GP) prediction,
that is regression performed in function spaces when using a Gaussian random process
as a prior.
The computational complexity of the GP predictor scales as O(n 3), where n is the size
Finite-Dimensional Approximation o/Gaussian Processes
219
of the datasetl . This suggests using a finite-dimensional approximating function space,
which we will assume has dimension m < n. The use of the finite-dimensional model is
motivated by the need for regression algorithms computationally cheaper than the G P
one. Moreover, GP regression may be used for the identification of dynamical systems
(De Nicolao and Ferrari Trecate, 1998), the next step being a model-based controller
design. In many cases it is easier to accomplish this second task if the model is low
dimensional.
Use of a finite-dimensional model leads naturally to the question as to which basis is
optimal. Zhu et al. (1997) show that, in the asymptotic regime, one should use the
first m eigenfunctions of the covariance function describing the Gaussian process. We
call this method Projected Bayes Regression (PBR) .
The main results of the paper are:
1. Although PBR is asymptotically optimal, for finite data we derive a predictor
hO(x) with computational complexity O(n 2m) which outperforms PBR, and
obtain an upper bound on the generalization error of hO(x) .
2. In practice we need to know how large to make m . We show that this depends
on n and provide a means of calculating the minimal m. We also provide
empirical results to back up the theoretical calculation.
2
Problem statement
Consider the problem of estimating an unknown function f(x) : JRd -T JR, from the
noisy observations
ti = f(Xi) + Ei,
i = 1, ... , n
where Ei are i.i.d. zero-mean Gaussian random variables with variance a 2 and the
samples Xi are drawn independently at random from a distribution p(x) . The prior
probability measure over the function f (.) is assumed to be Gaussian with zero mean
and autocovariance function C (6,6). Moreover we suppose that f (.), Xi, Ei, are
mutually independent. Given the data set 'Dn = {x , f}, where x = [Xl' . .. ' Xn] and
f = [tl, ... , t n]', it is well known that the posterior probability PUI'Dn) is Gaussian
and the GP prediction can be computed via explicit formula (e.g. Whittle, 1963)
j(x)
= E[JI'Dn](x) =
C(x , Xn)] H - lf,
[C(x, xd
{H} ij ~C(Xi ' Xj)
+ a20ij
where H is a n x n matrix and Oij is the Kronecker delta.
In this work we are interested in approximating j in a suitable m-dimensional space
that we are going to define. Consider the Mercer-Hilbert expansion of C(6, 6)
r C(6,6)'Pi(6)p(6)d6 =
iRd
Ai'Pi(6),
r 'Pi(~)'Pj(Op(~)d~ = Oij
iRd
(1)
+00
C(6,6)
=
L Ai'Pi(6)'Pi(6),
i=l
where the eigenvalues Ai are ordered in a decreasing way.
Then, in (Zhu et al., 1997) is shown that, at least asymptotically, the optimal model
belongs to M = Span {'Pi, i = 1, ... , m}. This motivates the choice of this space even
when dealing with a finite amount of data.
Now we introduce the finite-dimensional approximator which we call Projected Bayes
Regression.
lO(n3) arises from the inversion of a n x n matrix.
G. Ferrari- Trecate, C. K. I. Williams and M. Opper
220
Definition 1 The PBR approximator is b(x)
A = (A -1 + ,8iJ>' iJ? , (A)ij=Ai6ij and
k(x)=
[~I{X)
l'
= k' (x)w,
where w=,8A- 1iJ>'f, ,8=1/(12,
iJ>=
rpm (x)
The name PBR comes from the fact that b(x) is the GP predictor when using the
mis-specified prior
(2)
i=1
whose auto covariance function is the projection of C(6,6) on M. From the computational point of view, is interesting to note that the calculation of PBR scales with
the data as O(m 2 n), assuming that n ? m (this is the cost of computing the matrix
product A-I iJ>').
Throughout the paper the following measures of performance will be extensively used.
Definition 2 Let s{x) be a predictor that uses only information from Dn. Then its
x-error and generalization error are respectively defined as
Es(n,x)=Et.,x.,l [(t* - s(x*))2] , EHn)=Ex [Es{n,x)].
An estimator SO(x) belonging to a class 11. is said x-optimal or simply optimal if, respectively, Eso(n,x) ~ Es{n,x) or E;o(n) ~ E~(n), for all the s(x) E 11. and the data
sets x.
Note that x-optimality means optimality for each fixed vector x of data points. Obviously, if SO(x) is x-optimal it is also simply optimal. These definitions are motivated by
the fact that for Gaussian process priors over functions and a predictor s that depends
linearly on 1, the computation of Es(n, x) can be carried out with finite-dimensional
matrix calculations (see Lemma 4 below), although obtaining Ei{n) is more difficult,
as the average over x is usually analytically intractable.
3
Optimal finite-dimensional models
We start considering two classes of linear approximators,
namely
11.1 {g(x) = k' (x)LIJL E jRmxn} and 11.2= {h(x) = k' (x)FiJ>'IJF E jRmxm}, where
the matrices Land F are possibly dependent on the Xi samples. We point out that
11.2 C 11.1 and that the PBR predictor b(x) E 11. 2. Our goal is the characterization of
the optimal predictors in 11.1 and 11. 2. Before stating the main result, two preliminary
lemmas are given. The first one is proved in (Pilz, 1991) while the second follows from
a straightforward calculation.
=
Lemma 3 Let A E jRnxn, BE jRnxr, A> O. Then it holds that
inf
ZERrxn
Tr [(ZAZ' - ZB - B' Z')] = Tr [-B' A-I B]
and the minimum is achieved for the matrix Z* = B' A-I .
Lemma 4 Let g(x) E 11. 1 ? Then it holds that
+00
Eg(n,x) = LAi + (12 + q{L), q(L)=Tr [LHL' - 2LiJ>A].
i=1
Finite-Dimensional Approximation o/Gaussian Processes
Proof.
[C(x*, xd
Et< ,t
In view of the
C(x*, x n )]' ,it holds
[(t* - k' (x*)U)2]
(12
x-error
221
definition,
setting
r( x*)
+ C(X*, X*) + k' (x*)LH L' k(x*)
-2k' (x*)Lr(x*)
(12 + C(x*, x*)
(3)
+Tr [LHL'k(x*)k' (x*) - 2Lr(x*)k' (x*)] .
Note that Ex< [k(x*)k' (x*)] = fm, Ex' [r(x*)k' (x*)] = ?I> A , and, from the MercerHilbert expansion (1), Ex' [C(x*, x*)] = I:~~ Ai? Then, taking the mean of (3) w.r.t.
x*, the result follows.D
Theorem 5 The predictors gO(x) E 11.1 given by L = ?0 = A?I>' H- 1 and hO(x) E 11.2
given by F = FO = A?I>' ?1>(<<1>' H?I?-I, Vn 2: m, are x-optimal. Moreover
+00
Ego(n,x) = LAi+(12-Tr[A?I>'H-1?1>A]
(4)
i=1
+00
L Ai
+ (12
-
Tr [A?I>' ?1>(<<1>' H?I?-I?I>' ?I>A]
i=1
Proof. We start considering the gO(x) case. In view of Lemma 4 we need only to
minimize q(L) w.r.t. to the matrix L. By applying Lemma 3 with B = ?I>A, A = H > 0,
Z = L, one obtains
argmlnq(L)=Lo = A?I>'H- 1
mlnq(L) = -Tr [A?I>'H- 1?1>A]
(5)
so proving the first result . For the second case, we apply Lemma 4 with L = F?I>' and
then perform the minimization of q(F?I>'), w.r.t. the matrix F. This can be done as
before noting that ?1>' H-I?I> > 0 only when n 2: m. 0
Note that the only difference between gO(x) and the GP predictor derives from the
approximation of the fUnctions C(x, Xk) with I::l Ai'Pi(X)'Pi(Xk) . Moreover the complexity of gO (x) is O(n 3) the same of j(x). On the other hand hO(x) scales as O(n 2m),
so having a computational cost intermediate between the GP predictor and PBR. Intuitively, the PBR method is inferior to hO as it does not take into account the x locations
in setting up its prior. We can also show that the PBR predictor b(x) and hO(x) are
asymptotically equivalent.
l,From (4) is clear that the explicit evaluations of E%o(n) and Eho(n) are in
general very hard problems, because the mean w.r.t.
the Xi samples that
enters in the ?I> and H matrices.
In the remainder of this section we
will derive an upper bound on Eho(n).
Consider the class of approximators
11.3= {u(x) = k' (x) D ?1>' ~D E ffi.mxm , (D)ij = di 6ij }. Because of the inclusions 11.3 C
11.2 C 11. 1 , if UO(x) is the x-optimal predictor in 11. 3, then Ego(n) ::; Eho(n) ::; E;o(n).
Due the diagonal structure of the matrix D, an upper bound to E~o (n) may be explicitly computed, as stated in the next Theorem.
Theorem 6 The approximator UO(x) E 11.3 given by
(<<I>' ?I> A) .
(D)ij = (DO)ij = (<<I>' H?I?:: 6ij ,
(6)
G. Ferrari-Trecate, C. K. I. Williams and M. Opper
222
is x-optimal. Moreover an upper-bound on its generalization error is given by
+=
E;o
<
L Ai +
m
i=l
L qkAk,
k=l
(n -l)Ak
Ck
n
(J2 -
+
Ak
qk = Ck
JC(x,x)cp~(x)p(x)dx
(7)
+(J2.
Proof.
In order to find the x-optimal approximator in 11. 3 , we start applying the
Lemma 4 with L = Dq,'. Then we need to minimize
(8)
w.r.t. di so obtaining (6). To bound E;o(n), we first compute the generalization error
of a generic approximation u(x) that is
verifying that
EZ = Ex
[q(Dq,')] + L~~ Ai +
(J2.
After
we obtain from (8), assuming the d i constant,
E~
+=
=L
i=l
Ai +
(J2
+n
m
m
i=l
i=l
L d;Ci - 2n L diAi.
Minimizing EZ w.r.t. d i , and recalling that UO(x) is also simply optimal the formula
(7) follows.O When C(6, 6) is stationary, the expression of the Ci coefficient becomes
simply Ci = (n - l)Ai + L~~ Ai + (J2 .
Remark : A naive approach to estimating the coefficients in the estimator
L~lWi?i(X) would be to set Wi = n- 1 (q,'t)i as an approximation to the integral
Wi = f ?i(x)f(x)p(x)dx. The effect of the matrix D is to "shrink" the wi's of the
higher-frequency eigenfunctions. If there was no shrinkage it would be necessary to
limit m to stop the poorly-determined Wi'S from dominating, but equation 7 shows
that in fact the upper bound is improved as m increases. (In fact equation 7 can be
used as an upper bound on the GP prediction error; it is tightest when m ~ 00.) This
is consistent with the idea that increasing m under a Bayesian scheme should lead to
improved predictions. In practice one would keep m < n, otherwise the approximate
algorithm would be computationally more expensive than the O(n3) GP predictor.
4
Choosing m
For large n, we can show that
(9)
where b(x) is the PBR approximator of Definition 1. (This arises because the matrix
q,/q, becomes diagonal in the limit n ~ 00 due to the orthogonality of the eigenfunctions.)
In equation 9, the factor (Ail + ,Bn)-l indicates by how much the prior variance of
the ith eigenfunction ?i has been reduced by the observation of the n datapoints.
(Note that this expression is exactly the same as the posterior variance of the mean
223
Finite-Dimensional Approximation o/Gaussian Processes
--,
,
I
00/
tl
"
06
.'.
u.I
o os
- , ,,
,,
,
\"2,
: o~
? --------,,~,------~,,~,------~,,?
log,
(a)
(b)
Figure 1: (a) E~o(n) and detaching points for various model orders. Dashed: m
dash-dot: m = 5, dotted: m = 8, solid: Ejen). (b) Eg(n) - E~ o (n) plotted against n.
=
3,
of a Gaussian with prior N(O , Ai) given n observations corrupted by Gaussian noise
of variance {3-1 .) For an eigenfunction with Ai ? 0- 2 In, the posterior is considerably
tighter than the prior, but when Ai ? 0- 2 In, the prior and posterior have almost the
same width , which suggests that there is little point in including these eigenfunctions
in the finite-dimensional model. By omitting all but the first m eigenfunctions we add
a term L~m+1 Ai to the expected generalization error.
This means that for a finite-dimensional model using the first m eigenfunctions, we
expect that Eg(n) ~ Ej(n) up to a training set size n determined by n = 1/({3Am).
We call n the detatching point for the m-dimensional approximator. Conversely, in
practical regression problems the data set size n is known. Then, from the knowledge
of the auto covariance eigenvalues, is possible to determine, via the detatching points
formula, the order m of the approximation that should be used in order to guarantee
Eho (n) ~ Ej(n).
5
Experimental results
We have conducted experiments using the prior covariance function C(6,6) = (1 +
h)e- h where h = 16 - 61/p? with p = 0.1. This covariance function corresponds
to a Gaussian process which is once mean-squared differentiable, It lies in the family
of stationary covariance functions C(h) = hV Kv(h) (where Kv(-) is a modified Bessel
function) , with v = 3/2. The eigenvalues and eigenfunctions of this covariance kernel
for the density p(x) '" U(O, 1) have been calculated in Vivarelli (1998).
In our first experiment (using 0- 2 = 1) the learning curves of b(x), hO(x) and i(x) were
obtained; the average over the choice of training data sets was estimated by using 100
different x samples. It was noticed that Eg (n) and Eho(n) practically coincide, so only
the latter curve is drawn in the pictures.
In Figure l(a) we have plotted the learning curves for GP regression and the approximation hO(x) for various model orders. The corresponding detaching points are
also plotted, showing their effectiveness in determining the size of data sets for which
E~ o (n) ~ Ej(n). The minimum possible error attainable is (J2 = 1.0 For finitedimensional models this is increased by L~m+l Ai; these "plateaux" can be clearly
seen on the right hand side of Figure l(a).
224
G. Ferrari-Trecate, C. K. J. Williams and M. Opper
Our second experiment demonstrates the differences in performance for the hO(x) and
b(x) estimators, using (72 = 0.1. In Figure 1(b) we have plotted the average difference
Eg(n) - Eho(n). This was obtained by averaging Eb(n,x) - Eho(n,x) (computed with
the same x, i.e. a paired comparison) over 100 choices of x, for each n. Notice that
hO is superior to the PBR estimator for small n (as expected), but that they are
asymptotically equivalent.
6
Discussion
In this paper we have shown that a finite-dimensional predictor hO can be constructed which has lower generalization error than the PBR predictor. Its computational
complexity is O(n 2 m), lying between the O(n 3 ) complexity of the GP predictor and
O(m 2 n) complexity of PBR. We have also shown how to calculate m, the number of
basis functions required, according to the data set size.
We have used finite-dimensional models to approximate GP regression. An interesting
alternative is found in the work of Gibbs and MacKay (1997), where approximate
matrix inversion methods that have O(n 2 ) scaling have been investigated. It would be
interesting to compare the relative merits of these two methods.
Acknowledgements
We thank Francesco Vivarelli for his help in providing the learning curves for Ej(n) and the
eigenfunctions/values in section 5.
References
[1) De Nicolao, G., and Ferrari Trecate, G. (1998). Identification of NARX models using
regularization networks: a consistency result .. IEEE Int. Joint Conf. on Neural Networks,
Anchorage, US, pp. 2407-2412.
[2) Gibbs, M. and MacKay, D. J. C.'(1997). Efficient Implementation of Gaussian Processes. Cavendish Laboratory, Cambridge, UK. Draft manuscript, available from
http://wol.ra.phy.cam.ac.uk/mackay/homepage.html.
[3) Opper, M. (1997). Regression with Gaussian processes: Average case performance. In I. K.
Kwok-Yee, M. Wong and D.-Y. Yeung (eds), Theoretical Aspects of Neural Computation:
A Multidisciplinary Perspective. Springer-Verlag.
[4) Pilz, J. (1991). Bayesian estimation and experimental design in linear regression models.
Wiley & Sons.
[5) Ripley, B. D. (1996). Pattern recognition and neural networks. CUP.
[6) Wahba, G. (1990). Spline models for observational data. Society for Industrial and Applied
Mathematics. CBMS-NSF Regional Conf. series in applied mathematics.
[7) Whittle, P. (1963). Prediction and regUlation by linear least-square methods. English Universities Press.
[8) Williams C. K. I. (1998). Prediction with Gaussian processes: from linear regression to
linear prediction and beyond. In Jordan, M.I. editor, Learning and inference in graphical
models. Kluwer Academic Press.
[9] Vivarelli, F. (1998).Studies on generalization in Gaussian processes and Bayesian Neural
Networks. Forthcoming PhD thesis, Aston University, Birmingham, UK.
[10] Zhu, H., and Rohwer, R. (1996). Bayesian regression filters and the issue of priors. Neural
Computing and Applications, 4:130-142.
[11) Zhu, H., Williams, C. K. I. Rohwer, R. and Morciniec, M. (1997). Gaussian regression and
optimal finite dimensional linear models. Tech. Rep. NCRG/97/011. Aston University,
Birmingham, UK.
| 1497 |@word inversion:2 bn:1 covariance:7 attainable:1 tr:7 solid:1 phy:1 series:1 outperforms:1 dx:2 numerical:1 stationary:2 intelligence:1 xk:2 ith:1 lr:2 manfred:1 draft:1 characterization:1 location:1 dn:4 constructed:1 anchorage:1 introduce:1 ra:1 expected:2 growing:1 decreasing:1 pbr:14 little:1 considering:2 increasing:1 becomes:2 estimating:2 moreover:5 homepage:1 ail:1 guarantee:1 ti:1 xd:2 exactly:1 demonstrates:1 uk:7 uo:3 superiority:1 before:2 engineering:1 morciniec:1 limit:2 ak:2 eb:1 suggests:2 conversely:1 pui:1 practical:1 practice:2 lf:1 empirical:1 projection:1 applying:2 yee:1 wong:1 equivalent:2 backed:1 williams:6 straightforward:1 go:4 independently:1 estimator:4 datapoints:1 his:1 proving:1 ferrari:7 cavendish:1 suppose:1 us:1 ego:2 expensive:1 recognition:1 enters:1 verifying:1 hv:1 calculate:2 complexity:7 cam:1 division:1 basis:3 joint:1 various:2 artificial:1 choosing:1 whose:1 dominating:1 otherwise:1 gp:14 noisy:1 obviously:1 eigenvalue:3 differentiable:1 product:1 remainder:1 j2:6 poorly:1 kv:2 help:1 derive:3 ac:3 stating:1 ij:12 op:1 come:1 fij:1 filter:1 wol:1 observational:1 generalization:7 preliminary:1 tighter:1 dipartimento:1 hold:3 practically:1 lying:1 estimation:1 birmingham:3 minimization:1 clearly:1 gaussian:20 modified:1 ck:2 ej:4 shrinkage:1 indicates:1 tech:1 industrial:1 am:1 inference:1 dependent:1 going:1 interested:1 issue:1 html:1 mackay:3 once:1 having:1 spline:1 cheaper:1 recalling:1 interest:1 evaluation:1 integral:1 necessary:1 lh:1 autocovariance:1 plotted:4 theoretical:2 minimal:2 increased:1 cost:2 predictor:19 conducted:1 corrupted:1 accomplish:1 considerably:1 density:1 squared:1 thesis:1 possibly:1 conf:2 account:1 de:2 whittle:2 coefficient:2 int:1 jc:1 explicitly:1 depends:2 performed:1 view:3 start:3 bayes:3 minimize:2 square:1 variance:4 qk:1 bayesian:5 identification:2 plateau:1 fo:1 suffers:1 ed:2 definition:5 rohwer:2 against:1 frequency:1 pp:1 naturally:1 proof:3 di:4 mi:1 stop:1 proved:1 knowledge:1 hilbert:1 back:1 cbms:1 manuscript:1 higher:1 improved:2 zaz:1 done:1 shrink:1 hand:2 christopher:1 ei:4 o:1 multidisciplinary:1 name:1 effect:1 omitting:1 analytically:1 regularization:1 laboratory:1 eg:5 width:1 inferior:1 hill:1 cp:1 superior:1 ji:1 b4:1 ncrg:1 kluwer:1 cambridge:1 gibbs:2 ai:16 cup:1 consistency:1 mathematics:2 inclusion:1 dot:1 add:1 posterior:4 perspective:1 italy:1 belongs:1 inf:1 verlag:1 rep:1 approximators:2 seen:1 minimum:2 dai:1 determine:1 bessel:1 dashed:1 jrd:1 academic:1 calculation:5 lai:2 paired:1 prediction:8 regression:15 controller:1 mxm:1 yeung:1 kernel:1 achieved:1 regional:1 eigenfunctions:8 effectiveness:1 jordan:1 call:3 noting:1 intermediate:1 xj:1 forthcoming:1 fm:1 wahba:1 idea:1 trecate:6 motivated:2 expression:2 ird:2 remark:1 clear:1 amount:1 extensively:1 informatica:1 reduced:2 http:1 nsf:1 notice:1 dotted:1 delta:1 estimated:1 group:1 drawn:2 pj:1 asymptotically:5 family:1 throughout:1 almost:1 electronic:1 forrest:1 vn:1 scaling:2 rpm:1 bound:7 giancarlo:1 dash:1 detaching:2 kronecker:1 orthogonality:1 n3:3 aspect:1 span:1 optimality:2 department:1 according:1 belonging:1 jr:1 son:1 wi:4 intuitively:1 computationally:2 equation:3 mutually:1 describing:1 ffi:1 vivarelli:3 know:1 merit:1 available:1 tightest:1 apply:1 ckiw:1 kwok:1 generic:1 alternative:1 ho:11 graphical:1 eho:7 narx:1 calculating:1 approximating:2 universita:1 society:1 noticed:1 eh1:1 question:1 diagonal:2 said:1 thank:1 d6:1 assuming:2 providing:1 minimizing:1 ql:1 difficult:1 regulation:1 statement:1 stated:1 design:2 implementation:1 motivates:1 unknown:1 perform:1 upper:6 observation:3 francesco:1 finite:19 namely:1 required:1 specified:1 eigenfunction:2 beyond:1 dynamical:1 below:1 usually:1 pattern:1 regime:1 including:1 suitable:1 oij:2 zhu:4 scheme:1 aston:4 picture:1 carried:1 auto:2 naive:1 lij:1 prior:11 acknowledgement:1 determining:1 asymptotic:1 relative:1 expect:1 interesting:3 approximator:6 consistent:1 mercer:1 dq:2 editor:1 pi:8 land:1 lo:2 last:1 english:1 side:1 taking:1 edinburgh:2 curve:4 opper:6 dimension:1 xn:2 calculated:1 finitedimensional:1 projected:3 coincide:1 approximate:4 obtains:1 keep:1 dealing:1 assumed:1 xi:6 jrnxn:1 ripley:1 decade:1 obtaining:2 expansion:2 investigated:1 main:2 linearly:1 noise:1 tl:2 wiley:1 explicit:2 xl:1 lie:1 formula:3 theorem:3 showing:1 derives:1 intractable:1 ci:3 phd:1 easier:1 simply:4 nicolao:2 ez:2 ordered:1 springer:1 corresponds:1 ijf:1 lwi:1 goal:1 lhl:2 hard:1 determined:2 averaging:1 lemma:8 zb:1 e:4 experimental:2 latter:1 arises:2 ex:5 |
544 | 1,498 | Making Templates Rotationally Invariant:
An Application to Rotated Digit Recognition
Shurneet Baluja
baluja@cs.cmu.edu
Justsystem Pittsburgh Research Center &
School of Computer Science, Carnegie Mellon University
Abstract
This paper describes a simple and efficient method to make template-based
object classification invariant to in-plane rotations. The task is divided into two
parts: orientation discrimination and classification. The key idea is to perform
the orientation discrimination before the classification. This can be accomplished by hypothesizing, in turn, that the input image belongs to each class of
interest. The image can then be rotated to maximize its similarity to the training images in each class (these contain the prototype object in an upright orientation). This process yields a set of images, at least one of which will have the
object in an upright position. The resulting images can then be classified by
models which have been trained with only upright examples. This approach
has been successfully applied to two real-world vision-based tasks: rotated
handwritten digit recognition and rotated face detection in cluttered scenes.
1 Introduction
Rotated text is commonly used in a variety of situations, ranging from advertisements,
logos, official post-office stamps, and headlines in magazines, to name a few. For examples, see Figure 1. We would like to be able to recognize these digits or characters, regardless of their rotation.
Figure 1: Common examples of images which contain text that is not axis aligned include logos, post-office
stamps, magazine headlines and consumer advertisements.
848
S. Baluja
The focus of this paper is on the recognition of rotated digits. The simplest method for creating a system which can recognize digits rotated within the image-plane is to employ
existing systems which are designed only for upright digit recognition [Le Cun et aI.,
1990][Le Cun et a!., 1995a][Le Cun et ai., 1995b][Lee, 1991][Guyon et a!., 1989]. By
repeatedly rotating the input image by small increments and applying the recognition system at each rotation, the digit will eventually be recognized. As will be discussed in this
paper, besides being extremely computationally expensive, this approach is also errorprone. Because the classification of each digit must occur in many orientations, the likelihood of an incorrect match is high.
The procedure presented in this paper to make templates rotationally invariant is significantly faster and more accurate than the one described above. Detailed descriptions of the
procedure are given in Section 2. Section 3 demonstrates the applicability of this approach
to a real-world vision-based task, rotated handwritten digit recognition. Section 4 closes
the paper with conclusions and suggestions for future research. It also briefly describes the
second application to which this method has been successfully applied, face detection in
cluttered scenes.
2 Making Templates Rotationally Invariant
The process to make templates rotationally invariant is easiest to describe in the context of
a binary classification problem; the extension to multiple classes is discussed later in this
section. Imagine a simplified version of the digit recognition task: we want a detector for a
single digit. Suppose we wish to tell whether the input contains the digit '3' or not. The
challenge is that the '3' can be rotated within the image plane by an arbitrary amount.
Recognizing rotated objects is a two step process. In the first step, a "De-Rotation" network is applied to the input image. This network analyzes the input before it is given to a
"Detection" network. If the input contains a '3', the De-Rotation network returns the
digit's angle of rotation. The window can then be rotated by the negative of that angle to
make the '3' upright. Note that the De-Rotation network does not require a '3' as input. If
a non- ' 3' image is encountered, the De-Rotation network will return an unspecified rotation. However, a rotation of a non- '3' will yield another (perhaps different) image of a
non-'3'. When the resulting image is given to the Detection network it will not detect a
'3'. On the other hand, a rotated '3', which may not have been detected by the Detection
network alone, will be rotated to an upright position by the De-Rotation network, and will
subsequently be detected as a '3' by the Detection network.
The Detection network is trained to output a positive value only if the input contains an
upright '3', and a negative value otherwise (even if it contains a rotated '3 '). It should be
noted that the methods described here do not require neural networks. As shown in [Le
Cun et al., 1995a, Le Cun et ai., 1995b] a number of other classifiers can be used.
The De-Rotation and Detection networks are used sequentially. First, the input image is
processed by the De-Rotation network which returns an angle of rotation, assuming the
image contains a '3'. A simple geometric transformation of the image is performed to
undo this rotation. If the original image contained a '3', it would now be upright. The
resulting image is then passed to the Detection network. If the original image contained a
'3', it can now be successfully detected.
This idea can easily be extended to multiple-class classification problems: a De-Rotation
network is trained for each object class to be recognized. For the digit recognition problem, 10 De-Rotation networks are trained, one for each of the digits 0.. 9. To classify the
digits once they are upright, a single classification network is used with 10 outputs
(instead of the detection networks trained on individual digits - alternative approaches
will be described later in this paper). The classification network is used in the standard
manner; the output with the maximum value is taken as the classification. To classify a
new image, the following procedure is used:
849
Making Templates Rotationally Invariant
For each digitD (0 $; D
$;
9):
1.
Pass image through De-Rotation-network-D. This returns the rotation angle.
2.
Rotate the image by (-1.0 * returned rotation angle).
3.
Pass the de-rotated image to the classification network.
4.
If the classification network's maximum output is output D, the activation of
output D is recorded. Otherwise digit D is eliminated as a candidate.
In most cases, this will eliminate all but one of the candidates. However, in some cases
more than one candidate will remain. In these cases, the digit with the maximum recorded
activation (from Step 4) is returned. In the unlikely event that no candidates remain, either
the system can reject the sample as one it cannot classify, or it can return the maximum
value which would have been recorded in Step 4 if none of the examples were rejected.
2.1 Network Specifics
To train the De-Rotation networks, images of rotated digits were input, with the rotation
angle as the target output. Examples of rotated digits are shown in Figure 2. Each image is
28x28 pixels. The upright data sets are from the MNIST database [Le Cun et at. , 1995a].
.?-
?~~GJmmm? mm.
:
::&lIi?? ?\:? ?
IJ-~~---?.-:
.
?. :~..........
?. . .?IitIIUI:....
iB ? ? ?:R llil lll :? ?
11? ? ?? ? :? ?
n
:tM.",.m:IWMI:II.
~R.J aLIla U ~.:II.r. ~ . I!II:
~B.::mll~II:l\Wa
:...~ ...... _........~u. . . . . . . ;a . . . :. . .;aa ..........
? ? ? ?l ? ?:? ?!
? ?!:? ?
ill.?.11
Figure 2: 8 examples of each of the 10 digits to be recognized. The first example in each group of eight
is shown with no rotation; it is as it appears in the MNIST data set. The second through eighth examples
show the same digit rotated in-plane by random amounts.
In the classification network, each output represents a distinct class; therefore, the standard l-of-N output representation was used with 10 outputs. To represent a continuous
variable (the angle of rotation) in the outputs of the De-Rotation network, we used a Gaussian output encoding [Pomerleau, 1992] with 90 output units. With the Gaussian encoding, instead of only training the network to activate a single output (as is done in l-of-N
encoding), outputs close to the desired output are also activated in proportion to their distance from the desired output. This representation avoids the imposed discontinuities of
the strict l-of-N encoding for images which are similar, but have only slight differences in
rotations. Further, this representation allows finer granularity with the same number of
output units than would be possible if a l-of-N encoding was used [Pomerleau, 1992].
The network architecture for both the classification and the De-Rotation networks consists
of a single hidden layer. However, unlike a standard fully-connected network, each hidden
unit was only connected to a small patch of the 28x28 input. The De-Rotation networks
used groups of hidden units in which each hidden unit was connected to only 2x2, 3x3,
4x4 & 5x5 patches of the inputs (in each of these groups, the patches were spaced 2x2 pixels apart; therefore, the last three groups had overlapping patches). This is similar to the
networks used in [Baluja, 1997][Rowley et. at, 1998a, 1998b] for face detection. Unlike
the convolution networks used by [Le Cun et aI., 1990], the weights into the hidden units
were not shared. 1 Note that many different local receptive field configurations were tried;
almost all had equivalent performance.
S. Ba/uja
850
3 Rotated Handwritten Digit Recognition
To create a complete rotationally invariant digit recognition system, the first step is to segment each digit from the background. The second is to recognize the digit which has been
segmented. Many systems have been proposed for segmenting written digits from background clutter [Jain & Yu, 1997][Sato et ai., 1998][Satoh & Kanade, 1997]. In this paper,
we concentrate on the recognition portion of the task. Given a segmented image of a
potentially rotated digit, how do we recognize the digit?
The first experiment conducted was to establish the base-line performance. We used only
the standard, upright training set to train a classification network (this training set consists
of 60,000 digits). This network was then tested on the testing set (the testing set contains
10,000 digits) . In addition to measuring the performance on the upright testing set, the
entire testing set was also rotated. As expected, performance rapidly degrades with rotation. A graph of the performance with respect to the rotation angle is shown in Figure 3.
Perfonaaneeo' Network trained wtth t Jpript Dlpts
and Tcwhd on Rotated Dlgtu
i
]
1
J
Figure 3:
Performance of the classification
network trained only with upright images when
tested on rotated images. As the angle of
rotation increases, performance degrades. Note
the spike around 180 degrees, this is because
some digits look the same even when they are
upside-down. The peak performance is
approximately 97.5% (when the digits are
upright).
It is interesting to note that around 1800 rotation, performance slightly rises. This is
because some of the digits are symmetric across the center horizontal axis - for example
the digits '8', '1', '2' & '5' can be recognized upside-down. Therefore, at these orientations, the upright detector works well for these digits.
As mentioned earlier, the simplest method to make an upright digit classifier handle rotations is to repeatedly rotate the input image and classify it at each rotation. Thefirst drawback to this approach is the severe computational expense. The second drawback is that
because the digit is examined at many rotations, it may appear similar to numerous digits
in different orientations. One approach to avoid the latter problem is to classify the digit as
the one that is voted for most often when examined over all rotations. To ensure that this
process is not biased by the size of the increments by which the image is rotated, various
angle increments are tried. As shown in the first row of Table I, this method yields low
Table I: Exhaustive Search over all possible rotations
Number of Angle IncreQ1ents Tried
Exhaustive Search Method
360
100
50
(1 degree/increment) (3.6 degree/increment) (7.2 degreeslincrement
Most frequent vote (over all rotations)
59.5%
66.0%
65 .0%
Most frequent vote - counted onl y when votes are positi ve
(over all rotations)
75.2%
74.5%
74.0%
1. Note that in the empirical comparisons presented in [Le Cun et ai., 1995aJ, convolution networks performed
extremely well in the upright digit recognition task. However, due to limited computation resources, we were
unable to train these networks, as each takes 14-20 days to train. The network used here was trained in 3 hours,
and had approximately a 2.6% misclassification rate on the upright test set. The best networks reported in [Le
Cun et ai, 1995aJ have less than 1% error. It should be noted that the De-Rotation networks trained in this study
can easily be used in conjunction with any classification procedure, including convolutional networks.
Making Templates Rotationally Invariant
851
classification accuracies. One reason for this is that a vote is counted even when the classification network predicts all outputs to be less than 0 (the network is trained to predict
+1 when a digit is recognized, and -1 when it is not). The above experiment was repeated
with the following modification: a vote was only counted when the maximum output of
the classification network was above O. The result is shown in the second row of Table I.
The classification rate improved by more than 10%.
Given these base-line performance measures2 , we now have quantitative measurements
with which to compare the effectiveness of the approach described in this paper. The performance of the procedure used here, with 10 "De-Rotation" networks and a single classification network, is shown in Figure 4. Note that unlike the graph shown in Figure 3, there
is very little effect on the classification performance with the rotation angle.
Figure 4:
Performance ofthe combined DeRotation network and classification network
system proposed in this paper. Note that the
performance is largely unaffected by the
rotation. The average performance, over all
rotations, is 85.0%.
0.??
~~----------------~~
0,1.
-,*10.00
-100.00
0.00
100.110
Ito.OO
~011""'1oft
To provide some intuition of how the De-Rotation networks perform, Figure 5 shows
examples of how each De-Rotation networks transform each digit. Each De-Rotation network suggests a rotation which makes the digit look as much like the one with which the
network was trained. For example, De-Rotation-Network-5 will suggest a rotation that
will make the input digit look as much like the digit '5' as possible; for example, see DeRotation-Network-5's effect on the digit '4'.
o
2
Original Digit
345
6
7
8
9
Digit rotated by De-Rotation-Network-O
Digit rotated by De-Rotation-Network-l
Digit rotated by De-Rotation-Network-2
Digit rotated by De-Rotation-Network-3
Digit rotated by De-Rotation-Network-4
Digit rotated by De-Rotation-Network-5
Digit rotated by De-Rotation-Network-6
Digit rotated by De-Rotation-Network-7
Digit rotated by De-Rotation-Network-8
Digit rotated by De-Rotation-Network-9
Figure 5: Digits which have been rotated by the angles specified by each of the De-rotation networks. As
expected (if the method is working), the digits on the diagonal (upper left to bottom right) appear upright.
2. Another approach is to train a single network to handle both rotation and classification by using rotated digits
as inputs, and the digit's classification as the target output. Experiments with the approach yielded results far
below the techniques presented here.
852
S. Baluja
As shown in Figure 4, the average classification accuracy is approximately 85.0%. The
performance is not as good as with the upright case alone, which had a peak performance
of approximately 97.5% (Figure 3). The high level of performance achieved in the upright
case is unlikely for rotated digits: if all rotations are admissible, some characters are
ambiguous. The problem is that when working correctly, De-Rotation-Network-D will
suggest an angle of rotation that will make any input image look as much like the digit D
as possible through rotation. In most cases when the input image is not the digit D, the
rotation will not cause the image to look like D. However, in some cases, such as those
shown in Figure 6(right), the digit will be transformed enough to cause a classification
error. Some of these errors will most likely never be correctable (for example, '6' and '9'
in some instances); however, there is hope for correcting some of the others.
Figure 6 presents the complete confusion matrix. As can be seen in the examples in Figure
6(right), the digit '4' can be rotated to appear similar to a'S'. Nonetheless, there often
remain distinctive features that allow real '5's to be differentiated from the rotated '4's.
However, the classification network is unable to make these distinctions because it was
not trained with the appropriate examples. Remember, that since the classification network was only trained with the upright digit training set, rotated '4's are never encountered during training. This reflects a fundamental discrepancy in the training/testing
procedure. The distributions of images which were used to train the classification network
is different than the distributions on which the network is tested.
To address this problem, the classification mechanism is modified. Rather than using the
single 1-oj-1O neural network classifier used previously, 10 individual Detection networks
are used. Each detection network has a single binary output that signifies whether the
input contains the digit (upright) with which the network was trained. Each De-Rotation
network is paired with the respective Detection network. The crucial point is that rather
than training the Detection-Network-D with the original upright images in the training
set, each image (whether it is a positive or negative example) is first passed through DeRotation-Network-D. Although this makes training Detection-Network-D difficult since
all the digits are rotated to appear as much like upright-D's as possible by De-RotationNetwork-D, the distribution of training images matches the testing distribution more
closely. In use, when a new image is presented, it is passed through the lO network pairs.
Candidate digits are eliminated if the binary output from the detection network does not
signal a detection. Preliminary results with this new approach are extremely promising;
the classification accuracy increases dramatically - to 93% when averaged over all rotations. This is a more than a 50% reduction in error over the previously described approach.
Predicted Digit
o
a
2
94
--
--
90
1I
--
5
88
--
3
88
--
--
4
is
3
4
89
3
3
~
5
2
87
4
2
88
3
6
74
6
3
--
5
3
7
25
2
B.
!:a.:al liBII! 'JaBEI.
c?III?BEt: "11.'111 '-? .?:?.
3
]
10
:.! ,?.? ?!.?.?111
A. ::? ?
--
.~
7
Original Image
mage Rotated to Look Like Mistake Digit
r1mage Rotated to Look Like Correct Digit
2345678
:.1 :11.11111'.??~??
D. [? ?
3
--
89
--
--
--
64
l [? ??.la.II:.!
E?l? . ? ?
F.
::.:11:111
r.~~.:.:?.l
i?? . ?
111:11'1:: ;??11?
G. [? ? ??
Figure 6: Example errors. (LEFI) Confusion Matrix (only entries account for 2% or more entries are
filled in for ease of reading). (RlGH1) some of the errors made in classification. 3 examples of each of the
errors are shown. Row A: '4' mistaken as '5'. Row B: '5' mistaken as '6', Row C: '7' mistaken as '2'. Row
D: '7' mistaken as '6'. Row E: '8' mistaken as '4', Row F: '9' mistaken as '5', Row G: '9' mistaken as '6'.
Making Templates Rotationally Invariant
853
4 Conclusions and Future Work
This paper has presented results on the difficult problem of rotated digit recognition. First,
we presented base-line results with naive approaches such as exhaustively checking all
rotations. These approaches are both slow and have large error rates. Second, we presented results with a novel two-stage approach which is both faster and more effective
than the naive approaches. Finally, we presented preliminary results with a new approach
that more closely models the training and testing distributions.
We have recently applied the techniques presented in this paper to the detection of faces in
cluttered scenes. In previous studies, we presented methods for finding all upright frontal
faces [Rowley et aT., 1998aJ. By using the techniques presented here, we were able to
detect all frontal faces, including those which were rotated within the image plane [Baluja,
1997][Rowley et al., 1998bJ. The methods presented in this paper should also be directly
applicable to full alphabet rotated character recognition.
In this paper, we examined each digit individually. A straight-forward method to eliminate
some of the ambiguities between rotationally similar digits is to use contextual information. For example, if surrounding digits are all rotated to the same amount, this provides
strong hints about the rotation of nearby digits. Further, in most real-world cases, we
might expect digits to be close to upright; therefore, one method of incorporating this
information is to penalize matches which rely on large rotation angles.
This paper presented a general way to make template-based recognition rotation invariant.
In this study, both the rotation estimation procedures and the recognition templates were
implemented with neural-networks. Nonetheless, for classification, any technique which
implements a form of templates, such as correlation templates, support vector machines,
probabilistic networks, K-Nearest Neighbor, or principal component-based methods,
could have easily been employed.
Acknowledgements
The author would like to thank Kaari Aagstad for her reviews of many successive drafts of this paper.
References
Baluja, S. (1997) "Face Detection with In-Plane Rotation: Early Concepts and Preliminary Results," Justsystem
Pittsburgh Research Center Technical Report. JPRC-TR-97-001.
Guyon, I, Poujaud, I., Personnaz, L, Dreyfus, G., Denker, J. LeCun, Y. (1989) "Comparing Different Neural Net
Architectures for Classifying Handwritten Digits", in IlCNN II 127-132.
Jain, A. & Yu, B. (1997) "Automatic Text Location in Images and Video Frames", TR: MSUCPS: TR 97-33.
Le Cun, Y., Jackel, D., Bottou, L, Cortes, c., Denker, J. Drucker, J. Guyon, I, Miller, U. Sackinger, E. Simard, P.
Vapnik, V. (1995a) "Learning Algorithms for Classification: A Comparison on Handwritten Digit Recognition".
Neural Networks: The Statistical Mechanics Perspective, Oh, J., Kwon, C. & Cho, S. (Ed.), pp. 261-276.
LeCun, Y., Jackel, L. D., Bottou, L., Brunot, A., Cortes, C., Denker, J. S., Drucker, H., Guyon, I., Muller, U. A.,
Sackinger, E., Simard, P. and Vapnik, V. (1995b), Comparison of learning algorithms for handwritten digit recognition," ICANN, Fogelman, F. and Gallinari, P., 1995, pp. 53-60.
LeCun, Y., Boser, B., Denker, J. S., Henderson, D., Howard, R. E., Hubbard, W. and Jackel, L. D. (1990),
"Handwritten digit recognition with a back-propagation network," Advances in Neural Information Processing
Systems 2 (NIPS '89), Touretzky, David (Ed.), Morgan Kaufman.
Lee, Y. (1991) "Handwritten Digit Recognition using K-NN, RBF and Backpropagation Neural Networks", Neural Computation, 3, 3.
Pomerleau, D.A. (1993) Neural Network Perception for Mobile Robot Guidance, Kluwer Academic
Rowley, H., Baluja, S. & Kanade, T. (1998a) "Neural Network-Based Face Detection," IEEE-Transactions on
Pattern Analysis and Machine Intelligence (PAMI), Vol. 20, No.1, January, 1998.
Rowley, H., Baluja, S. & Kanade, T. (1998b) "Rotation Invariant Neural Network-Based Face Detection," to
appear in Proceedings of Computer Vzsion and Pattern Recognition, 1998.
Sato, T, Kanade, T., Hughes, E. & Smith, M. (1998) "Video OCR for Digital News Archives" to appear in IEEE
International Workshop on Content-Based Access of Image and Vzdeo Databases.
Satoh, S. & Kanade, T. (1997) "Name-It: Association of face and name in Video", in Proceedings of IEEE Conference on Computer Vzsion and Pattern Recognition, 1997.
| 1498 |@word version:1 briefly:1 proportion:1 tried:3 tr:3 reduction:1 configuration:1 contains:7 existing:1 contextual:1 comparing:1 activation:2 must:1 written:1 designed:1 discrimination:2 alone:2 intelligence:1 plane:6 smith:1 provides:1 draft:1 location:1 successive:1 incorrect:1 consists:2 manner:1 uja:1 expected:2 mechanic:1 little:1 window:1 lll:1 easiest:1 kaufman:1 unspecified:1 finding:1 transformation:1 quantitative:1 remember:1 demonstrates:1 classifier:3 gallinari:1 unit:6 appear:6 segmenting:1 before:2 positive:2 local:1 mistake:1 encoding:5 approximately:4 pami:1 logo:2 might:1 examined:3 suggests:1 ease:1 limited:1 averaged:1 lecun:3 testing:7 hughes:1 implement:1 x3:1 backpropagation:1 digit:86 procedure:7 empirical:1 poujaud:1 significantly:1 reject:1 suggest:2 cannot:1 close:3 context:1 applying:1 equivalent:1 imposed:1 center:3 regardless:1 cluttered:3 correcting:1 oh:1 handle:2 increment:5 imagine:1 suppose:1 target:2 magazine:2 recognition:22 expensive:1 predicts:1 database:2 bottom:1 connected:3 news:1 mentioned:1 intuition:1 rowley:5 exhaustively:1 trained:14 segment:1 distinctive:1 easily:3 various:1 alphabet:1 train:6 surrounding:1 distinct:1 jain:2 describe:1 activate:1 effective:1 detected:3 tell:1 exhaustive:2 otherwise:2 transform:1 net:1 frequent:2 aligned:1 rapidly:1 description:1 rotated:47 object:5 oo:1 nearest:1 ij:1 school:1 strong:1 implemented:1 c:1 predicted:1 concentrate:1 drawback:2 closely:2 correct:1 subsequently:1 require:2 preliminary:3 extension:1 mm:1 around:2 predict:1 bj:1 early:1 estimation:1 applicable:1 jackel:3 headline:2 individually:1 hubbard:1 create:1 successfully:3 reflects:1 hope:1 gaussian:2 modified:1 rather:2 avoid:1 bet:1 mobile:1 office:2 conjunction:1 focus:1 likelihood:1 detect:2 kaari:1 nn:1 eliminate:2 unlikely:2 entire:1 hidden:5 her:1 transformed:1 pixel:2 fogelman:1 classification:35 orientation:6 ill:1 field:1 once:1 never:2 eliminated:2 x4:1 represents:1 yu:2 look:7 hypothesizing:1 future:2 others:1 discrepancy:1 report:1 hint:1 few:1 employ:1 kwon:1 recognize:4 ve:1 individual:2 mll:1 detection:22 interest:1 severe:1 henderson:1 activated:1 accurate:1 respective:1 filled:1 rotating:1 desired:2 guidance:1 instance:1 classify:5 earlier:1 measuring:1 signifies:1 applicability:1 entry:2 recognizing:1 conducted:1 reported:1 combined:1 cho:1 peak:2 fundamental:1 international:1 lee:2 probabilistic:1 ambiguity:1 recorded:3 creating:1 lii:1 simard:2 return:5 account:1 de:35 later:2 performed:2 portion:1 voted:1 accuracy:3 convolutional:1 largely:1 miller:1 yield:3 spaced:1 ofthe:1 handwritten:8 none:1 finer:1 unaffected:1 straight:1 classified:1 detector:2 touretzky:1 ed:2 nonetheless:2 pp:2 back:1 appears:1 day:1 improved:1 done:1 rejected:1 stage:1 correlation:1 hand:1 working:2 horizontal:1 sackinger:2 overlapping:1 propagation:1 aj:3 perhaps:1 name:3 effect:2 contain:2 concept:1 symmetric:1 x5:1 during:1 ambiguous:1 noted:2 complete:2 confusion:2 image:43 ranging:1 dreyfus:1 novel:1 recently:1 common:1 rotation:76 discussed:2 slight:1 association:1 kluwer:1 mellon:1 measurement:1 ai:7 mistaken:7 automatic:1 satoh:2 had:4 robot:1 access:1 similarity:1 base:3 perspective:1 belongs:1 apart:1 binary:3 accomplished:1 muller:1 rotationally:9 analyzes:1 seen:1 morgan:1 employed:1 recognized:5 maximize:1 signal:1 ii:6 multiple:2 upside:2 full:1 segmented:2 technical:1 match:3 academic:1 faster:2 x28:2 divided:1 post:2 paired:1 vision:2 cmu:1 represent:1 achieved:1 penalize:1 background:2 want:1 addition:1 crucial:1 biased:1 unlike:3 archive:1 strict:1 undo:1 effectiveness:1 granularity:1 iii:1 enough:1 variety:1 architecture:2 idea:2 tm:1 prototype:1 drucker:2 whether:3 passed:3 returned:2 cause:2 repeatedly:2 dramatically:1 detailed:1 amount:3 clutter:1 processed:1 simplest:2 correctly:1 correctable:1 carnegie:1 vol:1 group:4 key:1 graph:2 angle:15 almost:1 guyon:4 patch:4 layer:1 encountered:2 yielded:1 sato:2 occur:1 scene:3 x2:2 nearby:1 extremely:3 wtth:1 describes:2 remain:3 slightly:1 character:3 across:1 cun:10 making:5 modification:1 brunot:1 invariant:11 taken:1 computationally:1 resource:1 previously:2 turn:1 eventually:1 mechanism:1 eight:1 denker:4 ocr:1 differentiated:1 appropriate:1 alternative:1 original:5 include:1 ensure:1 establish:1 personnaz:1 spike:1 receptive:1 degrades:2 diagonal:1 distance:1 unable:2 thank:1 reason:1 consumer:1 assuming:1 besides:1 difficult:2 onl:1 potentially:1 expense:1 negative:3 rise:1 ba:1 pomerleau:3 perform:2 upper:1 convolution:2 howard:1 january:1 situation:1 extended:1 frame:1 arbitrary:1 david:1 pair:1 specified:1 distinction:1 boser:1 hour:1 discontinuity:1 nip:1 address:1 able:2 below:1 perception:1 pattern:3 eighth:1 oft:1 challenge:1 reading:1 including:2 oj:1 video:3 event:1 misclassification:1 rely:1 numerous:1 axis:2 naive:2 text:3 review:1 geometric:1 acknowledgement:1 checking:1 fully:1 expect:1 suggestion:1 interesting:1 digital:1 degree:3 classifying:1 row:9 lo:1 last:1 allow:1 neighbor:1 template:12 face:10 world:3 avoids:1 forward:1 commonly:1 made:1 author:1 simplified:1 counted:3 far:1 transaction:1 sequentially:1 pittsburgh:2 continuous:1 search:2 table:3 kanade:5 promising:1 bottou:2 official:1 icann:1 justsystem:2 repeated:1 slow:1 position:2 wish:1 candidate:5 stamp:2 ib:1 advertisement:2 ito:1 admissible:1 down:2 specific:1 cortes:2 incorporating:1 workshop:1 mnist:2 vapnik:2 positi:1 mage:1 likely:1 contained:2 aa:1 rbf:1 shared:1 content:1 baluja:9 upright:27 ilcnn:1 principal:1 pas:2 la:1 vote:5 support:1 latter:1 rotate:2 frontal:2 tested:3 |
545 | 1,499 | VLSI Implementation of Motion Centroid
Localization for Autonomous Navigation
Ralph Etienne-Cummings
Dept. of ECE,
Johns Hopkins University,
Baltimore, MD
Viktor Gruev
Dept. of ECE,
Johns Hopkins University,
Baltimore, MD
Mohammed Abdel Ghani
Dept. ofEE,
S. Illinois University,
Carbondale, IL
Abstract
A circuit for fast, compact and low-power focal-plane motion centroid
localization is presented. This chip, which uses mixed signal CMOS
components to implement photodetection, edge detection, ON-set
detection and centroid localization, models the retina and superior
colliculus. The centroid localization circuit uses time-windowed
asynchronously triggered row and column address events and two
linear resistive grids to provide the analog coordinates of the motion
centroid. This VLSI chip is used to realize fast lightweight
autonavigating vehicles. The obstacle avoiding line-following
algorithm is discussed.
1 INTRODUCTION
Many neuromorphic chips which mimic the analog and parallel characteristics of visual,
auditory and cortical neural circuits have been designed [Mead, 1989; Koch, 1995] .
Recently researchers have started to combine digital circuits with neuromorphic aVLSI
systems [Boahen, 1996]. The persistent doctrine, however, has been that computation
should be performed in analog, and only communication should use digital circuits. We
have argued that hybrid computational systems are better equipped to handle the high
speed processing required for real-world problem solving , while maintaining
compatibility with the ubiquitous digital computer [Etienne, 1998]. As a further
illustration of this point of view, this paper presents a departure form traditional
approaches for focal plane centroid localization by offering a mixed signal solution that is
simultaneously high-speed, low power and compact. In addition, the chip is interfaced
with an 8-bit microcomputer to implement fast autonomous navigation.
Implementation of centroid localization has been either completely analog or completely
digital. The analog implementations, realized in the early 1990s, used focal plane current
mode circuits to find a global continuos time centroid of the pixels' intensities
[DeWeerth, 1992]. Due to their sub-threshold operation, these circuits are low power,
but slow. On the other hand, the digital solutions do not compute the centroid at the focal
R. Etienne-Cummings, V. Grnev and M A. Ghani
686
plane. They use standard CCO cameras, AID converters and OSP/CPU to compute the
intensity centroid [Mansfield, 1996]. These software approaches offer multiple centroid
localization with complex mathematical processing. However, they suffer from the usual
high power consumption and non-scalability of traditional digital visual processing
systems. Our approach is novel in many aspects. We benefit from the low power,
compactness and parallel organization of focal plane analog circuits and the speed,
robustness and standard architecture of asynchronous digital circuits. Furthermore, it
uses event triggered analog address read-out, which is ideal for the visual centroid
localization problem. Moreover, our chip responds to moving targets only by using the
ON-set of each pixel in the centroid computation. Lastly, our chip models the retina and
two dimensional saccade motor error maps of superior colliculus on a single chip
[Sparks, 1990] . Subsequently, this chip is interfaced with a IlC for autonomous obstacle
avoidance during line-following navigation. The line-following task is similar to target
tracking using the saccadic system, except that the "eye" is fixed and the "head" (the
vehicle) moves to maintain fixation on the target. Control signals provided to the vehicle
based on decisions made by the IlC are used for steering and accelerating/braking. Here
the computational flexibility and programmability of the IlC allows rapid prototyping of
complex and robust algorithms.
2 CENTROID LOCALIZATION
The mathematical computation of the centroid of an object on the focal plane uses
intensity weighted average of the position of the pixels forming the object [OeWeerth,
1992] . Equation (1) shows this formulation. The implementation of this representation
N
N
LI,x,
x=
j =1
N
L,I,
LI,y;
and
y=
;=1
N
(1)
L I,
,=1
1=1
can be quite involved since a product between the intensity and position is implied. To
eliminate this requirement, the intensity of the pixels can be normalized to a single value
within the object. This gives equation (2) since the intensity can be factored out of the
summations. Normalization of the intensity using a simple threshold is not advised since
Ix,
x=~
N
Iy,
and
y=~
N
(2)
Intensity Image
XI
X l+l
X 1+2
X I +3
x 1+4
Edges from pixels
Figure 1: Centroid computation
architecture.
Figure 2: Centroid computation method.
the value of the threshold is dependent on the brightness of the image and number of
pixels forming the object may be altered by the thresholding process. To circumvent
these problems, we take the view that the centroid of the object is defined in relation to its
boundaries. This implies that edge detection (second order spatial derivative of intensity)
can be used to highlight the boundaries, and edge labeling (the zero-crossing of the
edges) can be used to normalize the magnitude of the edges. Subsequently, the centroid
VLSI Implementation ofMotion Centroid Localization for Autonomous Navigation
687
of the zero-crossings is computed. Equation (2) is then realized by projecting the zerocrossing image onto the x- and y-axis and performing two linear centroid determinations.
Figure (1) shows this process.
The determination of the centroid is computed using a resistance grid to associate the
position of a column (row) with a voltage. In figure 2, the positions are given by the
voltages Vi . By activating the column (row) switch when a pixel of the edge image
appears in that column (row), the position voltage is connected to the output line through
the switch impedance, Rs. As more switches are activated, the voltage on the output line
approximates equation (2). Clearly, since no buffers are used to isolate the position
voltages, as more switches are activated, the position voltages will also change. This
does not pose a problem since the switch resistors are design to be larger than the position
resistors (the switch currents are small compared to the grid current). Equation (3) gives
the error between the ideal centroid and the switch loaded centroid in the worst case
when Rs
= on.
In the equation, N is the number of nodes, M is the number of switches
set and Xl and xM are the locations of the first and last set switches, respectively. This
error is improved as Rs gets larger, and vanishes as N
(M~N)
approaches infinity . The
terms Xi represent an ascending ordered list of the activated switches; x I may correspond
to column five, for example. This circuit is compact since it uses only a simple linear
resistive grid and MOS switches. It is low power because the total grid resistance, N x R,
can be large. It can be fast when the parasitic capacitors are kept small . It provides an
analog position value, but it is triggered by fast digital signals that activate the switches.
error = VmOll -Vmin
M(N + 1)
1
~[ X
?...J
.=1
?
xCN+l)
_ _--'-1-'--_-'--_
N + 1 + XI - Xm
(3)
3 MODELING THE RETINA AND SUPERIOR COLLICULUS
3.1 System Overview
The centroid computation approach presented in section 2 is used to isolate the location
of moving targets on a 20 focal plane array. Consequently, a chip which realizes a
neuromorphic visual target acquisition system based on the saccadic generation
mechanism of primates can be implemented. The biological saccade generation process is
mediated by the superior colliculus, which contains a map of the visual field [Sparks,
1990}. In laboratory experiments, cellular recordings suggest that the superior colliculus
provides the spatial location of targets to be foveated. Clearly, a great deal of neural
circuitry exists between the superior colliculus and the eye muscle. Horiuchi has built an
analog system which replicates most of the neural circuits (including the motor system)
which are believed to form the saccadic system [Horiuchi, 1996]. Staying true to the
anatomy forced his implementation to be a complex multi-chip system with many control
parameters. On the other hand, our approach focuses on realizing a compact single chip
solution by only mimicking the behavior of the saccadic system, but not its structure.
3.2 Hardware Implementation
Our approach uses a combination of analog and digital circuits to implement the
functions of the retina and superior colliculus at the focal plane. We use simple digital
control ideas, such as pulse-width modulation and stepper motors, to position the "eye".
The retina portion of this chip uses photodiodes, logarithmic compression, edge detection
and zero-crossing circuits. These circuits mimic the first three layers of cells in the retina
688
R. Etienne-Cummings, V. Grnev and M. A. Ghani
with mixed sub-threshold and strong inversion circuits. The edge detection circuit is
realized with an approximation of the Laplacian operator implemented using the
difference between a smooth (with a resistive grid) and unsmoothed version of the image
[Mead, 1989]. The high gain of the difference circuit creates a binary image of
approximate zero-crossings. After this point, the computation is performed using mixed
analog/digital circuits. The zero-crossings are fed to ON-set detectors (positive temporal
derivatives) which signal the location of moving or flashing targets. These circuits model
the behavior of some of the amacrine and ganglion cells of the primate retina [Barlow,
1982]. These first layers of processing constitute all the "direct" mimicry of the
biological models. Figure 3 shows the schematic of these early processing layers.
The ON-set detectors provide inputs to the model of the superior colliculus circuits. The
ON-set detectors allow us to segment moving targets against textured backgrounds. This
is an improvement on earlier centroid and saccade chips that used pixel intensity. The
essence of the superior colliculus map is to locate the target that is to be foveated. In our
case, the target chosen to be foveated will be moving. Here motion is define simply as
the change in contrast over time. Motion, in this sense, can be seen as being the earliest
measurable attribute of the target which can trigger a saccade without requiring any high
level decision making. Subsequently, the coordinates of the motion must be extracted and
provided to the motor drivers.
X M otion Cenuold
~!
~A
!
Edge Detc:ctlon
ON-set Detecu on
Figure 3: Schematic of
the model of the retina.
Figure 4: Schematic of the model of the superior
collicu Ius.
The circuits for locating the target are implemented entirely with mixed signal, nonneuromorphic circuits. The theoretical foundation for our approach is presented in
section 2. The ON-set detector is triggered when an edge of the target appears at a pixel.
At this time, the pixel broadcasts its location to the edge of the array by activating row
and column lines. This row (column) signal sets a latch at the right (top) of the array. The
latches asynchronously activate switches and the centroid of the activated positions is
provided. The latches remain set until they are cleared by an external control signal. This
control signal provides a time-window over which the centroid output is integrated. This
has the effect of reducing noise by combining the outputs of pixels which are activated at
different instances even if they are triggered by the same motion (an artifact of small fill
factor focal plane image processing). Furthermore, the latches can be masked from the
pixels' output with a second control signal. This signal is used to de-activate the centroid
689
VLSI Implementation of Motion Centroid Localization for Autonomous Navigation
circuit during a saccade (saccadic suppression). A centroid valid signal is also generated
by the chip. Figure 4 shows a portion of the schematic of the superior colliculus model.
3.3 Results
In contrast to previous work, this chip provides the 2-D coordinates of the centroid of a
moving target. Figure 5 shows the oscilloscope trace of the coordinates as a target moves
back and forth, in and out of the chip's field of view. The y-coordinate does change
while the x-coordinate increases and decreases as the target moves to the left and right,
respectively. The chip has been used to track targets in 2-D by making micro-saccades .
In this case, the chip chases the target as it attempts to escape from the center. The eye
movement is performed by converting the analog coordinates into PWM signals, which
are used to drive stepper motors. The system performance is limited by the contrast
sensitivity of the edge detection circuit, and the frequency response of the edge (high
frequency cut-off) and ON-set (low frequency cut-off) detectors. With the appropriate
optics, it can track walking or running persons under indoor or outdoor lighting
conditions at close or far distances. Table I gives a summary of the chip characteristics.
VarYing x?
l'Oordma te
Figure 5: Oscilloscope trace of 20
centroid for a moving target.
Technology
1.2um ORBIT
Chip Size
Array Size
4mm 1
12 x 10
Pixel Size
Fill Factor
llOxllOutn
11%
Intensity
Min . Contrast
0.lu-1OOmW/cm 2
Response Time
Power (chip)
2-10 6 Hz(@1 mW/cml)
10%
5 mW (@l m W/cm~ Vdd
=6V)
Table I: Chip characteristics.
4 APPLICATION: OBSTACLE AVOIDANCE DURING LINEFOLLOWING AUTONA VIGATION
4.1 System Overview
The frenzy of activity towards developing neuromorphic systems over the pass 10 years
has been mainly driven by the promise that one day engineers will develop machines that
can interact with the environment in a similar way as biological organisms. The prospect
of having a robot that can help humans in their daily tasks has been a dream of science
fiction for many decades. As can be expected, the key to success is premised on the
development of compact systems, with large computational capabilities, at low cost (in
terms of hardware and power) . Neuromorphic VLSI systems have closed the gap between
dreams and reality, but we are still very far from the android robot. For all these robots,
autonomous behavior, in the form of auto-navigation in natural environments, must be
one of their primary skills. For miniaturization, neuromorphic vision systems performing
most of the pre-processing, can be coupled with small fast computers to realize these
compact yet powerful sensor/processor modules.
4.2 Navigation Algorithm
The simplest form of data driven auto-navigation is the line-following task. In this task,
the robot must maintain a certain relationship with some visual cues that guide its motion.
In the case of the line-follower, the visual system provides data regarding the state of the
R. Etienne-Cummings, V. Gruev and M A. Ghani
690
line relative to the vehicle, which results in controlling steering and/or speed. If obstacle
avoidance is also required, auto-navigation is considerably more difficult. Our system
handles line-following and obstacle avoidance by using two neuromorphic visual sensors
that provide information to a micro-controller OlC) to steer, accelerate or decelerate the
vehicle. The sensors, which uses the centroid location system outlined above, provides
information on the position of the line and obstacles to the p,C, which provides PWM
signals to the servos for controlling the vehicle. The algorithm implemented in the p,C
places the two sensors in competition with each other to force the line into a blind zone
between the sensors. Simultaneously, if an object enters the visual field from outside, it
is treated as an obstacle and the p,C turns the car away from the object. Obstacle
avoidance is given higher priority than line-following to avoid collisions. The p,C also
keeps track of the direction of avoidance such that the vehicle can be re-oriented towards
the line after the obstacle is pushed out of the field of view. Lastly, for line following,
the position, orientation and velocity of drift, determined from the temporal derivative of
the centroid, are used to track the line. The control strategy is to keep the line in the blind
zone, while slowing down at corners, speeding up on straight aways and avoiding
obstacles. The angle which the line or obstacle form with the x-axis also affects the
speed. The value of the x-centroid relative to the y-centroid provides rudimentary
estimate of the orientation of the line or obstacle to the vehicle. For example, angles less
.
Follow
"
AV~8id~nce
...
,
0I0s1ade
L Zone
/
!
~ne
i
'.)", /
:
AV~na;ce
!
~\?': ~../
;i:~~~~???. .\j ~
\;... ?? ? ?~i=~s..s
Figure 6: Block diagram of the
autonomous line-follower system.
Figure 7: A picture of the vehicle.
(greater) than +/- 45 degrees tend to have small (large) x-coordinates and large (small) ycoordinates and require deceleration (acceleration). Figure 6 shows the organization of
the sensors on the vehicle and control spatial zones. Figure 7 shows the vehicle and
samples of the line and obstacles.
4.3 Hardware Implementation
The coordinates from the centroid localization circuits are presented to the p,C for
analysis. The p,C used is the Microchip PIC16C74. This chip is chosen because of its
five NO inputs and three PWM outputs. The analog coordinates are presented directly to
the NO inputs. Two of the PWM outputs are connected to the steering and speed control
servos. The PIC16C74 runs at 20 MHz and has 35 instructions, 4K by 8-b ROM and 80
by 20-b RAM. The program which runs on the PIC determines the control action to take,
based on the signal provided by the neuromorphic visual sensors. The vehicle used is a
four-wheel drive radio controlled model car (the radio receiver is disconnected) with
Digital Proportional Steering (DPS) .
VLSI Implementation ofMotion Centroid Localization for Autonomous Navigation
691
4.4 Results
The vehicle was tested on a track composed of black tape on a gray linoleum floor with
black and white obstacles. The track formed a closed loop with two sharp turns and some
smooth S-curves. The neuromorphic vision chip was equipped with a 12.5 mm variable
iris lens, which limited its field of view to about 100. Despite the narrow field of view ,
the car was able to navigate the track at an average speed of 1 mls without making any
errors. On less curvy parts of the track, it accelerated to about 2 mls and slowed down at
the corners. When the speed of the vehicle is scaled up, the errors made are mainly due
to over steering.
5 CONCLUSION
A 2D model of the saccade generating components of the superior colliculus is presented .
This model only mimics the functionality the saccadic system using mixed signal focal
plane circuits that realize motion centroid localization. The single chip combines a
silicon retina with the superior colliculus model using compact, low power and fast
circuits. Finally, the centroid chip is interfaced with an 8-bit IlC and vehicle for fast linefollowing auto navigation with obstacle avoidance. Here all of the required computation is
performed at the visual sensor, and a standard IlC is the high-level decision maker.
References
Barlow H., The Senses: Physiology of the Retina, Cambridge University Press,
Cambridge, England, 1982.
Boahen K., "Retinomorphic Vision Systems II: Communication Channel Design," ISCAS
96, Atlanta, GA, 1996.
DeWeerth, S. P., "Analog VLSI Circuits for Stimulus Localization and Centroid
Computation," Int'l 1. Computer Vision, Vol. 8, No.2, pp. 191-202, 1992.
Etienne-Cummings R., J Van der Spiegel and P. Mueller, "Neuromorphic and Digital
Hybrid Systems," Neuromorphic Systems: Engineering Silicon from Neurobiology, L.
Smith and A. Hamilton (Eds.), World Scientific, 1998.
Horiuchi T., T. Morris, C . Koch and S. P . DeWeerth, "Analog VLSI Circuits for
Attention-Based Visual Tracking," Advances in Neural Information Processing
Systems, Vol. 9, Denver, CO, 1996.
Koch C. and H. Li (Eds.), Vision Chips: Implementing Vision Algorithms with Analog
VLSI Circuits, IEEE Computer Press, 1995.
Mansfield, P., "Machine Vision Tackles Star Tracking," Laser Focus World, Vol. 30, No.
26, pp. S21 -S24, 1996.
Mead C. and M. Ismail (Eds.), Analog VLSI Implementation of Neural Networks, Kluwer
Academic Press, Newell, MA, 1989.
Sparks D., C. Lee and W. Rohrer, "Population Coding of the Direction, Amplitude and
Velocity of Saccadic Eye Movements by Neurons in the Superior Colliculus," Proc.
Cold Spring Harbor Symp. Quantitative Biology, Vol. LV, 1990.
| 1499 |@word version:1 inversion:1 compression:1 cco:1 instruction:1 r:3 pulse:1 cml:1 brightness:1 lightweight:1 contains:1 offering:1 detc:1 cleared:1 current:3 yet:1 follower:2 must:3 john:2 realize:3 s21:1 motor:5 designed:1 cue:1 vmin:1 slowing:1 plane:10 realizing:1 smith:1 provides:8 node:1 location:6 five:2 windowed:1 mathematical:2 direct:1 driver:1 rohrer:1 persistent:1 microchip:1 resistive:3 fixation:1 combine:2 symp:1 expected:1 rapid:1 behavior:3 oscilloscope:2 multi:1 cpu:1 equipped:2 window:1 provided:4 moreover:1 circuit:30 cm:2 microcomputer:1 temporal:2 quantitative:1 tackle:1 um:1 scaled:1 control:10 hamilton:1 positive:1 engineering:1 despite:1 id:1 mead:3 modulation:1 advised:1 black:2 co:1 limited:2 camera:1 block:1 implement:3 cold:1 physiology:1 pre:1 suggest:1 get:1 onto:1 close:1 wheel:1 operator:1 ga:1 measurable:1 map:3 center:1 attention:1 spark:3 factored:1 avoidance:7 array:4 fill:2 his:1 population:1 handle:2 autonomous:8 coordinate:10 deceleration:1 target:19 trigger:1 controlling:2 us:8 associate:1 crossing:5 velocity:2 walking:1 cut:2 module:1 enters:1 worst:1 ycoordinates:1 connected:2 decrease:1 movement:2 prospect:1 servo:2 boahen:2 vanishes:1 environment:2 vdd:1 solving:1 segment:1 localization:15 creates:1 completely:2 textured:1 accelerate:1 chip:27 laser:1 horiuchi:3 fast:8 forced:1 activate:3 labeling:1 outside:1 doctrine:1 quite:1 larger:2 spiegel:1 asynchronously:2 chase:1 triggered:5 product:1 combining:1 loop:1 flexibility:1 forth:1 ismail:1 normalize:1 scalability:1 competition:1 requirement:1 generating:1 cmos:1 staying:1 object:7 help:1 avlsi:1 develop:1 pose:1 strong:1 implemented:4 implies:1 direction:2 anatomy:1 attribute:1 functionality:1 subsequently:3 human:1 implementing:1 argued:1 require:1 activating:2 biological:3 summation:1 mm:2 koch:3 great:1 mo:1 circuitry:1 early:2 proc:1 realizes:1 radio:2 maker:1 weighted:1 clearly:2 sensor:8 amacrine:1 avoid:1 varying:1 voltage:6 earliest:1 focus:2 improvement:1 mainly:2 contrast:4 centroid:42 suppression:1 sense:1 mueller:1 dependent:1 eliminate:1 integrated:1 compactness:1 vlsi:10 relation:1 zerocrossing:1 ralph:1 compatibility:1 pixel:13 mimicking:1 orientation:2 development:1 retinomorphic:1 spatial:3 field:6 having:1 biology:1 mimic:3 stimulus:1 micro:2 escape:1 retina:10 oriented:1 composed:1 simultaneously:2 iscas:1 maintain:2 attempt:1 detection:6 organization:2 atlanta:1 replicates:1 navigation:11 activated:5 sens:1 programmability:1 edge:14 daily:1 carbondale:1 orbit:1 re:1 stepper:2 theoretical:1 android:1 instance:1 column:7 modeling:1 obstacle:15 earlier:1 steer:1 mhz:1 neuromorphic:11 cost:1 masked:1 considerably:1 person:1 sensitivity:1 lee:1 off:2 iy:1 hopkins:2 na:1 s24:1 broadcast:1 priority:1 external:1 corner:2 derivative:3 li:3 premised:1 de:1 star:1 coding:1 int:1 vi:1 blind:2 vehicle:15 performed:4 view:6 closed:2 portion:2 parallel:2 capability:1 il:1 formed:1 loaded:1 characteristic:3 interfaced:3 correspond:1 lu:1 lighting:1 researcher:1 drive:2 straight:1 processor:1 detector:5 ed:3 against:1 acquisition:1 frequency:3 involved:1 pp:2 gain:1 auditory:1 car:3 ubiquitous:1 amplitude:1 back:1 appears:2 higher:1 cummings:5 day:1 follow:1 response:2 improved:1 formulation:1 furthermore:2 lastly:2 deweerth:3 until:1 hand:2 mode:1 artifact:1 gray:1 scientific:1 effect:1 normalized:1 true:1 barlow:2 requiring:1 read:1 laboratory:1 deal:1 white:1 latch:4 during:3 width:1 essence:1 iris:1 motion:10 ilc:5 rudimentary:1 image:7 novel:1 recently:1 superior:14 denver:1 overview:2 viktor:1 analog:17 discussed:1 approximates:1 braking:1 organism:1 kluwer:1 silicon:2 cambridge:2 focal:10 grid:6 outlined:1 illinois:1 moving:7 robot:4 driven:2 mohammed:1 buffer:1 certain:1 binary:1 success:1 der:1 muscle:1 seen:1 greater:1 steering:5 floor:1 converting:1 signal:16 ii:1 multiple:1 photodiodes:1 smooth:2 determination:2 england:1 offer:1 believed:1 academic:1 laplacian:1 schematic:4 controlled:1 mansfield:2 controller:1 vision:7 normalization:1 represent:1 cell:2 addition:1 background:1 baltimore:2 diagram:1 isolate:2 recording:1 hz:1 tend:1 capacitor:1 mw:2 ideal:2 switch:13 affect:1 harbor:1 architecture:2 converter:1 idea:1 regarding:1 accelerating:1 suffer:1 locating:1 resistance:2 constitute:1 action:1 tape:1 collision:1 morris:1 hardware:3 simplest:1 fiction:1 track:8 promise:1 vol:4 key:1 four:1 threshold:4 ce:1 kept:1 ram:1 year:1 colliculus:13 nce:1 run:2 angle:2 powerful:1 place:1 decision:3 bit:2 entirely:1 layer:3 pushed:1 activity:1 optic:1 infinity:1 software:1 aspect:1 speed:8 min:1 unsmoothed:1 spring:1 performing:2 developing:1 combination:1 disconnected:1 remain:1 aways:1 primate:2 making:3 continuo:1 projecting:1 slowed:1 equation:6 turn:2 mechanism:1 ascending:1 fed:1 operation:1 decelerate:1 away:1 appropriate:1 robustness:1 top:1 pwm:4 running:1 maintaining:1 etienne:6 implied:1 move:3 realized:3 strategy:1 saccadic:7 primary:1 md:2 traditional:2 usual:1 responds:1 dp:1 distance:1 consumption:1 cellular:1 dream:2 rom:1 relationship:1 illustration:1 difficult:1 trace:2 implementation:11 design:2 av:2 neuron:1 neurobiology:1 communication:2 head:1 locate:1 sharp:1 intensity:11 drift:1 pic:1 required:3 narrow:1 address:2 able:1 prototyping:1 departure:1 xm:2 indoor:1 program:1 built:1 including:1 power:9 event:2 natural:1 hybrid:2 circumvent:1 force:1 treated:1 altered:1 technology:1 eye:5 ne:1 picture:1 axis:2 started:1 mediated:1 auto:4 coupled:1 speeding:1 relative:2 highlight:1 mixed:6 generation:2 proportional:1 lv:1 abdel:1 digital:13 foundation:1 degree:1 thresholding:1 row:6 summary:1 last:1 asynchronous:1 guide:1 allow:1 benefit:1 van:1 boundary:2 curve:1 cortical:1 world:3 valid:1 made:2 far:2 approximate:1 compact:7 skill:1 keep:2 ml:2 global:1 receiver:1 xi:3 decade:1 table:2 impedance:1 reality:1 channel:1 robust:1 interact:1 complex:3 noise:1 ghani:4 slow:1 aid:1 sub:2 position:13 resistor:2 xl:1 outdoor:1 ix:1 down:2 navigate:1 list:1 exists:1 flashing:1 magnitude:1 te:1 foveated:3 gap:1 logarithmic:1 simply:1 forming:2 ganglion:1 visual:12 ordered:1 tracking:3 saccade:7 newell:1 determines:1 extracted:1 ma:1 consequently:1 acceleration:1 towards:2 change:3 determined:1 except:1 reducing:1 engineer:1 total:1 lens:1 pas:1 ece:2 zone:4 parasitic:1 accelerated:1 dept:3 tested:1 avoiding:2 |
546 | 15 | 270
Correlational Strength and Computational Algebra
of Synaptic Connections Between Neurons
Eberhard E. Fetz
Department of Physiology & Biophysics,
University of Washington, Seattle, WA 98195
ABSTRACT
Intracellular recordings in spinal cord motoneurons and cerebral
cortex neurons have provided new evidence on the correlational strength of
monosynaptic connections, and the relation between the shapes of
postsynaptic potentials and the associated increased firing probability. In
these cells, excitatory postsynaptic potentials (EPSPs) produce crosscorrelogram peaks which resemble in large part the derivative of the EPSP.
Additional synaptic noise broadens the peak, but the peak area -- i.e., the
number of above-chance firings triggered per EPSP -- remains proportional to
the EPSP amplitude. A typical EPSP of 100 ~v triggers about .01 firings per
EPSP. The consequences of these data for information processing by
polysynaptic connections is discussed. The effects of sequential polysynaptic
links can be calculated by convolving the effects of the underlying
monosynaptic connections. The net effect of parallel pathways is the sum of
the individual contributions.
INTRODUCTION
Interactions between neurons are determined by the strength and
distribution of their synaptic connections.
The strength of synaptic
interactions has been measured directly in the central nervous system by two
techniques. Intracellular recording reveals the magnitude and time course of
postsynaptic potentials (PSPs) produced by synaptic connections, and crosscorrelation of extracellular spike trains measures the effect of the PSP's on the
firing probability of the connected cells. The relation between the shape of
excitatory postsynaptic potentials (EPSPs) and the shape of the crosscorrelogram peak they produce has been empirically investigated in cat
motoneurons 2,4,5 and in neocortical cells 10.
RELATION BETWEEN EPSP'S AND CORRELOGRAM PEAKS
Synaptic interactions have been studied most thoroughly in spinal
cord motoneurons. Figure 1 illustrates the membrane potential of a
rhythmically firing motoneuron, and the effect of EPSPs on its firing. An
EPSP occurring sufficiently close to threshold (8) will cause the motoneuron
to fire and will advance an action potential to its rising edge (top).
Mathematical analysis of this threshold-crossing process predicts that an
EPSP with shape e(t) will produce a firing probability f(t), which resembles
? American Institute of Phy~ics 1988
271
rI
f;
::
'I
8
'I
I'
/..-::-- ----.... ....
.",.""..,.",
.,...,.""
~-"""
/'
I
..,, .... ..
)
\
...",/
,;
I
i:
.. :
\
--.----r
,
'I
.,.,,,,
....
....
EPSP
e(t)
t
CROSS-
CORRELOGRAM
f(t)
TIME
t
Fig. 1. The relation between EPSP's and motoneuron firing. Top: membrane trajectory of
rhythmically firing motoneuron, showing EPSP crossing threshold (8) and shortening the
normal interspike interval by advancing a spike. V(t) is difference between membrane
potential and threshold. Middle: same threshold-crossing process aligned with EPSP, with
v(t) plotted as falling trajectory. Intercept (at upward arrow) indicates time of the advanced
action potential. Bottom: Cross-correlation histogram predicted by threshold crossings. The
peak in the firing rate f(t) above baseline (fo) is produced by spikes advanced from baseline,
as indicated by the changed counts for the illustrated trajectory. Consequently, the area in
the peak equals the area of the subsequent trough.
272
the derivative of the EPSP 4,8. Specifically, for smooth membrane potential
trajectories approaching threshold (the case of no additional synaptic noise):
f(t)
=fo + (fo/v) del dt
(1)
v
where fo is the baseline firing rate of the motoneuron and is the rate of
closure between motoneuron membrane potential and threshold. This
relation can be derived analytically by tranforming the process to a
coordinate system aligned with the EPSP (Fig. 1, middle) and calculating the
relative timing of spikes advanced by intercepts of the threshold trajectories
with the EPSP 4. The above relation (1) is also valid for the correlogram
trough during the falling phase of the EPSP, as long as del dt >
if the EPSP
falls more rapidly than
the trough is limited at zero firing rate (as
illustrated for the correlogram at bottom). The fact that the shape of the
correlogram peak above baseline matches the EPSP derivative has been
empirically confirmed for large EPSPs in cat motoneurons 4. This relation
implies that the height of the correlogram peak above baseline is proportional
to the EPSP rate of rise. The integral of this relationship predicts that the area
between the correlogram peak and baseline is proportional to the EPSP
amplitude.
This linear relation further implies that the effects of
simultaneously arriving EPSPs will add linearly.
The presence of additional background synaptic "noise", which is
normally produced by randomly occurring synaptic inputs, tends to make the
correlogram peak broader than the duration of the EPSP risetime. This
broadening is produced by membrane potential fluctuations which cause
additional threshold crossings during the decay of the EPSP by trajectories
that would have missed the EPSP (e.g., the dashed trajectory in Fig. 1,
middle). On the basis of indirect empirical comparisons it has been proposed
6,7 that the broader correlogram peaks can be described by the sum of two
linear functions of e(t):
-v,
f(t)
=fo + a e(t) + b deldt
-v;
(2)
This relation provides a reasonable match when the coefficients (a and b) can
be optimized for each case 5,7, but direct empirical comparisons 2,4 indicate
that the difference between the correlogram peak and the derivative is
typically briefer than the EPSP.
The effect of synaptic noise on the transform -between EPSP and
correlogram peak has not yet been analytically derived (except for the case of
However the threshold-crossing process has been
Gaussian noise1).
simulated by a computer model which adds synaptic noise to the trajectories
intercepting the EPSP 1. The correlograms generated by the simulation match
the correlograms measured empirically for small EPSP's in motoneurons 2,
confirming the validity of the model.
Although synaptic noise distributes the triggered firings over a wider
peak, the area of the correlogram peak, i.e., the number of motoneuron firings
produced by an EPSP, is essentially preserved and remains proportional to
EPSP amplitude for moderate noise levels. For unitary EPSP's (produced by
273
a single afferent fiber) in cat motoneurons, the number of firings triggered per
EPSP (Np) was linearly related to the amplitude (h) of the EPSP 2:
Np = (O.l/mv)? h (mv) + .003
(3)
The fact that the number of triggered spikes increases in proportion to EPSP
amplitude has also been confirmed for neocortical neurons 10; for cells
recorded in sensorimotor cortex slices (probably pyramidal cells) the
coefficient of h was very similar: 0.07/mv. This means that a typical unitary
EPSP with amplitude of 100 Ilv, raises the probability that the postsynaptic
cell fires by less than .01. Moreover, this increase occurs during a specific
time interval corresponding to the rise time of the EPSP - on the order of 1 - 2
msec. The net increase in firing rate of the postsynaptic cell is calculated by
the proportional decrease in interspike intervals produced by the triggered
spikes 4. (While the above values are typical, unitary EPSP's range in size
from several hundred IlV down to undetectable levels of severalllv., and
have risetimes of.2 - 4 msec.)
Inhibitory connections between cells, mediated by inhibitory
postsynaptic potentials (IPSPs), produce a trough in the cross-correlogram.
This reduction of firing probability below baseline is followed by a
subsequent broad, shallow peak, representing the spikes that have been
delayed during the IPSP. Although the effects of inhibitory connections
remain to be analyzed more quantitatively, preliminary results indicate that
small IPSP's in synaptic noise produce decreases in firing probability that are
similar to the increases produced by EPSP's 4,5.
DISYNAPTIC LINKS
The effects of polysynaptic links between neurons can be understood
as combinations of the underlying monosynaptic connections.
A
monosynaptic connection from cell A to cell B would produce a first-order
cross-correlation peak P1(BIA,t), representing the conditional probability that
neuron B fires above chance at time t, given a spike in cell A at time t = O. As
noted above, the shape of this first-order correlogram peak is largely
proportional to the EPSP derivative (for cells whose interspike interval
exceeds the duration of the EPSP). The latency of the peak is the conduction
time from A to B (Fig. 2 top left).
In contrast, several types of disynaptic linkages betw.een A and B,
mediated by a third neuron C, will produce a second-order correlation peak
between A and B. A disynaptic link may be produced by two serial
monosynaptic connections, from A to C and from C to B (Fig. 2, bottom left),
or by a common synaptic input from C ending on both A and B (Fig. 2,
bottom right). In both cases, the second-order correlation between A and B
produced by the disynaptic link would be the convolution of the two firstorder correlations between the monosynaptically connected cells:
(4)
274
As indicated by the diagram, the cross-correlogram peak P2(BIA,t) would be
smaller and more dispersed than the peaks of the underlying first-order
correlation peaks. For serial connections the peak would appear to the right
of the origin, at a latency that is the sum of the two monosynaptic latencies.
The peak produced by a common input typically straddles the origin, since its
timing reflects the difference between the underlying latencies.
=>
Monosynaptic connection
-----..'t-
I \
t \
@
First-order correlation
~(AIB,t)
LJA,,-_~_(_B_I_A_'t_)_
Disynaptic connection
=
~(~IA,-t)
1
~
Serial connection
Second-order correlation
Common input
r--A---t
t
I
t \
I \
: \. A
~ (C I A)
@
t
t
t
I
:
ll(AIC)
:"
V
H\~
t
:
j
\
I'"
t \
\
\
\
\
t
\~(BIC)
P(BIA)
_________ ~,_2__----
@
\.
P(BIC)
1 \/\ '
J t"-_------
_ _ _/\.~(BIA)
-L
Fig. 2. Correlational effects of monosynaptic and disynaptic links between two neurons.
Top: monosynaptic excitatory link from A to B produces an increase in firing probability of B
after A (left). As with all correlograms this is the time-inverted probability of increased firing
in A relative to B (right). Bottom: Two common disynaptic links between A and B are a
serial connection via C (left) and a common input from C. In both cases the effect of the
disynaptic link is the convolution of the underlying monosynaptic links.
275
This relation means that the probability that a spike in cell A will
produce a correlated spike in cell B would be the product of the two
probabilities for the intervening monosynaptic connections. Given a typical
Np of .Ol/EPSP, this would reduce the effectiveness of a given disynaptic
linkage by two orders of magnitude relative to a monosynaptic connection.
However, the net strength of all the disynaptic linkages between two given
cells is proportional to the number of mediating intemeurons (C}, since the
effects of parallel pathways add. Thus, the net potency of all the disynaptic
linkages between two cells could approach that of a monosynaptic linkage if
the number of mediating interneurons were sufficiently large. It should also
be noted that some intemeurons may fire more than once per EPSP and have
a higher probability of being triggered to fire than motoneurons 11.
For completeness, two other possible disynaptic links between A and B
involving a third cell C may be considered. One is a serial connection from B
to C to A, which is the reverse of the serial connection from A to B. This
would produce a P2(BIA) with peak to the left of the origin. The fourth
circuit involves convergent connections from both A and B to C; this is the
only combination that would not produce any causal link between A and B.
The effects of still higher-order polysynaptic linkages can be computed
similarly, by convolving the effects produced by the sequential connections.
For example, trisynaptic linkages between four neurons are equivalent to
combinations of disynaptic and monosynaptic connections.
The cross-correlograms between two cells have a certain symmetry,
depending on which is the reference cell. The cross-correlation histogram of
cell B referenced to A is identical to the time-inverted correlogram of A
referenced to B. This is illustrated for the monosynaptic connection in Fig.2,
top right, but is true for all correlograms. This symmetry represents the fact
that the above-chance probability of B firing after A is the same as the
probability of A firing before B:
P(BIA, t)
= P(AIB, -t)
(5)
As a consequence, polysynaptic correlational links can be computed as the
same convolution integral (Eq. 4), independent of the direction of impulse
propagation.
P ARALLEL PATHS AND FEEDBACK LOOPS
In addition to the simple combinations of pair-wise connections
between neurons illustrated above, additional connections between the same
cells may form circuits with various kinds of loops. Recurrent connections
can produce feedback loops, whose correlational effects are also calculated by
convolving effects of the underlying synaptic links. Parallel feed-forward
paths can form multiple pathways between the same cells. These produce
correlational effects that are the sum of the effects of the individual
underlying connections.
The simplest feedback loop is formed by reciprocal connections
between a pair of cells. The effects of excitatory feedback can be computed by
276
successive cO?1volutions of the underlying monosynaptic connections (Fig. 3
top). Note that such a positive feedback loop would be capable of sustaining
activity only if the connections were sufficiently potent to ensure
postsynaptic firing. Since the probabilities of triggered firings at a single
synapse are considerably less than one, reverberating activity can be
sustained only if the number of interacting cells is correspondingly increased.
Thus, if the probability for a single link is on the order of .01, reverberating
activity can be sustained if A and B are similarly interconnected with at least
a hundred cells in parallel.
Connections between three neurons may produce various kinds of
loops.
Feedforward parallel pathways are formed when cell A is
monosynaptically connected to B and in addition has a serial disynaptic
connection through C, as illustrated in Fig. 3 (bottom left); the correlational
effects of the two linkages from A to B would sum linearly, as shown for
excitatory connections. Again, the effect of a larger set of cells {C} would be
additive. Feedback loops could be formed with three cells by recurrent
connections between any pair; the correlational consequences of the loop
again are the convolution of the underlying links. Three cells can form
another type loop if both A and B are monosynaptically connected, and
simultaneously influenced by a common interneuron C (Fig. 3 bottom right).
In this case the expected correlogram between A and B would be the sum of
the individual components -- a common input peak around the origin plus a
delayed peak produced by the serial connection.
Feedback loop
1 - - - - -..
'l;---~
...
..../
-',
\....
.... ....
..
I:
'
....
.......
..........."
;'"
........
..?.:....
....
-',
....
?????
'"
???.
Parallel jeedfOrward path
I
I
:
t
t
t
Common input loop
I
t
PI (BIA) +P 2 (BIA)
PI (BIA)+P
2
(BIA)
:/\ ____
___.;.J
l~
Fig. 3. Correlational effects of parallel connections between two neurons. Top: feedback
loop between two neurons A and B produces higher-order effects equivalent to convolution
of mono~aptic effects. Bottom: Loops formed by parallel feed forward paths (left) and by a
common mput concurrent with a monosynaptic link (right) produce additive effects.
277
CONCLUSIONS
Thus, a simple computational algebra can be used to derive the
correlational effects of a given network structure. Effects of sequential
connections can be computed by convolution and effects of parallel paths by
summation. The inverse problem, of deducing the circuitry from the
correlational data is more difficult, since similar correlogram features may be
produced by different circuits 9.
The fact that monosynaptic links produce small correlational effects on
the order of .01 represents a significant constraint in the mechanisms of
information processing in real neural nets. For example, secure propagation
of activity through serial polysynaptic linkages requires that the small
probability of triggered firing via a given link is compensated by a
proportional increase in the number of parallel links. Thus, reliable serial
conduction would require hundreds of neurons at each level, with
appropriate divergent and convergent connections. It should also be noted
that the effect of intemeurons can be modulated by changing their activity.
The intervening cells need to be active to mediate the correlational effects. As
indicated by eq. I, the size of the correlogram peak is proportional to the
firing rate (fo) of the postsynaptic cell. This allows dynamic modulation of
polysynaptic linkages. The greater the number of links, the more susceptible
they are to modulation.
Acknowledgements: The author thanks Mr. Garrett Kenyon for stimulating
discussions and the cited colleagues for collaborative efforts. This work was
supported in part by Nll-I grants NS 12542 and RR00166.
REFERENCES
1. Bishop, B., Reyes, A.D., and Fetz E.E., Soc. for Neurosci Abst. 11:157 (1985).
2. Cope, T.C., Fetz, E.E., and Matsumura, M., J. Physiol. 390:161-18 (1987).
3. Fetz, E.E. and Cheney, P.D., J. Neurophysiol. 44:751-772 (1980).
4. Fetz, E.E. and Gustafsson, B., J. Physiol. 341:387-410 (1983).
5. Gustafsson, B., and McCrea, D., J. Physiol. 347:431-451 (1984).
6. Kirkwood, P.A., J. Neurosci. Meth. 1:107-132 (1979).
7. Kirkwood, P.A., and Sears, T._ J. Physiol. 275:103-134 (1978).
8. Knox, C.K., Biophys. J. 14: 567-582 (1974).
9. Moore, G.P., Segundo, J.P., Perkel, D.H. and Levitan, H., Biophys. J. 10:876900 (1970).
10. Reyes, A.D., Fetz E.E. and Schwindt, P.C., Soc. for Neurosci Abst. 13:157
(1987).
11. Surmeier, D.J. and Weinberg, R.J., Brain Res. 331:180-184 (1985).
| 15 |@word middle:3 rising:1 proportion:1 closure:1 simulation:1 t_:1 reduction:1 phy:1 yet:1 physiol:4 additive:2 subsequent:2 confirming:1 interspike:3 shape:6 nervous:1 reciprocal:1 completeness:1 provides:1 cheney:1 successive:1 height:1 mathematical:1 correlograms:5 direct:1 undetectable:1 gustafsson:2 sustained:2 pathway:4 expected:1 p1:1 ol:1 perkel:1 brain:1 provided:1 monosynaptic:18 underlying:9 moreover:1 circuit:3 kind:2 firstorder:1 normally:1 grant:1 appear:1 before:1 positive:1 understood:1 timing:2 referenced:2 tends:1 consequence:3 firing:26 fluctuation:1 path:5 modulation:2 plus:1 studied:1 resembles:1 sustaining:1 co:1 limited:1 range:1 area:5 empirical:2 physiology:1 close:1 intercept:2 equivalent:2 compensated:1 duration:2 monosynaptically:3 coordinate:1 trigger:1 origin:4 crossing:6 predicts:2 bottom:8 cord:2 connected:4 decrease:2 dynamic:1 raise:1 algebra:2 volution:1 basis:1 neurophysiol:1 indirect:1 cat:3 fiber:1 various:2 train:1 sears:1 broadens:1 whose:2 larger:1 transform:1 nll:1 triggered:8 abst:2 net:5 interaction:3 product:1 epsp:41 interconnected:1 aligned:2 loop:13 rapidly:1 intervening:2 seattle:1 produce:17 wider:1 depending:1 recurrent:2 derive:1 measured:2 eq:2 p2:2 epsps:5 soc:2 predicted:1 resemble:1 implies:2 indicate:2 involves:1 direction:1 require:1 preliminary:1 summation:1 sufficiently:3 considered:1 ic:1 normal:1 around:1 circuitry:1 ilv:2 concurrent:1 reflects:1 gaussian:1 broader:2 derived:2 indicates:1 contrast:1 secure:1 baseline:7 typically:2 relation:10 upward:1 equal:1 once:1 washington:1 identical:1 represents:2 broad:1 np:3 quantitatively:1 randomly:1 simultaneously:2 individual:3 delayed:2 phase:1 fire:5 interneurons:1 analyzed:1 edge:1 integral:2 capable:1 segundo:1 re:1 plotted:1 causal:1 increased:3 hundred:3 conduction:2 considerably:1 thoroughly:1 thanks:1 cited:1 eberhard:1 peak:30 straddle:1 potent:1 knox:1 intercepting:1 again:2 central:1 recorded:1 convolving:3 derivative:5 american:1 crosscorrelation:1 potential:12 coefficient:2 trough:4 afferent:1 mv:3 parallel:10 contribution:1 collaborative:1 formed:4 largely:1 produced:14 trajectory:8 confirmed:2 fo:6 influenced:1 synaptic:15 sensorimotor:1 disynaptic:14 colleague:1 associated:1 amplitude:6 garrett:1 feed:2 higher:3 dt:2 synapse:1 correlation:9 propagation:2 del:2 indicated:3 impulse:1 effect:31 validity:1 aib:2 true:1 kenyon:1 intemeurons:3 analytically:2 moore:1 illustrated:5 ll:1 during:4 noted:3 neocortical:2 reyes:2 wise:1 common:9 empirically:3 spinal:2 cerebral:1 discussed:1 significant:1 lja:1 similarly:2 cortex:2 add:3 moderate:1 reverse:1 certain:1 inverted:2 motoneuron:14 additional:5 greater:1 mr:1 dashed:1 multiple:1 smooth:1 exceeds:1 match:3 cross:7 long:1 ipsps:1 serial:10 biophysics:1 involving:1 essentially:1 histogram:2 cell:32 preserved:1 background:1 addition:2 interval:4 diagram:1 pyramidal:1 probably:1 recording:2 effectiveness:1 unitary:3 presence:1 feedforward:1 psps:1 bic:2 een:1 approaching:1 rhythmically:2 reduce:1 linkage:10 effort:1 cause:2 action:2 latency:4 shortening:1 simplest:1 inhibitory:3 per:4 four:1 threshold:11 falling:2 mono:1 changing:1 advancing:1 sum:6 bia:10 inverse:1 fourth:1 reasonable:1 missed:1 noise1:1 followed:1 aic:1 convergent:2 activity:5 strength:5 constraint:1 ri:1 extracellular:1 department:1 combination:4 membrane:6 psp:1 remain:1 smaller:1 postsynaptic:9 shallow:1 remains:2 count:1 mechanism:1 appropriate:1 top:7 ensure:1 calculating:1 levitan:1 spike:10 occurs:1 link:21 simulated:1 relationship:1 difficult:1 mediating:2 susceptible:1 weinberg:1 rise:2 neuron:14 convolution:6 interacting:1 pair:3 connection:39 optimized:1 below:1 reliable:1 tranforming:1 ia:1 advanced:3 meth:1 representing:2 mediated:2 acknowledgement:1 relative:3 proportional:9 pi:2 excitatory:5 course:1 changed:1 supported:1 arriving:1 fetz:6 institute:1 fall:1 correspondingly:1 slice:1 feedback:8 calculated:3 kirkwood:2 valid:1 ending:1 forward:2 author:1 cope:1 active:1 reveals:1 ipsp:2 symmetry:2 broadening:1 investigated:1 linearly:3 intracellular:2 arrow:1 noise:8 neurosci:3 mediate:1 fig:12 n:1 msec:2 third:2 down:1 specific:1 bishop:1 showing:1 reverberating:2 decay:1 divergent:1 evidence:1 sequential:3 magnitude:2 illustrates:1 occurring:2 biophys:2 interneuron:1 deducing:1 polysynaptic:7 correlogram:19 chance:3 dispersed:1 stimulating:1 conditional:1 consequently:1 briefer:1 typical:4 determined:1 specifically:1 except:1 distributes:1 correlational:13 mput:1 modulated:1 betw:1 surmeier:1 correlated:1 |
547 | 150 | 256
AN INFORMATION THEORETIC APPROACH TO
RULE-BASED CONNECTIONIST EXPERT SYSTEMS
Rodney M. Goodman, John W. Miller
Department of Electrical Engineering
C altech 116-81
Pasadena, CA 91125
Padhraic Smyth
Communication Systems Research
Jet Propulsion Laboratories 238-420
4800 Oak Grove Drive
Pasadena, CA 91109
Abstract
We discuss in this paper architectures for executing probabilistic rule-bases in a parallel manner, using as a theoretical basis recently introduced information-theoretic
models. We will begin by describing our (non-neural) learning algorithm and theory
of quantitative rule modelling, followed by a discussion on the exact nature of two
particular models. Finally we work through an example of our approach, going from
database to rules to inference network, and compare the network's performance with
the theoretical limits for specific problems.
Introduction
With the advent of relatively cheap mass storage devices it is common in many
domains to maintain large databases or logs of data, e.g., in telecommunications,
medicine, finance, etc. The question naturally arises as to whether we can extract
models from the data in an automated manner and use these models as the basis
for an autonomous rational agent in the given domain, i.e., automatically generate
"expert systems" from data. There are really two aspects to this problem: firstly
learning a model and, secondly, performing inference using this model. What we
propose in this paper is a rather novel and hybrid approach to learning and inference. Essentially we combine the qu'alitative knowledge representation ideas of
AI with the distributeq, computational advantages of connectionist models, using
an underlying theoretical basis tied to information theory. The knowledge representation formalism we adopt is the rule-based representation, a scheme which is
well supported by cognitive scientists and AI researchers for modeling higher level
symbolic reasoning tasks. We have recently developed an information-theoretic algorithm called ITRULE which extracts an optimal set of probabilistic rules from a
given data set [1, 2, 3]. It must be emphasised that we do not use any form of neural
learning such as backpropagation in our approach. To put it simply, the ITRULE
learning algorithm is far more computationally direct and better understood than
(say) backpropagation for this particular learning task of finding the most informative individual rules without reference to their collective properties. Performing
useful inference with this model or set of rules, is quite a difficult problem. Exact
theoretical schemes such as maximum entropy (ME) are intractable for real-time
applications.
An Infonnation Theoretic Approach to Expert Systems
We have been investigating schemes where the rules represent links on a directed
graph and the nodes correspond to propositions, i.e., variable-value pairs. Our
approach is characterised by loosely connected, multiple path (arbitrary topology)
graph structures, with nodes performing local non-linear decisions as to their true
state based on both supporting evidence and their a priori bias. What we have in
fact is a recurrent neural network. What is different about this approach compared
to a standard connectionist model as learned by a weight-adaptation algorithm such
as BP? The difference lies in the semantics of the representation [4]. Weights such
as log-odds ratios based on log transformations of probabilities possess a clear meaning to the user, as indeed do the nodes themselves. This explicit representation of
knowledge is a key requirement for any system which purports to perform reasoning,
probabilistic or otherwise. Conversely, the lack of explicit knowledge representation
in most current connectionist approaches, i.e., the "black box" syndrome, is a major limitation to their application in critical domains where user-confidence and
explanation facilities are key criteria for deployment in the field.
Learning the model
Consider that we have M observations or samples available, e.g., the number of
items in a database. Each sample datum is described in terms of N attributes
or features, which can assume values in a corresponding set of N discrete alphabets. For example our data might be described in the form of lO-component binary
vectors. The requirement for discrete rather than continuous-valued attributes is
dictated by the very nature of the rule-based representation. In addition it is important to note that we do not assume that the sample data is somehow exhaustive and
"correct." There is a tendency in both the neural network and AI learning literature
to analyse learning in terms of learning a Boolean function from a truth table. The
implicit assumption is often made that given enough samples, and a good enough
learning algorithm we can always learn the function exactly. This is a fallacy, since
it depends on the feature representation. For any problem of interest there are
always hidden causes with a consequent non-zero Bayes misclassification risk, i.e.,
the function is dependent on non-observable features (unseen columns of the truth
table). Only in artificial problems such as game playing is "perfect" classification
possible - in practical problems nature hides the real features. This phenomenon
is well known in the statistical pattern recognition literature and renders invalid
those schemes which simply try to perfectly classify or memorise the training data.
We use the following simple model of a rule, i.e.,
IT Y
=y
then
X
= x with probability p
where X and Yare two attributes (random variables) with "x" and "y" being values
in their respective discrete alphabets. Given sample data as described earlier we
pose the problem as follows: can we find the "best" rules from a given data set,
say the K best rules? We will refer to this problem as that of generalised rule
induction, in order to distinguish it from the special case of deriving classification
257
258
Goodman, Miller and Smyth
rules. Clearly we require both a preference measure to rank the rules and a learning
algorithm which uses the preference measure to find the K best rules.
Let us define the information which the event y yields about the variable X, say
!(Xj y). Based on the requirements that !(Xj y) is both non-negative and that
its expectation with respect to Y equals the average mutual information J(Xj Y),
Blachman [5] showed that the only such function is the j-measure, which is defined
as
i(Xj y) = p(x\y) log (p(x\y))
p(x)
+ p(x\y) log (p(x)~y))
p(x)
More recently we have shown that i(Xj y) possesses unique properties as a rule
information measure [6]. In general the j-measure is the average change in bits
required to specify X between the a priori distribution (p(X)) and the a posteriori
distribution (p(X\y)). It can also be interpreted as a special case of the cross-entropy
or binary discrimination (Kullback [7]) between these two distributions. We further
define J(Xj y) as the average information content where J(X; y) = p(Y)-i(Xj y).
J(Xj y) simply weights the instantaneous rule information i(X; y) by the probability
that the left-hand side will occur, i.e., that the rule will be fired. This definition
is motivated by considerations of learning useful rules in a resource-constrained
environment. A rule with high information content must be both a good predictor
and have a reasonable probability of being fired, i.e., p(y) can not be too small.
Interestingly enough our definition of J(Xj y) possesses a well-defined interpretation
in terms of classical induction theory, trading off hypothesis simplicity with the
goodness-of-fit of the hypothesis to the data [8].
The ITRULE algorithm [1, 2, 3] uses the J-measure to derive the most informative
set of rules from an input data set. The algorithm produces a set of K probabilistic
rules, ranked in order of decreasing information content. The parameter K may be
user-defined or determined via some statistical significance test based on the size of
the sample data set available. The algorithm searches the space of possible rules,
trading off generality of the rules with their predictiveness, and using informationtheoretic bounds to constrain the search space.
Using the Model to Perform Inference
Having learned the model we now have at our disposal a set of lower order constraints on the N-th order joint distribution in the form of probabilistic rules. This
is our a priori model. In a typical inference situation we are given some initial
conditions (i.e., some nodes are clamped), we are allowed to measure the state of
some other nodes (possibly at a cost), and we wish to infer the state or probability
of one more goal propositions or nodes from the available evidence. It is important
to note that this is a much more difficult and general problem than classification of
a single, fixed, goal variable, since both the initial conditions and goal propositions
may vary considerably from one problem instance to the next. This is the inference problem, determining an a posteriori distribution in the face of incomplete and
uncertain information. The exact maximum entropy solution to this problem is in-
An Information Theoretic Approach to Expert Systems
tractable and, despite the elegance of the problem formulation, stochastic relaxation
techniques (Geman [9]) are at present impractical for real-time robust applications.
Our motivation then is to perform an approximation to exact Bayesian inference
in a robust manner. With this in mind we have developed two particular models
which we describe as the hypothesis testing network and the uncertainty network.
Principles of the Hypothesis Testing Network
In the first model under consideration each directed link from Y to x is assigned a
weight corresponding to the weight of evidence of yon x. This idea is not necessarily
new, although our interpretation and approach is different to previous work [10, 4].
Hence we have
W
-1
:r.y -
p{xIY) -1
og p(x)
p(:xIY)
og p(x)
R
and
:r.
= -log p(x)
p(x)
and the node x is assigned a threshold term corresponding to a priori bias. We use
a sigmoidal activation function, i.e.,
a(x)
=
1
--~7'""E=-t----;;R'--,
l+e
n
where
l:J.E:r.
=
I:
W:r.y; .
q(y,) -
R:r.
,=1
T
based on multiple binary inputs Y1 ... Yn to x. Let 8 be the set of all Yi which are
hypothesised true (Le., a{yd = 1), so that
AE =
L.l:r.
Iog p(x)
+ '" (1 p(xlYd _ 1 p(x IY,))
p(x)
L- og p(x)
og p(x)
y;ES
If each y, is conditionally independent given x then we can write
p(xIS) = p(x)
p(xIS)
p(x)
II
y;ES
p(xIY,)
p(xlYd
Therefore the updating rule for conditionally independent y, is:
T . log
a(x)
= log
1 - a(x)
p(xI8)
1 - p(x/S)
Hence a(x) > ~ iff p{xI8) > ~ and if T == 1, a(x) is exactly p(xIS). In terms of a
hypothesis test, a(x) is chosen true iff:
'L" Iog p(XIYi) > - Iog-p{x)
p(XIYi) -
p(x)
Since this describes the Neyman-Pearson decision region for independent measurements (evidence or yd with R:r. = -log :~~~ [11], this model can be interpreted as
a distributed form of hypothesis testing.
259
260
Goodman, Miller and Smyth
Principles of the Uncertainty Network
For this model we defined the weight on a directed link from Yi to x as
W XYi
=
.
(
p(XIYi)
_
p(xIYi)))
si.1(XjYi) = Si? p(XIYi}log( p(x) ) + p(xly,)log( p(x)
where Si = ?1 and the threshold is the same as the hypothesis model. We can
interpret W:Z lli as the change in bits to specify the a posteriori distribution of x. H
P(XIYi) > p{x), w:ZYi has positive support for x, i.e., Si = +1. H P{XIYi) < p(x), W:Z lli
has negative support for x, Le., Si = -1. IT we interpret the activation a(Yi) as an
estimator (p(y)) for p(Yi), then for multiple inputs,
i
~ ..
(
)
P(XIYi)
(_
P(XIYi) )
- ~ p(Yi).Si. p(XIYi log( p{x) ) + P xly,) log( p(x) )
?
This sum over input links weighted by activation functions can be interpreted as
the total directional change in bits required to specify x, as calculated locally by the
node x. One can normalise !:1Ex to obtain an average change in bits by dividing by
a suitable temperature T. The node x can make a local decision by recovering p(x)
from an inverse J-measure transformation of !:1E (the sigmoid is an approximation
to this inverse function).
Experimental Results and Conclusions
In this section we show how rules can be generated from example data and automatically incorporated into a parallel inference network that takes the form of a
multi-layer neural network. The network can then be "run" to perform parallel
inference. The domain we consider is that of a financial database of mutual funds,
using published statistical data [12]. The approach is, however, typical of many
different real world domains.
Figure 1 shows a portion of a set of typical raw data on no-load mutual funds.
Each line is an instance of a fund (with name omitted), and each column represents
an attribute (or feature) of the fund. Attributes can be numerical or categorical.
Typical categorical attributes are the fund type which reflect the investment objectives of the fund (growth, growth and income, balanced, and agressive growth) and
a typical numerical attribute is the five year return on investment expressed as a
percentage. There are a total of 88 fund examples in this data set. From this raw
data a second quantized set of the 88 examples is produced to serve as the input to
ITRULE (Figure 2). In this example the attributes have been categorised to binary
values so that they can be directly implemented as binary neurons. The ITRULE
software then processes this table to produce a set of rules. The rules are ranked
in order of decreasing information according to the J-measure. Figure 3 shows a
An Infonnation Theoretic Approach to Expert Systems
portion (the top ten rules) of the ITRULE output for the mutual fund data set. The
hypothesis test log-likelihood metric h(Xj y), the instantaneous j-measure j(Xj y),
and the average J-measure J(Xj y), are all shown, together with the rule transition
probability p{x/y).
In order to perform inference with the ITRULE rules we need to map the rules
into a neural inference net. This is automatically done by ITRULE which generates a network file that can be loaded into a neural network simulator. Thus rule
information metrics become connection weights. Figure 4 shows a typical network
derived from the ITRULE rule output for the mutual funds data. For clarity not
all the connections are shown. The architecture consists of two layers of neurons
(or "units"): an input layer and an output layer, both of which have an activation
within the range {O,l}. There is one unit in the input layer (and a corresponding
unit in the output layer) for each attribute in the mutual funds data. The output
feeds back to the input layer, and each layer is synchronously updated. The output
units can be considered to be the right hand sides of the rules and thus receive
inputs from many rules, where the strength of the connection is the rule's metric.
The output units implement a sigmoid activation function on the sum of the inputs, and thus compute an activation which is an estimator of the right hand side
posteriori attribute value. The input units simply pass this value on to the output
layer and thus have a linear activation.
To perform inference on the network, a probe vector of attribute values is loaded
into the input and output layers. Known values are clamped and cannot change
while unknown or desired attribute values are free to change. The network then
relaxes and after several feedback cycles converges to a solution which can be read
off the input or output units. To evaluate the models we setup fo~r standard classification tests with varying number of nodes clamped as inPlits. Undamped nodes
were set to their a priori probability. After relaxing the network, the activation of
the "target" node was compared with the true attribute values for that sample in
order to determine classification performance. The two models were each trained
on 10 randomly selected sets of 44 samples. The performance results given in Table
1 are the average classification rate of the models on the other 44 unseen samples.
The Bayes risk (for a uniform loss matrix) of each classification test was calculated
from the 88 samples. The actual performance of the networks occasionally exceeded
this value due to small sample variations on the 44/44 cross validations.
Table 1
Units Cramped Uncertainty Test
9
5
2
1
66.8%
70.1%
48.2%
51.4%
HYPOthesis Test
70.4%
70.1%
63.0%
65.7%
1 - Bayes' Risk
88.6%
80.6%
63.6%
64.8%
261
262
Goodman, Miller and Smyth
We conclude from the performance of the networks as classifiers that they have
indeed learned a model of the data using a rule-based representation. The hypothesis network performs slightly better than the uncertainty model, with both being
quite close to the estimated optimal rate (the Bayes' risk). Given that we know
that the independence assumptions in both models do not hold exactly, we coin the
term robust inference to describe this kind of accurate behaviour in the presence of
incomplete and uncertain information. Based on these encouraging initial results,
our current research is focusing on higher-order rule networks and extending our
theoretical understanding of models of this nature.
Acknowledgments
This work is supported in part by a grant from Pacific Bell, and by Caltech's
program in Advanced Technologies sponsored by Aerojet General, General Motors
and TRW. Part of the research described in this paper was carried out by the Jet
Propulsion Laboratory, California Institute of Technology, under a contract with
the National Aeronautics and Space Administration. John Miller is supported by
NSF grant no. ENG-8711673.
References
1. R. M. Goodman and P. Smyth, 'An information theoretic model for rule-based
expert systems,' presented at the 1988 International Symposium on Information
Theory, Kobe, Japan.
2. R. M. Goodman and P. Smyth, 'Information theoretic rule induction,' Proceedings of the 1988 European Conference on AI, Pitman Publishing: London.
3. R. M. Goodman and P. Smyth, 'Deriving rules from databases: the ITRULE
algorithm,' submitted for publication.
4. H. Geffner and J. Pearl, 'On the probabilistic semantics of connectionist networks,' Proceedings of the 1987 IEEE ICNN, vol. II, pp. 187-195.
5. N. M. Blachman, 'The amount of information that y gives about X,' IEEE
Transactions on Information Theory, vol. IT-14 (1), 27-31, 1968.
6. P. Smyth and R. M. Goodman, 'The information content of a probabilistic
rule,' submitted for publication.
7. S. Kullback, Information Theory and Statistics, New York: Wiley, 1959.
8. D. Angluin and C. Smith, 'Inductive inference: theory and methods,' ACM
Computing Surveys, 15(9), pp. 237-270, 1984.
9. S. Geman, 'Stochastic relaxation methods for image restoration and expert systems,' in Maximum Entropy and Bayesian Methods in Science and Engineering
(Vol. 2), 265-311, Kluwer Academic Publishers, 1988.
10. G. Hinton and T. Sejnowski, 'Optimal perceptual inference,' Proceedings of the
IEEE CVPR 1989.
11. R. E. Blahut, Principles and Practice of Information Theory, Addison-Wesley:
Reading, MA, 1987.
12. American Association of Investors, The individual investor's guide to no-load
mutual funds, International Publishing Corporation: Chicago, 1987.
An Infonnation Theoretic Approach to Expert Systems
Fund Type
5 Year Diver- Beta
Bull Bear Stocks
Return sity
(Risk) Perf. Perf. 0/0
0/0
Balanced
136
C
0.8 B
D
87
Growth
32 .5
C
1.05 E
B
81
Growth& Income 88.3
A
0.96 C
82
D
Agressive
-24
A
1.23 E
E
95
Growth&lncome
172
0.59 A
E
73
B
Balanced
144
C
0.71 B
B
51
Flgure1.
Type Type Type Type 5 Year
A
B
G (?J Return 0/0
S&P=1380/0
above S&P
below S&P
no no yes no below
no no yes no below
no no no yes below
no no no yes above
no no no yes below
no no yes no above
Beta
under1
over1
under1
under1
under1
under1
Figure 2.
Invest- Net
Distri- Expense Turn- Total
ment Asset butions Ratio % over Assets
Incm. $ Value $ (%NAV\
Rate %$M
0.67 37 .3 17 . 63
0 .79
34 415
- 0.02 12.5
0.88
1.4 200
16
0.14 11.9
4 .78
1.34 127
27
0.02 6.45
9 .30
1.4 1 61
64
0.53 13.6
9.97
1.09
31 113
0.72
13 10 .44
0 .98 239 190
Raw Mutual Funds Data
Stocks Turn>90% over
no
no
no
no
yes
no
IF
IF
IF
IF
IF
IF
IF
IF
IF
IF
5yrRebS&P
BullJ)erf
Assets
BullJ)erf
typeA
BullJ)erf
typeGl
BullJ)erf
typeG
Assets
above
low
large
high
yes
low
yes
high
yes
small
Diver- Bull Bear
sity Perf. Perf.
<100% <$100M <150/0NAV C.D.E C.D.E C.D.E
>100% >$100M >150/0NAV A.B
AB A,B
large
high low
low
high
low
high
small
low
low
low high
high
small
low
high
low low
low
large
low
low
high high
low
small
high
high
low high
large
high
high high low
low
Quantized Mutual Funds Data
ITRULE rule output: Mutual Funds
1
2
3
4
5
6
7
8
9
10
Distributions
Assets
lHEN
lHEN
lHEN
lHEN
lHEN
lHEN
lHEN
lHEN
lHEN
lHEN
BullJ)erf
5yrRet>S&P
BullJ)erf
5yrRet>s&P
typeG
Assets
typeG
Assets
typeA
Bull perf
high
below
high
above
no
small
no
large
no
low
p(x/y)
j(X;y)
0.97
0.98
0.81
0.40
0 .04
0.18
0.05
0.72
0.97
0 .26
0.75
0.41
0.28
0.25
0 .50
0.25
0 .49
0.21
0.27
0.19
J(X;y) h(X;y)
0.235
0 .201
0.127
0.127
0 .123
0.121
0.109
0.109
0.108
0.103
4.74
4.31
2.02
-1.71
-3 . 87
-1 . 95
-3.74
1.64
3 .54
-1.57
Figure 3. Top Ten Mutual Funds Rules
nfo2atl~
DD
00~
metric connection
weights
one unit per attribute
I
I
I
~ ~ Input layer - linear units
Feedback connections
weight 1
=
o DOD 0 DD0 o
I
I
output layer - sigmoid units
Figure 4. Rule Network
263
| 150 |@word agressive:2 eng:1 initial:3 xiy:3 interestingly:1 current:2 activation:8 xiyi:10 si:6 must:2 john:2 chicago:1 numerical:2 informative:2 cheap:1 motor:1 sponsored:1 fund:16 discrimination:1 selected:1 device:1 item:1 smith:1 quantized:2 node:12 preference:2 firstly:1 oak:1 sigmoidal:1 five:1 direct:1 become:1 symposium:1 beta:2 consists:1 combine:1 manner:3 indeed:2 themselves:1 multi:1 simulator:1 decreasing:2 automatically:3 actual:1 encouraging:1 begin:1 distri:1 underlying:1 mass:1 advent:1 what:3 nav:3 kind:1 interpreted:3 developed:2 finding:1 transformation:2 corporation:1 impractical:1 quantitative:1 growth:6 finance:1 exactly:3 classifier:1 unit:11 grant:2 yn:1 generalised:1 positive:1 engineering:2 scientist:1 understood:1 local:2 limit:1 despite:1 path:1 yd:2 black:1 might:1 conversely:1 relaxing:1 deployment:1 range:1 directed:3 practical:1 unique:1 acknowledgment:1 testing:3 investment:2 practice:1 implement:1 backpropagation:2 bell:1 confidence:1 symbolic:1 cannot:1 close:1 storage:1 put:1 risk:5 map:1 survey:1 simplicity:1 rule:49 estimator:2 deriving:2 financial:1 autonomous:1 xjyi:1 variation:1 updated:1 target:1 user:3 smyth:8 exact:4 us:2 hypothesis:10 recognition:1 updating:1 aerojet:1 geman:2 database:5 electrical:1 region:1 connected:1 hypothesised:1 cycle:1 balanced:3 environment:1 trained:1 serve:1 basis:3 joint:1 stock:2 alphabet:2 describe:2 london:1 sejnowski:1 artificial:1 pearson:1 exhaustive:1 quite:2 valued:1 cvpr:1 say:3 otherwise:1 statistic:1 erf:6 unseen:2 analyse:1 advantage:1 net:2 xi8:2 propose:1 ment:1 adaptation:1 iff:2 fired:2 invest:1 requirement:3 extending:1 produce:2 perfect:1 executing:1 converges:1 sity:2 derive:1 recurrent:1 pose:1 dividing:1 recovering:1 implemented:1 trading:2 xly:2 correct:1 attribute:14 stochastic:2 require:1 behaviour:1 really:1 icnn:1 itrule:11 proposition:3 secondly:1 hold:1 predictiveness:1 considered:1 major:1 vary:1 adopt:1 omitted:1 infonnation:3 weighted:1 clearly:1 always:2 rather:2 og:4 varying:1 publication:2 derived:1 modelling:1 rank:1 likelihood:1 posteriori:4 inference:16 dependent:1 pasadena:2 hidden:1 going:1 semantics:2 classification:7 priori:5 constrained:1 special:2 mutual:11 field:1 equal:1 having:1 represents:1 connectionist:5 kobe:1 randomly:1 national:1 individual:2 blahut:1 maintain:1 ab:1 interest:1 accurate:1 grove:1 respective:1 incomplete:2 loosely:1 desired:1 theoretical:5 uncertain:2 instance:2 formalism:1 modeling:1 boolean:1 column:2 classify:1 earlier:1 goodness:1 restoration:1 bull:3 cost:1 predictor:1 uniform:1 dod:1 too:1 considerably:1 international:2 probabilistic:7 off:3 contract:1 together:1 iy:1 reflect:1 padhraic:1 possibly:1 geffner:1 cognitive:1 expert:8 american:1 return:3 japan:1 depends:1 try:1 portion:2 bayes:4 investor:2 parallel:3 rodney:1 loaded:2 miller:5 correspond:1 yield:1 directional:1 yes:10 bayesian:2 raw:3 produced:1 xyi:1 lli:2 drive:1 researcher:1 published:1 asset:7 submitted:2 fo:1 definition:2 pp:2 naturally:1 elegance:1 rational:1 knowledge:4 back:1 trw:1 focusing:1 feed:1 disposal:1 higher:2 exceeded:1 wesley:1 specify:3 formulation:1 done:1 box:1 generality:1 implicit:1 hand:3 lack:1 somehow:1 name:1 true:4 facility:1 hence:2 assigned:2 inductive:1 read:1 laboratory:2 conditionally:2 game:1 criterion:1 butions:1 theoretic:9 performs:1 temperature:1 reasoning:2 meaning:1 image:1 instantaneous:2 novel:1 recently:3 consideration:2 sigmoid:3 common:1 association:1 interpretation:2 kluwer:1 interpret:2 refer:1 measurement:1 ai:4 etc:1 base:1 aeronautics:1 hide:1 dictated:1 showed:1 occasionally:1 binary:5 yi:5 caltech:1 syndrome:1 determine:1 ii:2 multiple:3 infer:1 jet:2 academic:1 cross:2 iog:3 ae:1 essentially:1 expectation:1 metric:4 represent:1 receive:1 addition:1 publisher:1 goodman:8 posse:3 file:1 odds:1 presence:1 enough:3 relaxes:1 automated:1 xj:12 fit:1 independence:1 architecture:2 topology:1 perfectly:1 idea:2 administration:1 whether:1 motivated:1 render:1 york:1 cause:1 useful:2 clear:1 amount:1 locally:1 ten:2 generate:1 angluin:1 percentage:1 nsf:1 estimated:1 per:1 discrete:3 write:1 vol:3 key:2 threshold:2 clarity:1 graph:2 relaxation:2 sum:2 year:3 run:1 inverse:2 uncertainty:4 telecommunication:1 reasonable:1 decision:3 bit:4 bound:1 layer:12 followed:1 datum:1 distinguish:1 strength:1 occur:1 constraint:1 constrain:1 bp:1 software:1 generates:1 aspect:1 performing:3 relatively:1 department:1 pacific:1 according:1 describes:1 slightly:1 qu:1 computationally:1 resource:1 neyman:1 discus:1 describing:1 turn:2 mind:1 know:1 addison:1 tractable:1 available:3 yare:1 probe:1 coin:1 top:2 publishing:2 medicine:1 emphasised:1 classical:1 objective:1 question:1 link:4 normalise:1 propulsion:2 me:1 induction:3 ratio:2 difficult:2 setup:1 expense:1 memorise:1 negative:2 collective:1 unknown:1 perform:6 observation:1 neuron:2 supporting:1 situation:1 hinton:1 communication:1 incorporated:1 y1:1 synchronously:1 arbitrary:1 introduced:1 pair:1 required:2 connection:5 california:1 learned:3 pearl:1 below:6 pattern:1 reading:1 program:1 explanation:1 critical:1 misclassification:1 event:1 hybrid:1 ranked:2 suitable:1 dd0:1 advanced:1 scheme:4 technology:2 carried:1 perf:5 categorical:2 extract:2 literature:2 understanding:1 determining:1 loss:1 bear:2 limitation:1 validation:1 undamped:1 agent:1 principle:3 dd:1 playing:1 lo:1 supported:3 free:1 bias:2 side:3 guide:1 institute:1 face:1 pitman:1 diver:2 distributed:1 feedback:2 calculated:2 world:1 transition:1 made:1 far:1 income:2 transaction:1 observable:1 informationtheoretic:1 kullback:2 categorised:1 investigating:1 conclude:1 xi:3 continuous:1 fallacy:1 search:2 table:5 learn:1 nature:4 robust:3 ca:2 necessarily:1 european:1 domain:5 significance:1 motivation:1 allowed:1 wiley:1 purport:1 explicit:2 wish:1 lie:1 clamped:3 tied:1 perceptual:1 zyi:1 load:2 specific:1 consequent:1 evidence:4 intractable:1 entropy:4 simply:4 expressed:1 truth:2 acm:1 ma:1 goal:3 invalid:1 content:4 change:6 characterised:1 determined:1 typical:6 called:1 total:3 pas:1 tendency:1 e:2 experimental:1 support:2 arises:1 evaluate:1 phenomenon:1 ex:1 |
548 | 1,500 | Outcomes of the Equivalence of Adaptive Ridge
with Least Absolute Shrinkage
Yves Grandvalet
Stephane Canu
Heudiasyc, UMR CNRS 6599, Universite de Technologie de Compiegne,
BP 20.529, 60205 Compiegne cedex, France
Yves.Grandvalet@hds.utc.fr
Abstract
Adaptive Ridge is a special form of Ridge regression, balancing the
quadratic penalization on each parameter of the model. It was shown to
be equivalent to Lasso (least absolute shrinkage and selection operator),
in the sense that both procedures produce the same estimate. Lasso can
thus be viewed as a particular quadratic penalizer.
From this observation, we derive a fixed point algorithm to compute the
Lasso solution. The analogy provides also a new hyper-parameter for tuning effectively the model complexity. We finally present a series ofpossible extensions oflasso performing sparse regression in kernel smoothing,
additive modeling and neural net training.
1
INTRODUCTION
In supervised learning, we have a set of explicative variables x from which we wish to predict a response variable y. To solve this problem, a learning algorithm is used to produce a
predictor x) from a learning set Sf. = {(Xi, yd
of examples. The goal of prediction
may be: 1) to provide an accurate prediction of future responses, accuracy being measured
by a user-defined loss function; 2) to quantify the effect of each explicative variable in the
response; 3) to better understand the underlying phenomenon.
J(
H=l
Penalization is extensively used in learning algorithms. It decreases the predictor variability
to improve the prediction accuracy. It is also expected to produce models with few non-zero
coefficients if interpretation is planned.
Ridge regression and Subset Selection are the two main penalization procedures. The former is stable, but does not shrink parameters to zero, the latter gives simple models, but is
unstable [1]. These observations motivated the search for new penalization techniques such
as Garrotte, Non-Negative Garrotte [1], and Lasso (least absolute shrinkage and selection
operator) [10].
Y. Grandvalet and S. Canu
446
Adaptive Ridge was proposed as a means to automatically balance penalization on different
coefficients. It was shown to be equivalent to Lasso [4]. Section 2 presents Adaptive Ridge
and recalls the equivalence statement. The following sections give some of the main outcomes ofthis connection. They concern algorithmic issues in section 3, complexity control
in section 4, and some possible generalizations oflasso to non-linear regression in section 5.
2
ADAPTIVE RIDGE REGRESSION
For clarity of exposure, the formulae are given here for linear regression with quadratic loss.
The predictor is defined as j( x) = rff x, with rff = (f31, ... , f3d). Adaptive Ridge is a
modification of the Ridge estimate, which is defined by the quadratic constraint ~~ = 1 f3; ~
C applied to the parameters. It is usually computed by minimizing the Lagrangian
l
jj =
Argmin
i=l
{3
d
L (L
d
f3j Xij - Yi) 2 + A
j=l
L
f3; ,
(1)
j=l
where A is the Lagrange multiplier varying with the bound C on the norm of the parameters.
When the ordinary least squares (OLS) estimate maximizes likelihood 1 , the Ridge estimate
may be seen as a maximum a posteriori estimate. The Bayes prior distribution is a centered
normal distribution, with variance proportional to 1/ A. This prior distribution treats all covariates similarly. It is not appropriate when we know that all covariates are not equally
relevant.
----0
The garrotte estimate [1] is based on the OLS estimate,8 . The standard quadratic constraint
is replaced by ~~ = 1 f3] (iJf ~ C. The coefficients with smaller OLS estimate are thus
more heavily penalized. Other modifications are better explained with the prior distribution viewpoint. Mixtures of Gaussians may be used to cluster different set of covariates.
Several models have been proposed, with data dependent clusters [9], or classes defined a
priori [7]. The Automatic Relevance Determination model [8] ranks in the latter type. In [4],
we propose to use such a mixture, in the form
t
d
d
Xij - Yi) 2
,8=
ArgmmL.,.,
---. " " (L.,.,f3j
""
{3
i=l j=l
+ "L.,.,Ajf3j
"
2
(2)
j=l
Here, each coefficient has its own prior distribution. The priors are centered normal distributions with variances proportional to 1/ Aj. To avoid the simultaneous estimation of these
d hyper-parameters by trial, the constraint
1
d
1
j=l
J
1
dL ~ = ~
, Aj > 0
(3)
is applied on A = (A1, .. . , Ad)T, where A is a predefined value. This constraint is a link
between the d prior distributions. Their mean variance is proportional to 1/ A. The values of
Aj are automatically2 induced from the sample, hence the qualifieradaptative. Adaptativity
refers here to the penalization balance on {Pj }, not to the tuning of the hyper-parameter A.
{(.C,} are independently and identically drawn from some distribution, and that some{3" exists,
such that Y. = {3" T (.C, + e, where c is a centered normal random variable, then the empirical cost
1 If
"'0
based on the quadratic loss is proportional to the log-likelihood of the sample. The OLS estimate{3
is thus the maximum likelihood estimate offJ'.
2 Adaptive Ridge, as Ridge or Lasso, is not scale invariant, so that the covariates should be normalized to produce sensible estimates.
447
Equivalence of Adaptive Ridge with Least Absolute Shrinkage
It was shown [4] that Adaptive Ridge and least absolute value shrinkage are equivalent, in
the sense that they yield the same estimate. We remind that the Lasso estimate is defined by
e
j3 =
Argmin
j3
d
L (L (3j
i=l
d
Xij -
Yi ) 2
L
subject to
j =l
l{3j l ~
f{
(4)
.
j =l
The only difference in the definition of the Adaptive Ridge and the Lasso estimate is that
the Lagrangian form of Adaptive Ridge uses the constraint CL1=1l{3j 1) 2/ d ~ f{ 2.
3
OPTIMIZATION ALGORITHM
Tibshirani [10] proposed to use quadratic programming to find the l,asso solution, with 2d
variables (positive and negative parts of (3j ) and 2d + 1 constraints (signs of positive and
negative parts of (3j plus constraint (4)). Equations (2) and (3) suggest to use a fixed point
(FP) algorithm. At each step s, the FP algorithm estimates the optimal parameters ). of
.y)
the Bayes prior based on the estimate (3) S -1 ) , and then maximizes the posterior to compute
the current estimate (3) S ) ?
As the parameterization (j3, A) may lead to divergent solutions, we define new variables
and
Cj
= Vr;:
I;
.= 1, .. . , d
for J
(5)
The FP algorithm updates alternatively c and -y as follows:
(S)2
{
cj
d ,jS -1 )2
=
,,",d
(s -1 )2
(6)
L.., k =l /k
-y (s) = (diag( c (s) )XT
X
diag( c (s) )
+ AI) -1 diag( c (S) )XT y
where Xi j = X i j , I is the identity matrix, and diag( c) is the square matrix with the vector
c on its diagonal.
The algorithm can be initialized by the Ridge or the OLS estimate. In the latter case,,B(1) is
the garrotte estimate.
Practically, 'if
,ys)
lys-1 )is small compared to numerical accuracy, then c~s) is set to zero. In
turn,
is zero, and the system to be solved in the second step to determine -y can be
reduced to the other var~ables. If cJ' is set to zero at any time during the optimization process, the final estimate {3j will be zero. The computations are simplified, but it is not clear
whether global convergence can be obtained with this algorithm. It is easy to show the convergence towards a local minimum, but we did not find general conditions ensuring global
convergence. If these conditions exist, they rely on initial conditions.
Finally, we stress that the optimality conditions for c (or in a less rigorous sense for A) do
not depend on the first part of the cost minimized in (2). In consequence, the equivalence
between Adaptive Ridge and lasso holds/or any model or loss/unction . The FP algorithm
can be applied to these other problems, without modifying the first step.
4
COMPLEXITY TUNING
The Adaptive Ridge estimate depends on the learning set Sf. and on the hyper-parameter
A. When the estimate is defined by (2) and (3), the analogy with Ridge suggests A as the
448
Y. Grandvalet and S. Canu
~
"natural" hyper-parameter for tuning the complexity of the regressor. As ..\ goes to zero, j3
r-<>
approac~es the OLS estimatej3 , and the number of effective parameters is d. As ..\ goes to
infinity, (3 goes to zero and the number of effective parameters is zero.
When the estimate is defined by (4), there is no obvious choice for the hyper-parameter controlling complexity. Tibshirani [10] proposed to use v = 'Lf=1
~
r-<>
l.8j 1/ 'Lf=l ~ I. As v goes
~
to one,{3 approaches{3 ; as v goes to infinity, {3goes to zero.
The weakness of v is that it is explicitly defined from the OLS estimate. As a result, it is
variable when the design matrix is badly conditioned. The estimation of v is thus harder,
and the overall procedure looses in stability. This is illustrated on an experiment following
Breiman's benchmark [1] with 30 highly correlated predictors lE(XjX k ) = plj-k l , with
p = 1 - 10- 3 .
We generate 1000 Li.d. samples of size ? = 60. For each sampie s1, the modeling error (ME) is computed for several values of v and'\. We select v k and ,\k achieving the
lowest ME. For one sample, there is a one to one mapping from v to'\. Thus ME is the
same for v k and ,\k. Then, we compute v* and ..\* achieving the best average ME on
the 1000 samples. As v k and ,\k achieve the lowest ME for s1, the ME for
is higher
or equal for v* and ,\ *. Due to the wide spread of {Vk }, the average loss encountered is
twice for v* than for ,\*: 1/1000 'L!~10 (ME(s~, v*) - ME(s; , v k )) = 4.6 10- 2 , and
1/1000'L!~010 (ME(s~ , ..\ * ) - ME(s1 , ,\k)) = 2.310- 2 . The average modeling error are
ME(v*) = 1.910- 1 and ME("\*) = 1.710- 1.
s1
The estimates of prediction error, such as leave-one-out cross-validation tend to be variable.
Hence, complexity tuning is often based on the minimization of some estimate of the mean
prediction error (e.g bootstrap, K-fold cross-validation). Our experiment supports that, regarding mean prediction error, the optimal ,\ performs better than the optimal v . Thus, ,\ is
the best candidate for complexity tuning.
Although,\ and v are respectively the control parameter of the FP and QP algorithms, the
preceding statement does not imply that we should use the FP algorithm. Once the solution
73 is known, v or ,\ are easily computed. The choice of one hyper-parameter is not linked to
the choice of the optimization algorithm.
5
APPLICATIONS
Adaptive Ridge may be applied to a variety of regression techniques. They include kernel
smoothing, additive and neural net modeling.
5.1
KERNEL SMOOTHING
Soft-thresholding was proved to be efficient in wavelet functional e~timation [2]. Kernel
smoothers [5] can also benefit from the sparse representation given by soft-thresholding
methods. For these regressors, l( x) = 'L1=1 f3i K(x , xd+f3o, there are as many covariates
as pairs in the sample. The quadratic procedure of Lasso with 2? + 1 constraints becomes
computationally expensive, but the FP algorithm of Adaptive Ridge is reasonably fast to
converge.
An example of least squares fitting is shown in fig. 1 for the motorcycle dataset [5]. On
this example, the hyperparameter ,\ has been estimated by .632 bootstrap (with 50 bootstrap replicates) for Ridge and Adaptive Ridge regressions. For tuning..\, it is not necessary
to determine the coefficients {3 with high accuracy. Hence, compared to Ridge regression,
449
Equivalence ofAdaptive Ridge with Least Absolute Shrinkage
the overall amount of computation required to get the Adaptive Ridge estimate was about
six times more important. For evaluation, Adaptive Ridge is ten times faster as Ridge regression as the final fitting uses only a few kernels (11 out of 133).
-AR
-- - - R
+ "'+
-1+
+ ++
+t+
+ + +
+
x
Figure 1: Adaptive Ridge (AR) and Ridge (R) in kernel smoothing on the
motorcycle data. The + are data points, and. are the prototypes corresponding to the kernels with non-zero coefficients in AR. The Gaussian
kernel used is represented dotted in the lower right-hand corner.
Girosi [3] showed an equivalence between a version of least absolute shrinkage applied to
kernel smoothing, and Support Vector Machine (SVM). However, Adaptive Ridge, as applied here, is not equivalent to SVM, as the cost minimized is different. The fit and prototypes are thus different from the fit and support vectors that would be obtained from SVM.
5.2
ADDITIVE MODELS
LJ=
Additive models [6] are sums of univariate functions , f( x) =
1
fj (x j ).
In the non-
parametric setting, {fj} are smooth but unspecified functions. Additive models are easily
represented and thus interpretable, but they require the ch~ice of the relevant covariates to
be included in the model, and of the smoothness of each Ij.
In the form presented in the two previous sections, Adaptive Ridge regression penalizes
differently each individual coefficient, but it is easily extended to the pooled penalization of
coefficients. Adaptive Ridge may th~ be used as an alternative to BRUTO [6] to balance
the penalization parameters on each Ij .
A classical choice for fj is cubic spline smoothing. Let B j denote the ? x (? + 2) matrix of
the unconstrained B-spline basis, evaluated at Xij. Let 51 j be the (? + 2) x (f + 2) matrix
corresponding to the penalization of the second derivative of J;. The coefficients of fj in
the unconstrained B-spline basis are noted /3j. The "natural" extension of Adaptive Ridge
is to minimize
d
II L
d
B j/3j -
YI12 + L
j=l
,\jf3]' 51 j /3j
(7)
,
j=l
subject to constraint (3). This problem is easily shown to have the same solution as the
minimization of
I
t,
Bjf3j -
yll' + A
(t,
Jf3J
fljf3j)
2
(8)
Note that if the cost (8) is optimized with respect to a single covariate, the solution is a usual
smoothing spline regression (with quadratic penalization). In the multidimensional case,
450
Y. Grandvalet and S. Canu
J
ex] =rif o'j/3j = {Ij'(t)}2dt may be used to summarize the non-linearity of Ij, thus lajl
can be interpreted as a relevance index operating besides linear dependence of feature j.
The penalizer in (8) is a least absolute shrinkage operator applied to ex j. Hence, formula (8)
may be interpreted as "quadratic penalization within, and soft-thresholding between covariates".The FP algorithm of section 3 is easily modified to minimize (8), and backfitting may
be used to solve the second step of this procedure.
A simulated example in dimension five is shown in fig. 2. The fitted univariate functions
are plotted for five values of'\. There is no dependency between the the explained variable
and the last covariate. The other covariates affect the response, but the dependency on the
first features is smoother, hence easier to capture and more relevant for the spline smoother.
For a small value of '\, the univariate functions are unsmooth, and the additive model is
interpolating the data. For,\ 10- 4 , the dependencies are well estimated on all covariates.
As ,\ increases, the cov~riates with higher coordinate number are more heavily penalized,
and the corresponding Ij tend to be linear.
=
~
I
0
II
"<t
I
0
"
.
"
"
....<
"
....<
.., ......
..
....+-t-.
'
~----~-------+------~------~----~
t
+. .. .... :.:... :.
.. ..
"
....<~:______J -_ _ _ _ _ _~_ _ _ _ _ _~_ _ _ _~_ _ _ _ _ _- J
Figure 2: Adaptive Ridge in additive modeling on simulated data. The
true model is y = Xl + cos( rrx2) + cOS(2rrx3) + cos(3rrx4) + E. The covariates are independently drawn from a uniform distribution on [-1, 1]
and E is a Gaussian noise of standard deviation (j 0.3. The solid curves
are the estimated univariate functions for different values of '\, and + are
partial residuals.
Linear trends are not penalized in cubic spline smoothing. Thus, when after convergence
=
~T
~
/3j n j/3j
= 0, the jth covariate is not eliminated. This can be corrected by applying Adap-
bve Ridge a second time. To test if a significant linear trend can be detected, a linear (pek #- j being cubic splines.
nalized) model may be used for 1;, the remaining
h,
5.3
MLP FITTING
The generalization to the pooled penalization of coefficients can also be applied to MultiLayered Perceptrons to control the complexity of the fit. If weights are penalized individually, Adaptive Ridge is equivalent to the Lasso. If weights are pooled by layer, Adaptive
Ridge automatically tunes the amount of penalization on each layer, thus avoiding the multiple hyper-parameter tuning necessary in weight-decay [7].
Equivalence ofAdaptive Ridge with Least Absolute Shrinkage
451
Figure 3: groups of weights for two examples of Adaptive Ridge in MLP
fitting. Left: hidden node soft-thresholding. Right: input penalization
and selection, and individual smoothing coefficient for each output unit.
Two other interesting configurations are shown in fig. 3. If weights are pooled by incoming and outcoming weights of a unit, node penalization/pruning is performed. The weight
groups may also gather the outcoming weights from each input unit, orthe incoming weights
from each output unit (one set per input plus one per output). The goal here is to penalize/select the input variables according to their relevance, and each output variable according to the smoothness of the corresponding mapping. This configuration proves itself especially useful in time series prediction, where the number of inputs to be fed into the network
is not known in advance. There are also more complex choices of pooling, such as the one
proposed to encourage additive modeling in Automatic Relevance Determination [8].
References
[1] L. Breiman. Heuristics of instability and stabilization in model selection. The Annals
of Statistics, 24(6):2350-2383, 1996.
[2] D.L Donoho and I.M. Johnstone. Minimax estimation via wavelet shrinkage. Ann.
Statist., 26(3):879--921,1998.
[3] F. Girosi. An equivalence between sparse approximation and support vector machines.
Technical Report 1606, M.LT AI Laboratory, Cambridge, MA., 1997.
[4] Y. Grandvalet. Least absolute shrinkage is equivalent to quadratic penalization. In
L. Niklasson, M. Boden, and T Ziemske, editors, ICANN'98, volume 1 of Perspectives in Neural Computing, pages 201-206. Springer, 1998.
[5] W. HardIe. Applied Nonparametric Regression, volume 19 of Economic Society
Monographs. Cambridge University Press, New York, 1990.
[6] TJ. Hastie and R.J. Tibshirani. Generalized Additive Models, volume 43 of Monographs on Statistics and Applied Probability. Chapman & Hall, New York, 1990.
[7] D.J.C. MacKay. A practical Bayesian framework for backprop networks. Neural Computation, 4(3):448-472,1992.
[8] R. M. Neal. Bayesian Learning for Neural Networks. Lecture Notes in Statistics.
Springer, New York, 1996.
[9] S.I. Nowlan and G.E. Hinton. Simplifying neural networks by soft weight-sharing.
Neural Computation, 4(4):473-493, 1992.
[10] R.I. Tibshirani. Regression shrinkage and selection via the lasso. Journal ofthe Royal
Statistical Society, B, 58(1):267-288, 1995.
| 1500 |@word trial:1 version:1 norm:1 simplifying:1 solid:1 harder:1 initial:1 configuration:2 series:2 current:1 unction:1 nowlan:1 additive:9 numerical:1 girosi:2 interpretable:1 update:1 parameterization:1 provides:1 node:2 five:2 backfitting:1 fitting:4 expected:1 utc:1 automatically:2 becomes:1 underlying:1 linearity:1 maximizes:2 lowest:2 argmin:2 unspecified:1 interpreted:2 loos:1 multidimensional:1 xd:1 lys:1 control:3 unit:4 positive:2 ice:1 local:1 treat:1 consequence:1 yd:1 plus:2 umr:1 twice:1 equivalence:8 suggests:1 co:3 practical:1 lf:2 bootstrap:3 procedure:5 empirical:1 refers:1 suggest:1 get:1 selection:6 operator:3 applying:1 instability:1 equivalent:6 lagrangian:2 penalizer:2 exposure:1 go:6 independently:2 hds:1 stability:1 coordinate:1 pek:1 controlling:1 annals:1 heavily:2 user:1 programming:1 us:2 trend:2 expensive:1 solved:1 capture:1 decrease:1 monograph:2 complexity:8 covariates:10 technologie:1 depend:1 basis:2 easily:5 differently:1 represented:2 fast:1 effective:2 detected:1 hyper:8 outcome:2 heuristic:1 solve:2 cov:1 statistic:3 itself:1 final:2 net:2 propose:1 fr:1 relevant:3 motorcycle:2 achieve:1 cl1:1 rff:2 cluster:2 convergence:4 adap:1 produce:4 leave:1 bve:1 derive:1 measured:1 ij:5 quantify:1 stephane:1 modifying:1 centered:3 stabilization:1 backprop:1 require:1 generalization:2 extension:2 hold:1 practically:1 hall:1 normal:3 algorithmic:1 predict:1 mapping:2 estimation:3 ofpossible:1 individually:1 minimization:2 gaussian:2 modified:1 avoid:1 shrinkage:12 breiman:2 varying:1 heudiasyc:1 vk:1 rank:1 likelihood:3 rigorous:1 sense:3 posteriori:1 dependent:1 cnrs:1 lj:1 hidden:1 france:1 issue:1 overall:2 priori:1 smoothing:9 special:1 mackay:1 equal:1 once:1 f3:3 eliminated:1 chapman:1 future:1 minimized:2 report:1 spline:7 few:2 individual:2 replaced:1 mlp:2 highly:1 ziemske:1 evaluation:1 replicates:1 weakness:1 mixture:2 tj:1 predefined:1 accurate:1 encourage:1 partial:1 necessary:2 initialized:1 penalizes:1 plotted:1 fitted:1 modeling:6 soft:5 planned:1 ar:3 ordinary:1 cost:4 deviation:1 subset:1 predictor:4 uniform:1 dependency:3 regressor:1 corner:1 derivative:1 li:1 de:2 pooled:4 coefficient:11 explicitly:1 ad:1 depends:1 performed:1 linked:1 bayes:2 minimize:2 yves:2 square:3 accuracy:4 variance:3 yield:1 ofthe:1 bayesian:2 simultaneous:1 sharing:1 definition:1 obvious:1 universite:1 proved:1 dataset:1 recall:1 f3j:2 cj:3 rif:1 higher:2 dt:1 supervised:1 response:4 evaluated:1 shrink:1 hand:1 hardie:1 aj:3 effect:1 normalized:1 multiplier:1 true:1 former:1 hence:5 laboratory:1 orthe:1 neal:1 illustrated:1 during:1 noted:1 generalized:1 compiegne:2 stress:1 ridge:41 performs:1 l1:1 fj:4 ols:7 niklasson:1 functional:1 qp:1 volume:3 interpretation:1 significant:1 cambridge:2 ai:2 smoothness:2 tuning:8 automatic:2 unconstrained:2 canu:4 similarly:1 stable:1 operating:1 j:1 posterior:1 own:1 showed:1 perspective:1 unsmooth:1 yi:3 seen:1 minimum:1 preceding:1 determine:2 converge:1 ii:2 smoother:3 multiple:1 asso:1 technical:1 faster:1 smooth:1 determination:2 cross:2 f3o:1 equally:1 y:1 a1:1 ensuring:1 j3:4 prediction:7 regression:14 kernel:9 penalize:1 cedex:1 pooling:1 induced:1 subject:2 tend:2 identically:1 easy:1 variety:1 affect:1 fit:3 hastie:1 lasso:12 f3i:1 economic:1 regarding:1 prototype:2 whether:1 motivated:1 six:1 york:3 jj:1 useful:1 clear:1 tune:1 amount:2 nonparametric:1 extensively:1 ten:1 statist:1 reduced:1 generate:1 xij:4 exist:1 dotted:1 sign:1 estimated:3 tibshirani:4 per:2 hyperparameter:1 group:2 achieving:2 drawn:2 clarity:1 pj:1 sum:1 bound:1 layer:2 fold:1 quadratic:11 encountered:1 badly:1 constraint:9 infinity:2 bp:1 ables:1 optimality:1 performing:1 according:2 smaller:1 explicative:2 modification:2 s1:4 explained:2 invariant:1 computationally:1 equation:1 turn:1 know:1 fed:1 gaussians:1 appropriate:1 alternative:1 remaining:1 include:1 plj:1 prof:1 especially:1 classical:1 garrotte:4 society:2 parametric:1 dependence:1 usual:1 diagonal:1 link:1 simulated:2 sensible:1 me:12 unstable:1 besides:1 index:1 remind:1 balance:3 minimizing:1 statement:2 negative:3 design:1 observation:2 benchmark:1 extended:1 variability:1 hinton:1 pair:1 required:1 connection:1 optimized:1 yll:1 usually:1 fp:8 summarize:1 yi12:1 royal:1 natural:2 rely:1 residual:1 minimax:1 improve:1 imply:1 prior:7 f3d:1 loss:5 lecture:1 interesting:1 proportional:4 analogy:2 var:1 penalization:16 validation:2 gather:1 thresholding:4 viewpoint:1 editor:1 grandvalet:6 balancing:1 penalized:4 last:1 jth:1 understand:1 johnstone:1 wide:1 absolute:10 sparse:3 benefit:1 f31:1 curve:1 dimension:1 adaptive:27 regressors:1 simplified:1 pruning:1 global:2 incoming:2 xi:2 alternatively:1 search:1 timation:1 reasonably:1 correlated:1 boden:1 interpolating:1 complex:1 diag:4 did:1 icann:1 main:2 spread:1 multilayered:1 noise:1 fig:3 cubic:3 vr:1 wish:1 sf:2 xl:1 candidate:1 wavelet:2 formula:2 xt:2 covariate:3 divergent:1 svm:3 decay:1 concern:1 dl:1 ofthis:1 exists:1 effectively:1 conditioned:1 easier:1 lt:1 univariate:4 lagrange:1 springer:2 ch:1 outcoming:2 ijf:1 ma:1 viewed:1 goal:2 identity:1 donoho:1 ann:1 towards:1 included:1 corrected:1 e:1 perceptrons:1 select:2 support:4 latter:3 relevance:4 phenomenon:1 avoiding:1 ex:2 |
549 | 1,501 | Learning curves for Gaussian processes
Peter Sollich *
Department of Physics, University of Edinburgh
Edinburgh EH9 3JZ, U.K. Email: P.Sollich<Oed.ac . uk
Abstract
I consider the problem of calculating learning curves (i.e., average
generalization performance) of Gaussian processes used for regression. A simple expression for the generalization error in terms of
the eigenvalue decomposition of the covariance function is derived,
and used as the starting point for several approximation schemes.
I identify where these become exact, and compare with existing
bounds on learning curves; the new approximations, which can
be used for any input space dimension, generally get substantially
closer to the truth.
1
INTRODUCTION: GAUSSIAN PROCESSES
Within the neural networks community, there has in the last few years been a
good deal of excitement about the use of Gaussian processes as an alternative to
feedforward networks [lJ. The advantages of Gaussian processes are that prior
assumptions about the problem to be learned are encoded in a very transparent
way, and that inference-at least in the case of regression that I will consider-is
relatively straightforward. One crucial question for applications is then how 'fast'
Gaussian processes learn, i.e., how many training examples are needed to achieve a
certain level of generalization performance. The typical (as opposed to worst case)
behaviour is captured in the learning curve, which gives the average generalization
error ? as a function of the number of training examples n. Several workers have
derived bounds on ?(n) [2,3, 4J or studied its large n asymptotics. As I will illustrate
below, however, the existing bounds are often far from tight; and asymptotic results
will not necessarily apply for realistic sample sizes n. My main aim in this paper
is therefore to derive approximations to ?( n) which get closer to the true learning
curves than existing bounds, and apply both for small and large n.
In its simplest form, the regression problem that I am considering is this: We are
trying to learn a function 0* which maps inputs x (real-valued vectors) to (realvalued scalar) outputs O*(x) . We are given a set of training data D, consisting of n
'Present address: Department of Mathematics, King's College London, Strand, London
WC2R 2LS, U.K. Email peter.sollicMlkcl.ac . uk
345
Learning Curves for Gaussian Processes
input-output pairs (Xl, yt); the training outputs Yl may differ from the 'clean' target
outputs 9* (xL) due to corruption by noise. Given a test input x, we are then asked
to come up with a prediction 9(x) for the corresponding output, expressed either in
the simple form of a mean prediction 9(x) plus error bars, or more comprehensively
in terms of a 'predictive distribution' P(9(x)lx, D). In a Bayesian setting, we do this
by specifying a prior P(9) over our hypothesis functions, and a likelihood P(DI9)
with which each 9 could have generated the training data; from this we deduce the
posterior distribution P(9ID) ex P(DI9)P(9). In the case of feedforward networks,
where the hypothesis functions 9 are parameterized by a set of network weights, the
predictive distribution then needs to be extracted by integration over this posterior,
either by computationally intensive Monte Carlo techniques or by approximations
which lead to analytically tractable integrals. For a Gaussian process, on the other
hand, obtaining the predictive distribution is trivial (see below); one reason for
this is that the prior P(9) is defined directly over input-output functions 9. How
is this done? Any 9 is uniquely determined by its output values 9(x) for all x
from the input domain, and for a Gaussian process, these are simply assumed to
have a joint Gaussian distribution (hence the name). This distribution can be
specified by the mean values (9(x))o (which I assume to be zero in the following,
as is commonly done), and the covariances (9(x)9(x ' ))o = C(x, x'); C(x, x') is
called the covariance function of the Gaussian process. It encodes in an easily
interpretable way prior assumptions about the function to be learned. Smoothness,
for example, is controlled by the behaviour of C(x, x') for x' -+ x: The OrnsteinUhlenbeck (OU) covariance function C(x, x') ex exp( -IX-X'l/l) produces very rough
(non-differentiable) functions, while functions sampled from the squared exponential
(SE) prior with C(X,X') ex exp(-Ix - x ' 12/(2l2)) are infinitely differentiable. The
'length scale' parameter l, on the other hand, corresponds directly to the distance in
input space over which we expect our function to vary significantly. More complex
properties can also be encoded; by replacing l with different length scales for each
input component, for example, relevant (smalll) and irrelevant (large l) inputs can
be distinguished.
How does inference with Gaussian processes work? I only give a brief summary here
and refer to existing reviews on the subject (see e.g. [5, 1]) for details. It is simplest
to assume that outputs yare generated from the 'clean' values of a hypothesis
function 9(x) by adding Gaussian noise of x-independent variance 0'2. The joint
distribution of a set of training outputs {yd and the function values 9(x) is then
also Gaussian, with covariances given by
here I have defined an n x n matrix K and x-dependent n-component vectors k(x).
The posterior distribution P(9ID) is then obtained by simply conditioning on the
{Yl}. It is again Gaussian and has mean and variance
(B(x))OID
( (9(x) - 9(X))2)
9(x)
=
k(X)TK-1y
C(x, x) - k(X)TK-lk(x)
(1)
(2)
OlD
Eqs. (1,2) solve the inference problem for Gaussian process: They provide us directly
with the predictive distribution P(9(x)lx, D). The posterior variance, eq. (2), in
fact also gives us the expected generalization error at x. Why? If the teacher
is 9*, the squared deviation between our mean prediction and the teacher output
is 1 (9(x) - 9* (X))2; averaging this over the posterior distribution of teachers P(9* ID)
just gives (2). The underlying assumption is that our assumed Gaussian process
lOne can also one measure the generalization by the squared deviation between the
prediction O(x) and the noisy teacher output; this simply adds a term 0'2 to eq. (3).
P. Sollich
346
prior is the true one from which teachers are actually generated (and that we are
using the correct noise model). Otherwise, a more complicated expression for the
expected generalization error results; in line with most other work on the subject, I
only consider the 'correct prior' case in the following. Averaging the generalization
error at x over the distribution of inputs gives then
(3)
This form of the generalization error (which is well known [2, 3, 4, 5]) still depends
on the training inputs (the fact that the training outputs have dropped out already
is a signature of the fact that Gaussian processes are linear predictors, compare (1)).
Averaging over data sets yields the quantity we are after,
?
(4)
= (t(D)}D?
This average expected generalization error (I will drop the 'average expected' in the
following) only depends on the number of training examples n; the function ?(n)
is called the learning curve. Its exact calculation is difficult because of the joint
average in eqs. (3,4) over the training inputs Xl and the test input x.
2
LEARNING CURVES
As a starting point for an approximate calculation of ?(n), I first derive a representation of the generalization error in terms of the eigenvalue decomposition of
the covariance function. Mercer's theorem (see e.g. [6]) tells us that the covariance
function can be decomposed into its eigenvalues Ai and eigenfunctions (/Ji(x):
00
C(x, x') =
L Ai<Pi(X)qJi(x')
(5)
i=1
This is simply the analogue of the eigenvalue decomposition of a finite symmetric
matrix; the eigenfunctions can be taken to be normalized such that (<Pi (x) <Pj (x)} x =
Oij . Now write the data-dependent generalization error (3) as ?(D) = (C(x,x)}xtr (k(x)k(x)T)x K- 1 and perform the x-average in the second term:
?(k(x)k(x)T)lm)x =
L AiAj<Pi(XI) (<Pi (x)<pj (x)} <pj(xm) = L A;<Pi (Xt}<Pi (x
ij
m)
i
This suggests introducing the diagonal matrix (A)ij = AiOij and the 'design matrix'
(<J?li = <Pi (xt), so that (k(x)k(x)T)x = <J>A2<J>T. One then also has (C(x,x)}x =
tr A, and the matrix K is expressed as K = a 21 + <J>A<J>T, 1 being the identity
matrix. Collecting these results, we have
?(D)
= tr A -
tr (a 21 + <J>A<J>T)-I<J>A 2<J>T
This can be simplified using the Woodbury formula for matrix inverses (see e.g. [7]),
which applied to our case gives (a 2I+<J>A<J>T)-1 = a- 2[I-<J>(a 21+A<J>T<J?-1 A<J>TJ;
after a few lines of algebra, one then obtains the final result
t= (t(D))D'
?(D) =tra 2A(a 21+A<J>T<J?-1 =tr(A- 1 +a- 2 <J>T<J?-1
(6)
This exact representation of the generalization error is one of the main results of this
paper. Its advantages are that the average over the test input X has already been
carried out, and that the remainingf dependence on the training data is contained
entirely in the matrix <J> T <J>. It also includes as a special case the well-known result
for linear regression (see e.g. [8]); A-I and <J> T <J> can be interpreted as suitably
generalized versions of the weight decay (matrix) and input correlation matrix.
Starting from (6), one can now derive approximate expressions for the learning
347
Learning Curves for Gaussian Processes
curve I:(n). The most naive approach is to entirely neglect the fluctuations in cJ>TcJ>
over different data sets and replace it by its average, which is simply (( cJ> T cJ> )ij ) D =
I:l (?i(Xt)?j(XI)) D = n8ij . This leads to the Naive approximation
I:N(n) = tr (A -1 + O'- 2 nI)-1
(7)
which is not, in general, very good. It does however become exact in the large noise
limit 0'2 -t 00 at constant nlO' 2 : The fluctuations of the elements of the matrix
O'- 2cJ>TcJ> then become vanishingly small (of order foO'- 2 = (nlO' 2 )/fo -t 0) and
so replacing cJ> T cJ> by its average is justified.
To derive better approximations, it is useful to see how the matrix 9 = (A -1 +
O'- 2cJ>TcJ?-1 changes when a new example is added to the training set. One has
9(n
+ 1) -
9(n)
= [9- 1 (n) + O'- 2 1j11j1 T
r
l -
9(n)
=_
9(n)1jI1jI T 9(n)
+ 1jIT 9(n)1jI
(8)
0'2
in terms of the vector 1jI with elements (1jI)i = ?i(Xn+I); the second identity uses
again the Woodbury formula. To get exact learning curves, one would have to average this update formula over both the new training input Xn+1 and all previous
ones. This is difficult, but progress can be made by again neglecting some fluctuations: The average over Xn +1 is approximated by replacing 1jI1jIT by its average,
which is simply the identity matrix; the average over the previous training inputs
by replacing 9(n) by its average G(n) = (9(n)) D' This yields the approximation
G 2 (n)
G(n + 1) - G(n) = - 2
G()
(9)
a +tr n
Iterating from G(n = 0) = A, one sees that G(n) remains diagonal for all n, and
so (9) is trivial to implement numerically. I call the resulting I:D(n) = tr G(n) the
Discrete approximation to the learning curve, because it still correctly treats n as
a variable with discrete, integer values. One can further approximate (9) by taking
n as continuously varying, replacing the difference on the left-hand side by the
derivative dG( n) 1dn. The resulting differential equation for G( n) is readily solved;
taking the trace, one obtains the generalization error
I:uc(n) = tr (A -1 + O'- 2 n'I)-1
(10)
with n' determined by the self-consistency equation n' + tr In(I + O'- 2 n' A) = n.
By comparison with (7), n' can be thought of as an 'effective number of training
examples'. The subscript DC in (10) stands for Upper Continuous approximation.
As the name suggests, there is another, lower approximation also derived by treating
n as continuous. It has the same form as (10), but a different self-consistent equation
for n', and is derived as follows. Introduce an auxiliary offset parameter v (whose
usefulness will become clear shortly) by 9- 1 = vI+A -1 +O'- 2cJ>TcJ>; at the end ofthe
calculation, v will be set to zero again. As before, start from (8)-which also holds
for nonzero v-and approximate 1jI1jIT and tr 9 by their averages, but retain possible
fluctuations of 9 in the numerator. This gives G(n+ 1) - G(n) = - (9 2 (n)) 1[0'2 +
tr G(n)]. Taking the trace yields an update formula for the generalization error 1:,
where the extra parameter v lets us rewrite the average on the right-hand side as
-tr (9 2 ) = (olov)tr (9) = ol:lov. Treating n again as continuous, we thus arrive
at the partial differential equation Eh{on = (oI:l ov) 1 (0'2 + 1:). This can be solved
using the method of characteristics [8 and (for v = 0) gives the Lower Continuous
approximation to the learning curve,
I:Lc(n)
= tr (A -1 + O'- 2 n'I)-1 ,
n'
=
nO' 2
0'2
+ I:LC
(11)
By comparing derivatives w.r.t. n, it is easy to show that this is always lower than
the DC approximation (10). One can also check that all three approximations that
I have derived (D, LC and DC) converge to the exact result (7) in the large noise
limit as defined above.
P. Sol/ich
348
3
COMPARISON WITH BOUNDS AND SIMULATIONS
I now compare the D, LC and UC approximations with existing bounds, and with
the 'true' learning curves as obtained by simulations. A lower bound on the generalization error was given by Michelli and Wahba [2J as
?(n) ~ ?Mw(n) = 2::n+l Ai
This is derived for the noiseless
jections of 0* (x) along the first
to be tight for the case of 'real'
information theoretic methods, a
(12)
case by allowing 'generalized observations' (pron eigenfunctions of C (x, x') ), and so is unlikely
observations at discrete input points. Based on
different Lower bound was obtained by Opper [3J:
1
?(n) ~ ?Lo(n) = 4"tr (A -1
+ 2a- 2 nl)-1
x [I + (I
+ 2a- 2 nA)-lJ
This is always lower than the naive approximation (7); both incorrectly suggest that
? decreases to zero for a 2 -+ 0 at fixed n, which is clearly not the case (compare (12)).
There is also an Upper bound due to Opper [3J,
i(n) ~ ?uo(n)
= (a- 2 n)-1 tr In(1 + a- 2 nA) + tr (A -1 + a- 2 nl)-1
(13)
Here i is a modified version of ? which (in the rescaled version that I am using)
becomes identical to ? in the limit of small generalization errors (? ? a 2 ), but never
gets larger that 2a 2 ; for small n in particular, ?(n) can therefore actually be much
larger than i(n) and its bound (13). An upper bound on ?(n) itself was derived
by Williams and Vivarelli [4J for one-dimensional inputs and stationary covariance
functions (for which C(x, x') is a function of x - x' alone). They considered the
generalization error at x that would be obtained from each individual training example, and then took the minimum over all n examples; the training set average
of this 'lower envelope' can be evaluated explicitly in terms of integrals over the
covariance function [4J. The resulting upper bound, ?wv(n), never decays below a 2
and therefore complements the range of applicability of the UO bound (13).
In the examples in Fig. 1, I consider a very simple input domain, x E [0, 1Jd,
with a uniform input distribution. I also restrict myself to stationary covariance
functions, and in fact I use what physicists call periodic boundary conditions. This
is simply a trick that makes it easy to calculate the required eigenvalue spectra of
the covariance function, but otherwise has little effect on the results as long as the
length scale of the covariance function is smaller than the size of the input domain 2 ,
l ? 1. To cover the two extremes of 'rough' and 'smooth' Gaussian priors, I
consider the OU [C(x, x') = exp( -lx-xll/l)J and SE [C(x, x') = exp( -lx-x' 12 /2l 2 )J
covariance functions. The prior variance of the values of the function to be learned is
simply C (x, x) = 1; one generically expects this 'prior ignorance' to be significantly
larger than the noise on the training data, so I only consider values of a 2 < 1.
I also fix the covariance function length scale to l = 0.1; results for l = 0.01
are qualitatively similar. Several observations can be made from Figure 1. (1)
The MW lower bound is not tight, as expected. (2) The bracket between Opper's
lower and upper bounds (LO /UO) is rather wide (1-2 orders of magnitude); both
give good representations of the overall shape of the learning curve only in the
asymptotic regime (most clearly visible for the SE covariance function), i. e., once
? has dropped below a 2 . (3) The WV upper bound (available only in d = 1) works
21n d = 1 dimension, for example, a 'periodically continued' stationary covariance
function on [0,1] can be written as C(X,X') = 2:::_ooc(x - x' + r). For I ? 1, only
the r = 0 term makes a significant contribution, except when x and x' are within ~ I of
opposite ends of the input space. With this definition, the eigenvalues of C(x, x') are given
dx c(x) exp( -2rriqx), for integer q.
by the Fourier transform
1:"00
349
Learning Curves for Gaussian Processes
2
OU, d=l, 1=0.1 , cr =10
-3
2
-3
SE, d=l, 1=0.1, cr =10
10?
10?
E
(b)
10-1
10- 1
-_WV
10-2
10-3
10-2
\.
MW--y~ - -
--
' ,L?
,\
'i- ,
10-4
, - __
'MW ---
\
10-3
10-5
0
200
400
600
~O
50
0
2
100
200
150
2
OU, d=l, 1=0.1, cr =0.1
SE, d=l , 1=0.1, cr =0.1
10?
10?
(c)
E
___
10- 1
-~~-::::.-::::.-
10-1
D/uC
,
\\~
WV
10-2
--VO
LC
''''',---
0
,- - - - - - - - - - - --- - - _ _ -VO- ,
10-3
'-.\.
10-2
'-.
(d)
wv
- _ _ -l-O
IMW
- ..!-O-
-
- --
10-4
200
400
2
OU, d=2, 1=0.1, cr =10
600
0
200
-3
400
2
600
-3
SE, d=2, 1=0.1, cr =10
10?
10?
(e)
E
\.
'-.
-- -- - -~
LC
vo--- ---
10- 1
\
0
10-2
10-3
, - ___ ~o
200
\.
-
n
400
- -- 600
'.MW
,
,
\
,
\~---,~-
10-4
-- -
10-2
(D
10-1
DIVC
\
10-5
0
200
n
400
600
Figure 1: Learning curves c(n): Comparison of simulation results (thick solid lines;
the small fluctuations indicate the order of magnitude of error bars), approximations
derived in this paper (thin solid lines; D = discrete, UC/LC = upper/lower continuous) , and existing upper (dashed; UO = upper Opper, WV = Williams-Vivarelli)
and lower (dot-dashed; LO = lower Opper, MW = Michelli-Wahba) bounds. The
type of covariance function (Ornstein-Uhlenbeck/Squared Exponential), its length
scale l, the dimension d of the input space, and the noise level (72 are as shown.
Note the logarithmic y-axes. On the scale of the plots, D and UC coincide (except
in (b)); the simulation results are essentially on top of the LC curve in (c-e) .
350
P'Sollich
well for the OU covariance function, but less so for the SE case. As expected, it
is not useful in the asymptotic regime because it always remains above (72. (4)
The discrete (D) and upper continuous (UC) approximations are very similar, and
in fact indistinguishable on the scale of most plots. This makes the UC version
preferable in practice, because it can be evaluated for any chosen n without having
to step through all smaller values of n. (5) In all the examples, the true learning
curve lies between the UC and LC curves. In fact I would conjecture that these two
approximations provide upper and lower bounds on the learning curves, at least
for stationary covariance functions. (6) Finally, the LC approximation comes out
as the clear winner: For (72 = 0.1 (Fig. 1c,d), it is indistinguishable from the true
learning curves. But even in the other cases it represents the overall shape of the
learning curves very well, both for small n and in the asymptotic regime; the largest
deviations occur in the crossover region between these two regimes.
In summary, I have derived an exact representation of the average generalization c
error of Gaussian processes used for regression, in terms of the eigenvalue decomposition of the covariance function. Starting from this, I have obtained three different
approximations to the learning curve c(n) . All of them become exact in the large
noise limit; in practice, one generically expects the opposite case ((72 /C(x, x) ? 1),
but comparison with simulation results shows that even in this regime the new
approximations perform well. The LC approximation in particular represents the
overall shape of the learning curves very well, both for 'rough' (OU) and 'smooth'
(SE) Gaussian priors, and for small as well as for large numbers of training examples
n. It is not perfect, but does get substantially closer to the true learning curves
than existing bounds. Future work will have to show how well the new approximations work for non-stationary covariance functions and/or non-uniform input
distributions, and whether the treatment of fluctuations in the generalization error
(due to the random selection of training sets) can be improved, by analogy with
fluctuation corrections in linear perceptron learning [8].
Acknowledgements: I would like to thank Chris Williams and Manfred Opper for
stimulating discussions, and for providing me with copies of their papers [3,4] prior
to publication. I am grateful to the Royal Society for financial support through a
Dorothy Hodgkin Research Fellowship.
[1] See e.g. D J C MacKay, Gaussian Processes, Tutorial at NIPS 10, and recent papers
by Goldberg/Williams/Bishop (in NIPS 10), Williams and Barber/Williams (NIPS 9),
Williams/Rasmussen (NIPS 8).
[2] C A Michelli and G Wahba. Design problems for optimal surface interpolation. In
Z Ziegler, editor, Approximation theory and applications, pages 329-348. Academic
Press, 1981.
[3] M Opper. Regression with Gaussian processes: Average case performance. In I K
Kwok-Yee, M Wong, and D-Y Yeung, editors, Theoretical Aspects of Neural Computation: A Multidisciplinary Perspective. Springer, 1997.
[4] C K I Williams and F Vivarelli. An upper bound on the learning curve for Gaussian
processes. Submitted for publication.
[5] C K I Williams. Prediction with Gaussian processes: From linear regression to linear
prediction and beyond. In M I Jordan, editor, Learning and Inference in Graphical
Models. Kluwer Academic. In press.
[6] E Wong. Stochastic Processes in Information and Dynamical Systems. McGraw-Hill,
New York, 1971.
[7] W H Press, S A Teukolsky, W T Vetterling, and B P Flannery. Numerical Recipes in
C (2nd ed.). Cambridge University Press, Cambridge, 1992.
[8] P Sollich. Finite-size effects in learning and generalization in linear perceptrons. Journal of Physics A, 27:7771- 7784, 1994.
| 1501 |@word version:4 nd:1 suitably:1 simulation:5 decomposition:4 covariance:21 tr:17 solid:2 existing:7 comparing:1 dx:1 written:1 readily:1 visible:1 periodically:1 numerical:1 realistic:1 shape:3 drop:1 interpretable:1 update:2 treating:2 plot:2 stationary:5 alone:1 manfred:1 lx:4 dn:1 along:1 become:5 differential:2 introduce:1 lov:1 expected:6 ol:1 decomposed:1 little:1 considering:1 becomes:1 underlying:1 what:1 interpreted:1 substantially:2 lone:1 collecting:1 preferable:1 uk:2 uo:4 before:1 dropped:2 treat:1 limit:4 physicist:1 id:3 subscript:1 fluctuation:7 interpolation:1 yd:1 plus:1 studied:1 specifying:1 suggests:2 range:1 woodbury:2 practice:2 implement:1 ooc:1 asymptotics:1 crossover:1 significantly:2 thought:1 suggest:1 get:5 selection:1 yee:1 wong:2 map:1 yt:1 straightforward:1 williams:9 starting:4 l:1 continued:1 financial:1 aiaj:1 target:1 exact:8 us:1 goldberg:1 hypothesis:3 trick:1 element:2 approximated:1 solved:2 worst:1 calculate:1 region:1 sol:1 decrease:1 rescaled:1 asked:1 oed:1 signature:1 ov:1 tight:3 rewrite:1 algebra:1 grateful:1 predictive:4 tcj:4 easily:1 joint:3 fast:1 effective:1 london:2 monte:1 tell:1 whose:1 encoded:2 larger:3 valued:1 solve:1 otherwise:2 transform:1 noisy:1 itself:1 final:1 advantage:2 eigenvalue:7 differentiable:2 took:1 vanishingly:1 relevant:1 achieve:1 recipe:1 produce:1 perfect:1 tk:2 illustrate:1 derive:4 ac:2 ij:3 progress:1 eq:4 auxiliary:1 imw:1 come:2 indicate:1 differ:1 thick:1 correct:2 stochastic:1 myself:1 wc2r:1 behaviour:2 transparent:1 generalization:21 fix:1 correction:1 hold:1 considered:1 exp:5 lm:1 vary:1 a2:1 ziegler:1 largest:1 rough:3 xtr:1 clearly:2 gaussian:27 always:3 aim:1 modified:1 rather:1 cr:6 varying:1 publication:2 derived:9 ax:1 likelihood:1 check:1 am:3 inference:4 dependent:2 vetterling:1 lj:2 unlikely:1 overall:3 integration:1 special:1 uc:8 mackay:1 once:1 never:2 having:1 identical:1 represents:2 thin:1 future:1 few:2 dg:1 individual:1 consisting:1 generically:2 extreme:1 nl:2 bracket:1 tj:1 integral:2 closer:3 worker:1 neglecting:1 partial:1 old:1 michelli:3 theoretical:1 cover:1 applicability:1 introducing:1 deviation:3 expects:2 predictor:1 usefulness:1 uniform:2 teacher:5 periodic:1 my:1 retain:1 physic:2 yl:2 continuously:1 na:2 squared:4 again:5 opposed:1 derivative:2 li:1 includes:1 tra:1 explicitly:1 depends:2 vi:1 pron:1 ornstein:1 start:1 di9:2 complicated:1 contribution:1 oi:1 ni:1 variance:4 characteristic:1 yield:3 identify:1 ofthe:1 bayesian:1 carlo:1 corruption:1 submitted:1 fo:1 ed:1 email:2 definition:1 sampled:1 treatment:1 cj:8 ou:7 actually:2 improved:1 done:2 evaluated:2 just:1 correlation:1 hand:4 replacing:5 multidisciplinary:1 name:2 effect:2 normalized:1 true:6 analytically:1 hence:1 symmetric:1 nonzero:1 deal:1 ignorance:1 indistinguishable:2 numerator:1 self:2 uniquely:1 generalized:2 trying:1 hill:1 theoretic:1 vo:3 oid:1 ornsteinuhlenbeck:1 ji:4 conditioning:1 winner:1 kluwer:1 numerically:1 refer:1 significant:1 cambridge:2 ai:3 smoothness:1 consistency:1 mathematics:1 dot:1 surface:1 deduce:1 add:1 posterior:5 recent:1 perspective:1 irrelevant:1 certain:1 wv:6 captured:1 minimum:1 converge:1 dashed:2 smooth:2 academic:2 calculation:3 nlo:2 long:1 controlled:1 prediction:6 regression:7 noiseless:1 essentially:1 yeung:1 uhlenbeck:1 justified:1 fellowship:1 dorothy:1 crucial:1 extra:1 envelope:1 eigenfunctions:3 subject:2 jordan:1 call:2 integer:2 mw:6 feedforward:2 easy:2 wahba:3 restrict:1 opposite:2 intensive:1 whether:1 expression:3 peter:2 york:1 generally:1 useful:2 se:8 iterating:1 clear:2 simplest:2 tutorial:1 correctly:1 write:1 discrete:5 pj:3 clean:2 year:1 inverse:1 parameterized:1 ich:1 hodgkin:1 arrive:1 eh9:1 bound:20 entirely:2 occur:1 encodes:1 fourier:1 aspect:1 relatively:1 conjecture:1 department:2 smaller:2 sollich:5 taken:1 computationally:1 equation:4 remains:2 vivarelli:3 excitement:1 needed:1 tractable:1 end:2 available:1 apply:2 yare:1 kwok:1 distinguished:1 alternative:1 shortly:1 jd:1 top:1 graphical:1 qji:1 calculating:1 neglect:1 society:1 question:1 already:2 quantity:1 added:1 dependence:1 diagonal:2 distance:1 thank:1 chris:1 me:1 barber:1 trivial:2 reason:1 jit:1 length:5 providing:1 difficult:2 trace:2 design:2 xll:1 perform:2 allowing:1 upper:12 observation:3 finite:2 incorrectly:1 dc:3 community:1 complement:1 pair:1 required:1 specified:1 learned:3 nip:4 address:1 beyond:1 bar:2 below:4 dynamical:1 xm:1 regime:5 royal:1 analogue:1 eh:1 oij:1 scheme:1 brief:1 realvalued:1 lk:1 carried:1 naive:3 prior:12 review:1 l2:1 acknowledgement:1 asymptotic:4 expect:1 analogy:1 consistent:1 mercer:1 editor:3 pi:7 lo:3 summary:2 last:1 copy:1 rasmussen:1 side:2 perceptron:1 wide:1 comprehensively:1 taking:3 edinburgh:2 curve:27 dimension:3 xn:3 stand:1 opper:7 boundary:1 commonly:1 made:2 qualitatively:1 simplified:1 coincide:1 far:1 approximate:4 obtains:2 mcgraw:1 assumed:2 xi:2 spectrum:1 continuous:6 why:1 jz:1 learn:2 obtaining:1 necessarily:1 complex:1 domain:3 main:2 noise:8 fig:2 foo:1 lc:11 exponential:2 xl:3 lie:1 ix:2 theorem:1 formula:4 xt:3 bishop:1 offset:1 decay:2 adding:1 magnitude:2 flannery:1 logarithmic:1 simply:8 infinitely:1 expressed:2 strand:1 contained:1 scalar:1 springer:1 corresponds:1 truth:1 extracted:1 teukolsky:1 stimulating:1 identity:3 king:1 replace:1 change:1 typical:1 determined:2 except:2 averaging:3 called:2 perceptrons:1 college:1 support:1 ex:3 |
550 | 1,502 | Adding Constrained Discontinuities to Gaussian
Process Models of Wind Fields
Ian T. Nabney
Christopher K. I. Williams t
Neural Computing Research Group
Aston University, BIRMINGHAM, B4 7ET, UK
d.comford@aston.ac.uk
Dan Cornford*
Abstract
Gaussian Processes provide good prior models for spatial data, but can
be too smooth. In many physical situations there are discontinuities
along bounding surfaces, for example fronts in near-surface wind fields.
We describe a modelling method for such a constrained discontinuity
and demonstrate how to infer the model parameters in wind fields with
MCMC sampling.
1 INTRODUCTION
We introduce a model for wind fields based on Gaussian Processes (GPs) with 'constrained
discontinuities'. GPs provide a flexible framework for modelling various systems. They
have been adopted in the neural network community and are interpreted as placing priors
over functions.
Stationary vector-valued GP models (Daley, 1991) can produce realistic wind fields when
run as a generative model; however, the resulting wind fields do not contain some features
typical of the atmosphere. The most difficult features to include are surface fronts. Fronts
are generated by complex atmospheric dynamics and are marked by large changes in the
surface wind direction (see for example Figures 2a and 3b) and temperature. In order
to account for such features, which appear discontinuous at our observation scale, we have
developed a model for vector-valued GPs with constrained discontinuities which could also
be applied to surface reconstruction in computer vision, and geostatistics.
In section 2 we illustrate the generative model for wind fields with fronts. Section 3 explains what we mean by GPs with constrained discontinuities and derives the likelihood of
data under the model. Results of Bayesian estimation of the model parameters are given,
?To whom correspondence should be addressed.
tNowat: Division of Informatics, University of Edinburgh, 5 Forrest Hill, Edinburgh EHI 2QL,
Scotland, UK
D. Com/ord, I. T. Nabney and C. K. I. Williams
862
using a Markov Chain Monte Carlo (MCMC) procedure. In the final section, the strengths
and weaknesses of the model are discussed and improvements suggested.
2 A GENERATIVE WIND FIELD MODEL
We are primarily interested in retrieving wind fields from satellite scatterometer observations of the ocean surface!. A probabilistic prior model for wind fields will be used in
a Bayesian procedure to resolve ambiguities in local predictions of wind direction. The
generative model for a wind field including a front is taken to be a combination of two
vector-valued GPs with a constrained discontinuity.
A common method for representing wind fields is to put GP priors over the velocity potential ~ and stream function 'It, assuming the processes are uncorrelated (Daley, 1991). The
horizontal wind vector u = (u, v) can then be derived from:
8'lt
8y
8~
u=--+-,
(1)
8x
This produces good prior models for wind fields when a suitable choice of covariance
function for ~ and 'It is made. We have investigated using a modified Bessel function
based covariance2 (Handcock and Wallis, 1994) but found, using three years of wind data
for the North Atlantic, that the maximum a posteriori value for the smoothness paramete~
in this covariance function was'" 2.5. Thus we used the correlation function:
p(r) =
(1 + .!:. + ~)
L
3L2
exp
(-.!:..)
L
(2)
where L is the correlation length scale, which is equivalent to the modified Bessel function
and less computationally demanding (Cornford, 1998).
N
Simulate Frontal Position. Orientation and Direction
Simulate Along Both Sides of Front using GPl
Simulate 'Mnd Raids Either Side of Front Conditionally
on that Sides Frontal 'Mnds using GP2
(a)
Origin
(b)
Figure 1: (a) Flowchart describing the generative frontal model. See text for full description. (b) A description of the frontal model.
The generative model has the form outlined in Figure 1a. Initially the frontal position and
orientation are simulated. They are defined by the angle clockwise from north (?/) that
the front makes and a point on the line (x/, Y/). Having defined the position of the front,
lS~
http://www.ncrg.aston.ac.uk/Projects/NEUROSAT/NEUROSAT.htm1
for details of the scatterometer work. Technical reports describing, in more detail, methods for
generating prior wind field models can also be accessed from the same page.
2The modified Bessel function allows us to control the differentiability of the sample realisations
through the 'smoothness parameter', as well as the length scales and variances.
3This varies with season, but is the most temporally stable parameter in the covariance function.
863
Adding Constrained Discontinuities to GP Models o/Wind Fields
the angle of the wind across the front (a J) is simulated from a distribution covering the
range [0,71"). This angle is related to the vertical component of vorticity ?() across the front
through ( = k? V x u ex: cos
and the constraint a J E [0,71") ensures cyclonic vorticity
at the front. It is assumed that the front bisects a J. The wind speed (8 J) is then simulated at
the front. Since there is generally little change in wind speed across the front, one value is
simulated for both sides of the front. These components 8 f = (? J , x J , YJ, a J, 8 J) define
the line of the front and the mean wind vectors just ahead of and just behind the front
(Figure Ib):
(? )
A realistic model requires some variability in wind vectors along the front. Thus we use a
GP with a non-zero mean (mla or mlb) along the line of the front. In the real atmosphere
we observe a smaller variability in the wind vectors along the line of the front compared
with regions away from fronts . Thus we use different GP parameters along the front (G Pl ),
from those used in the wind field away from the front (GP2 ), although the same GPl
parameters are used both sides of the front, just with different means. The winds just ahead
of and behind the front are assumed conditionally independent given ml a and mlb, and
are simulated at a regular 50 km spacing. The final step in the generative model is to
simulate wind vectors using G P2 in both regions either side of the front, conditionally on
the values along that side of the front. This model is flexible enough to represent fronts, yet
has the required constraints derived from meteorological principles, for example that fronts
should always be associated with cyclonic vorticity and that discontinuities at the model
scale should be in wind direction but not in wind speed4 . To make this generative model
useful for inference, we need to be able to compute the data likelihood, which is the subject
of the next section.
3
GPs WITH CONSTRAINED DISCONTINUITIES
"
. ;....
-].
1
.
!
I
D2
>
Dl
(a)
(b)
Figure 2: (a) The discontinuity in one ofthe vector components in a simulation. (b) Framework for GPs with boundary conditions. The curve Dl has nl sample points with values
Zt. The domain D2 has n2 points with values Z2.
4The model allows small discontinuities in wind speed, which are consistent with frontal dynamics.
D. Cornford, 1. T Nabney and C. K. 1. Williams
864
We consider data from two domains D1 and D2 (Figure 2b), where in this case D1 is a
curve in the plane which is intended to be the front and D2 is a region of the plane. We
obtain n1 variables Zl at points Xl along the curve, and we assume these are generated
under G P1 (a GP which depends on parameters 81 and has mean m1 = m1l which will be
determined by (3) or (4?. We are interested in determining the likelihood of the variables
Z2 observed at n2 points X2 under GP2 which depends on parameters 82, conditioned on
the 'constrained discontinuities' at the front.
We evaluate this by calculating the likelihood of Z2 conditioned on the n1 values of Zl
from G P1 along the front and marginalising out Zl:
i:
p(Z2182,81) =
p(Z2I Z 1,82,81,m1)p(ZlI81,m1) dZ1.
(5)
From the definition ofthe likelihood of a GP (Cressie, 1993) we find:
p(Z2IZ1,82,81,m1) =
~
(271") ISd'2
1
exp
2
(--21Z;'S2;lZ;)
(6)
where:
To understand the notation consider the joint distribution of Zl, Z2 and in particular its
covariance matrix:
(7)
where K 1112 is the n1 x n1 covariance matrix between the points in D1 evaluated using
8 2, K1212 = K~112 the n1 x n2 (cross) covariance matrix between the points in D1 and D2
evaluated using 8 2 and K2212 is the usual n2 x n2 covariance for points in D 2. Thus we
can see that S22 is the n2 x n2 modified covariance for the points in D2 given the points
along D 1 , while the Z; is the corrected mean that accounts for the values at the points in
D 1 ? which have non-zero mean.
We remove the dependency on the values Zl by evaluating the integral in (5).
p(ZlI8 1, m1) is given by:
p(ZlI81, m1) =
(271")
~1
1.
IK111112
exp
(--21(Zl -
m1)' Kill1 (Zl - m 1
?)
(8)
where K 1111 is the n1 x n1 covariance matrix between the points in D1 evaluated under
the covariance given by 8 1 . Completing the square in Zl in the exponent, the integral (5)
can be evaluated to give:
p
(z
188m ) 2
2,
1,
1
-
1
(271")~
_1_
1
_1_ x
IS221 t IK11111t IBlt
exp
(~ (C' B- 1C -
(9)
Z2' S2;l Z2 - m1' Kill1 m1) )
where:
B
C'
' K-1 )'S-lK' K- 1
K- 1
(K 1212
1112 22 1212 1112 + 1111
1
1
Z 2'S-lK'
22 1212 K1112 + m1 'K1111
The algorithm has been coded in MATLAB and can deal with reasonably large numbers of
points quickly. For a two dimensional vector-valued GP with n1 = 12 and n2 = 200 5 and
5This is equivalent to nl = 24 and n2 = 400 for a scalar GP.
865
Adding Constrained Discontinuities to GP Models of Wind Fields
a covariance function given by (2), computation of the log likelihood takes 4.13 seconds on
an SGI Indy R5000.
The mean value just ahead and behind the front define the mean values for the constrained
discontinuity (i.e. m1 in (9?. Conditional on the frontal parameters the wind fields either
side (Figure 3a) are assumed independent:
p(Z2a, Z2b\02, 01, Of) = p(Z2a\02, 01, m1a)p(m1a\Of) x
p(Z2b\02, 01, m1b)p(m1b\Of)
where we have performed the integration (5) to remove the dependency on Z1a and Z1b.
Thus the likelihood of the data Z2 = (Z2a, Z2b) given the model parameters O2,01 , Of
is simply the product of the likelihoods of two GPs with a constrained discontinuity which
can be computed using (9).
SOIl
_ _- . . . " " " " , , - -
""
---'
......,"--.............. ,"
-von
" , .... , ' - - - -
_II ::::
.-
,!. 100
---
'
,
.... , ,
-- - - -
, ''I. \ , - - -.... ,
- '
,,,'\--_
"" _-......... ", ,
' \ \, \, - - - ..... , , , ,
\,
' " _--....'''''' "
I
,
"DC
Front
(a)
(b)
Figure 3: (a) The division of the wind field using the generative frontal model. Z1a, Z1b
are the wind fields just ahead and behind the front, along its length, respectively. Z2a,
Z2b are the wind fields in the regions ahead of and behind the front respectively. (b) An
example from the generative frontal model: the wind field looks like a typical 'cold front'.
The model outlined above was tested on simulated data generated from the model to assess
parameter sensitivity. We generated a wind field ZO = (Z2a' Z2b) using known model
parameters (e.g. Figure 3b). We then sampled the model parameters from the posterior
distribution:
(10)
where p( ( 2), p( ( 1 ), p( Of) are prior distributions over the parameters in the GPs and front
models. This brings out one advantage of the proposed model. All the model parameters
have a physical interpretation and thus expert knowledge was used to set priors which
produce realistic wind fields. We will also use (10) to help set (hyper)priors using real data
in Zoo
MCMC using the Metropolis algorithm (Neal, 1993) is used to sample from (to) using the
NETLAB 6 library. Convergence of the Markov chain is currently assessed using visual inspection of the univariate sample paths since the generating parameters are known, although
other diagnostics could be used (Cowles and Carlin, 1996). We find that the procedure is
insensitive to the initial value of the GP parameters, but that the parameters describing the
location ofthe front (1/>" d,) need to be initialised 'close' to the correct values if the chain
is to converge on a reasonable time-scale. In the application some preliminary analysis of
the wind field would be necessary to identify possible fronts and thus set the initial parameters to 'sensible' values. We intend to fit a vector-valued GP without any discontinuities
6Available from http://www.ncrg.aston.ac . uk/netlab/index. html.
866
D. Comjord, I. T. Nabney and C. K. 1. Williams
2
3
Sample nurrber
4 '
5
2
3
4
Sample number
? In'
w 104
(b)
(a)
Figure 4: Examples from the Markov chain of the posterior distribution (10). (a) The
energy = negative log posterior probability. Note that the energy when the chain was initialised was 2789 and the first 27 values are outside the range of the y-axis. (b) The angle
of the front relative to north (?> I) '
and then measure the 'strain' or misfit of the locally predicted winds with the winds fitted
by the GP. Lines of large 'strain' will be used to initialise the front parameters.
3000
1000
500
2
3
sample number
(a)
~
~-
~~-an1.5~uw~2ww~2.~5~~3L-~3. 5
Angle of wind (radians)
(b)
Figure 5: Examples from the Markov chain of the posterior distribution (10). (a) The angle
of the wind across the front (01 ). (b) Histogram of the posterior distribution of 01 allowing
a 10000 iteration bum-in period.
Examples of samples from the Markov chain from the simulated wind field shown in Figure 3a can be seen in Figures 4 and 5. Figure 4a shows that the energy level (= negative log
posterior probability) falls very rapidly to near its minimum value from its large starting
value of 2789. In these plots the true parameters for the front were ?> I = 0.555,01 = 2.125
while the initial values were set at ?>I = 0.89,01 = 1.49. Other parameters were also incorrectly set. The Metropolis algorithm seems to be able to find the minimum and then
stays in it.
Figure 4b and 5a show the Markov chains for ?>I and 0/ ' Both converge quickly to an apparently stationary distributions, which have mean values very close to the 'true' generating
parameters. The histogram of the distribution of 01 is shown in Figure 5b.
Adding Constrained Discontinuities to GP Models of Wind Fields
867
4 DISCUSSION AND CONCLUSIONS
Simulations from our model are meteorologically plausible wind fields which contain
fronts. It is possible similar models could usefully be applied to other modelling problems where there are discontinuities with known properties. A method for the computation
of the likelihood of data given two GP models, one with non-zero mean on the boundary
and another in the domain in which the data is observed, has been given. This allows us
to perform inference on the parameters in the frontal model using a Bayesian approach of
sampling from the posterior distribution using a MCMC algorithm.
There are several weaknesses in the model specifically for fronts, which could be improved
with further work. Real atmospheric fronts are not straight, thus the model would be improved by allowing 'curved' fronts. We could represent the position of the front, oriented
along the angle defined by ?, using either another smooth GP, B-splines or possibly polynomials.
Currently the points along the line of the front are simulated at the mean observation spacing in the rest of the wind field ('" 50 km). Interesting questions remain about the (in-fill)
asymptotics (Cressie, 1993) as the distance between the points along the front tends to zero.
Empirical evidence suggests that as long as the spacing along the front is 'much less' than
the length scale of the GP along the front (which is typically'" 1000 km) then the spacing
does not significantly affect the results.
Although we currently use a Metropolis algorithm for sampling from the Markov chain,
the derivative of (9) with respect to the GP parameters 8 1 and 8 2 could be computed analytically and used in a hybrid Monte Carlo procedure (Neal, 1993).
These improvements should lead to a relatively robust procedure for putting priors over
wind fields which will be used with real data when retrieving wind vectors from scatterometer observations over the ocean.
Acknowledgements
This work was partially supported by the European Union funded NEUROSAT programme
(grant number ENV 4 CT96-0314) and also EPSRC grant GRlL03088 Combining Spatially
Distributed Predictions from Neural Networks.
References
Cornford, D. 1998. Flexible Gaussian Process Wind Field Models. Technical Report
NCRG/98/017, Neural Computing Research Group, Aston University, Aston Triangle, Birmingham, UK.
Cowles, M. K. and B. P. Carlin 1996. Markov-Chain Monte-Carlo Convergence
Diagnostics-A Comparative Review. Journal of the American Statistical Association 91, 883-904.
Cressie, N. A. C. 1993. Statistics for Spatial Data. New York: John Wiley and Sons.
Daley, R. 1991. Atmospheric Data Analysis. Cambridge: Cambridge University Press.
Handcock, M. S. and J. R. Wallis 1994. An Approach to Statistical Spatio-Temporal
Modelling of Meteorological Fields. Journal of the American Statistical Association 89, 368-378.
Neal, R. M. 1993. Probabilistic Inference Using Markov Chain Monte Carlo Methods.
Technical Report CRG-TR-93-1, Department of Computer Science, University of
Toronto. URL: http://www.cs.utoronto.ca/ ... radford.
| 1502 |@word polynomial:1 seems:1 km:3 d2:6 simulation:2 dz1:1 covariance:11 tr:1 initial:3 o2:1 atlantic:1 com:1 z2:7 yet:1 john:1 realistic:3 remove:2 plot:1 stationary:2 generative:10 plane:2 inspection:1 scotland:1 location:1 toronto:1 accessed:1 along:16 mla:1 retrieving:2 dan:1 gpl:2 introduce:1 p1:2 scatterometer:3 resolve:1 little:1 project:1 notation:1 what:1 interpreted:1 developed:1 temporal:1 usefully:1 uk:6 control:1 zl:8 grant:2 appear:1 local:1 tends:1 path:1 suggests:1 co:1 range:2 yj:1 union:1 procedure:5 cold:1 asymptotics:1 empirical:1 significantly:1 regular:1 close:2 put:1 www:3 equivalent:2 williams:4 starting:1 l:1 d1:5 fill:1 initialise:1 gps:9 cressie:3 origin:1 velocity:1 observed:2 epsrc:1 cornford:4 region:4 ensures:1 dynamic:2 division:2 triangle:1 joint:1 various:1 zo:1 describe:1 monte:4 hyper:1 outside:1 valued:5 z2i:1 plausible:1 statistic:1 gp:18 final:2 advantage:1 reconstruction:1 product:1 combining:1 rapidly:1 description:2 sgi:1 convergence:2 satellite:1 produce:3 generating:3 comparative:1 help:1 illustrate:1 ac:3 ex:1 p2:1 indy:1 predicted:1 c:1 direction:4 discontinuous:1 correct:1 explains:1 atmosphere:2 preliminary:1 crg:1 pl:1 exp:4 nurrber:1 estimation:1 birmingham:2 currently:3 bisects:1 gaussian:4 always:1 modified:4 season:1 derived:2 improvement:2 modelling:4 likelihood:9 posteriori:1 inference:3 typically:1 initially:1 interested:2 flexible:3 orientation:2 html:1 exponent:1 constrained:13 spatial:2 integration:1 field:32 having:1 sampling:3 env:1 placing:1 s22:1 look:1 report:3 spline:1 realisation:1 primarily:1 oriented:1 intended:1 n1:8 weakness:2 nl:2 diagnostics:2 behind:5 z2b:5 bum:1 chain:11 integral:2 necessary:1 fitted:1 front:54 too:1 dependency:2 varies:1 ehi:1 sensitivity:1 stay:1 probabilistic:2 informatics:1 quickly:2 von:1 ambiguity:1 possibly:1 expert:1 derivative:1 american:2 account:2 potential:1 north:3 depends:2 stream:1 performed:1 wind:51 apparently:1 ass:1 square:1 variance:1 ofthe:3 identify:1 misfit:1 bayesian:3 carlo:4 zoo:1 straight:1 definition:1 energy:3 initialised:2 associated:1 radian:1 sampled:1 knowledge:1 improved:2 evaluated:4 marginalising:1 just:6 correlation:2 horizontal:1 christopher:1 meteorological:2 brings:1 contain:2 true:2 analytically:1 spatially:1 neal:3 deal:1 conditionally:3 covering:1 hill:1 m1l:1 demonstrate:1 temperature:1 common:1 physical:2 b4:1 insensitive:1 ncrg:3 discussed:1 interpretation:1 m1:11 association:2 cambridge:2 smoothness:2 outlined:2 handcock:2 mlb:2 funded:1 stable:1 surface:6 vorticity:3 posterior:7 seen:1 minimum:2 gp2:3 converge:2 period:1 bessel:3 clockwise:1 ii:1 full:1 infer:1 smooth:2 technical:3 cross:1 long:1 coded:1 prediction:2 vision:1 histogram:2 represent:2 iteration:1 spacing:4 addressed:1 rest:1 subject:1 near:2 enough:1 affect:1 carlin:2 fit:1 url:1 york:1 matlab:1 generally:1 useful:1 locally:1 differentiability:1 http:3 group:2 putting:1 uw:1 isd:1 year:1 run:1 angle:7 reasonable:1 forrest:1 z2a:5 netlab:2 completing:1 nabney:4 correspondence:1 strength:1 ahead:5 constraint:2 x2:1 speed:3 simulate:4 relatively:1 department:1 combination:1 across:4 smaller:1 meteorologically:1 remain:1 son:1 metropolis:3 taken:1 computationally:1 describing:3 adopted:1 available:1 observe:1 away:2 ocean:2 include:1 calculating:1 intend:1 question:1 mnd:1 usual:1 distance:1 simulated:8 sensible:1 whom:1 assuming:1 length:4 index:1 difficult:1 ql:1 negative:2 zt:1 perform:1 allowing:2 ord:1 observation:4 vertical:1 markov:9 curved:1 incorrectly:1 situation:1 variability:2 strain:2 dc:1 ww:1 community:1 atmospheric:3 required:1 flowchart:1 discontinuity:19 geostatistics:1 able:2 suggested:1 including:1 suitable:1 demanding:1 hybrid:1 representing:1 aston:6 library:1 temporally:1 lk:2 axis:1 text:1 prior:10 review:1 l2:1 acknowledgement:1 determining:1 relative:1 interesting:1 consistent:1 principle:1 uncorrelated:1 soil:1 supported:1 side:8 understand:1 fall:1 edinburgh:2 distributed:1 boundary:2 curve:3 evaluating:1 made:1 lz:1 programme:1 ml:1 assumed:3 spatio:1 an1:1 reasonably:1 robust:1 ca:1 investigated:1 complex:1 european:1 domain:3 bounding:1 s2:2 n2:9 raid:1 wiley:1 position:4 daley:3 xl:1 ib:1 ian:1 utoronto:1 evidence:1 derives:1 dl:2 adding:4 conditioned:2 lt:1 simply:1 univariate:1 visual:1 partially:1 scalar:1 radford:1 conditional:1 marked:1 change:2 typical:2 determined:1 corrected:1 specifically:1 wallis:2 htm1:1 assessed:1 frontal:10 evaluate:1 mcmc:4 tested:1 cowles:2 |
551 | 1,503 | Learning from Dyadic Data
Thomas Hofmann?, Jan Puzicha+, Michael I. Jordan?
? Center for Biological and Computational Learning, M .I.T
Cambridge , MA , {hofmann , jordan}@ai.mit.edu
+ Institut fi.ir Informatik III , Universitat Bonn, Germany, jan@cs.uni-bonn.de
Abstract
Dyadzc data refers to a domain with two finite sets of objects in
which observations are made for dyads , i.e., pairs with one element
from either set. This type of data arises naturally in many application ranging from computational linguistics and information
retrieval to preference analysis and computer vision . In this paper,
we present a systematic, domain-independent framework of learning from dyadic data by statistical mixture models. Our approach
covers different models with fiat and hierarchical latent class structures. We propose an annealed version of the standard EM algorithm for model fitting which is empirically evaluated on a variety
of data sets from different domains.
1
Introduction
Over the past decade learning from data has become a highly active field of research distributed over many disciplines like pattern recognition, neural computation , statistics, machine learning, and data mining. Most domain-independent
learning architectures as well as the underlying th eories of learning have been focusing on a feature-based data representation by vectors in an Euclidean space. For
this restricted case substantial progress has been achieved. However, a variety of
important problems does not fit into this setting and far less advances have been
made for data types based on different representations.
In this paper, we will present a general framework for unsupervised learning from
dyadic data . The notion dyadic refers to a domain with two (abstract) sets of objects, ;r = {Xl , ... , XN} and Y = {YI, ... , YM} in which observations S are made for
dyads (Xi, Yk). In the simplest case - on which we focus - an elementary observation
consists just of (Xi, Yk) itself, i.e., a co-occurrence of Xi and Yk, while other cases
may also provide a scalar value Wik (strength of preference or association). Some exemplary application areas are: (i) Computational linguistics with the corpus-based
statistical analysis of word co-occurrences with applications in language modeling ,
word clustering, word sense disambiguation , and thesaurus construction. (ii) Textbased znJormatzon retrieval, where ,:{, may correspond to a document collection , Y
467
Learningfrom Dyadic Data
to keywords , and (Xi, Yk) would represent the occurrence of a term Yk in a document
Xi. (iii) Modeling of preference and consumption behavior by identifying X with individuals and Y with obj ects or stimuli as in collaborative jilterzng. (iv) Computer
VIS tOn , in particular in the context of image segmentation, where X corresponds to
imagE' loc ations , y to discretized or categorical feature values , and a dyad (Xi , Yk)
represents a feature Yk observed at a particular location Xi.
2
Mixture Models for Dyadic Data
Across different domains there are at least two tasks which playa fundamental role
in unsupervised learning from dyadic data: (i) probabilistic modeling, i.e., learning
a joint or conditional probability model over X xY , and (ii) structure discovery, e.g. ,
identifying clusters and data hierarchies. The key problem in probabilistic modeling
is the data sparseness: How can probabilities for rarely observed or even unobserved
co-occurrences be reliably estimated? As an answer we propose a model-based approach and formulate latent class or mixture models . The latter have the further
advantage to offer a unifying method for probabilistic modeling and structure discovery. There are at least three (four, if both variants in (ii) are counted) different
ways of defining latent class models:
The most direct way is to introduce an (unobserved) mapping c : X X Y --+
{Cl , . . . , CK} that partitions X x Y into K classes. This type of model is
called aspect-based and the pre-image c- l (cO') is referred to as an aspect.
n. Alternatively, a class can be defined as a subset of one of the spaces X (or Y
by symmetry, yielding a different model) , i.e., C : X --+ {Cl, . .. , CK} which
induces a unique partitioning on X x Y by C(Xi , yk) == C(Xi) . This model is
referr ed to as on e-szded clustering and c-l(c a ) ~ X is called a cluster.
Ill. If latent classes are defined for both sets, c : X --+ {ci , .. . , cK} and C :
Y --+ {cI , . .. , cD, respectively, this induces a mapping C which is a K . L partitioning of X x y. This model is called two-sided clustering.
I.
2.1
Aspect Model for Dyadic Data
In order to sp ecify an aspect model we make the assumption that all co-occurrences
in the sample set S are i.i .d. and that Xi and Yk are conditionally independent given
the class. With parameters P(x i lca ), P(Yklca) for the class-conditional distributions
and prior probabilities P( cO' ) the complete data probability can be written as
P(S , c) =
IT [P(Cik)P(Xilcik)P(Yklcik)t (x"Yk)
(1)
,
i,k
where n(xi, Yk) are the empirical counts for dyads in Sand Cik == C(Xi, Yk) . By
summing over the latent variables C the usual mixture formulation is obtained
P(S)
= IT P(Xi, Ykt(X"Yk),
i,k
where P(Xi , Yk)
=L
P(ca)P(xilca)P(Yk Ic a )
.
(2)
a
Following the standard Expectation Maximi zation approach for maximum likelihood
t's timation [DE'mpster et al .. 1977], the E-step equations for the class posterior probabilities arE' given byl
(3)
1 In the case of multiple observations of dyads it has been assumed that each observation
may have a different latent class. If only one latent class variable is introduced for each
dyad, slightly different equations are obtained.
T. Hofmann, J puzicha and M. 1. Jordan
468
??
~
..... .
. . . . _0 . . . . . . . .
............... --
..... __ .........
Ill , U.UU4
114, U.UU:l
! P (Ca)
: maximal
!P(XiICcx)
: maximal
!p(YklcCX>
two 0.18
seven 0.10
tbree 0.10
four 0.06
five 0.06
years 0.11
bousand 0.1
buodred 0.1
days 0.07
cubits 0.05
bave
batb
bad
bast
0.38
0.22
0.11
0.09
be 0.02
up 0.40
dowoO.17
fortb 0.15
out 0.09
ioO.Ol
I"', U.U~9
sbalt 0.18
bast 0.08
wilt 0.08
art 0.07
if 0.05
tbe 0.95
bis 0.006
my 0.005
our 0.003
tby 0.003
tbou 0.85
DOt 0.01
also 0.004
ndeed 0.00
aooiot 0.003
lord 0.09
bildreo 0.0
SOD 0.02
land 0.02
o Ie 0.02
<.>
<:>
<,>
<;>
<?>
0.52
0.16
0.14
0.07
0.04
aDd 0.33
for 0.08
but 0.07
ben 0.0
so 0.02
ee O.
me 0.03
him 0.03
it 0.02
you 0.02
<?>
<,>
<.>
<:>
<.>
0.27
0.23
0.12
0.06
0.04
Figure 1: Some aspects of the Bible (bigrams) .
It is straightforward to derive the M-step re-estimation formulae
P(c a) ex
L n(xi' Yk)P{Cik = Ca }, P(xilca) ex L n(xi, Yk)P{Cik = Ca },
i,k
(4)
k
and an analogous equation for P(Yk Ic a ). By re-parameterization the aspect model
can also be characterized by a cross-entropy criterion. Moreover, formal equivalence to the aggregate Markov model, independently proposed for language modeling in [Saul, Pereira, 1997], has been established (cf. [Hofmann, Puzicha, 1998] for
details).
2.2
One-Sided Clustering Model
The complete data model proposed for the one-sided clustering model is
P(S, c) = P( c)P(SIc) =
(If
P( c(x;)) )
(IT
[P( x;)P(Y' Ic( X;))]n(x",,)) ,
(5)
where we have made the assumption that observations (Xi, Yk) for a particular Xi
are conditionally independent given c( xd . This effectively defines the mixture
P(S) =
IT P(S;) ,
P(S;) =
L P(c a ) IT [P(XdP(Yklea)r(X"Yk)
a
,
(6)
k
where Si are all observations involving Xi. Notice that co-occurrences in Si are not
independent (as they are in the aspect model) , but get coupled by the (shared)
latent variable C(Xi). As before, it is straightforward to derive an EM algorithm
with update equations
P{ c( Xi) =Ca } ex P( Ca )
IT P(Yk Icat(x. ,Yk), P(Yk lea) ex L n(Xi, Yk )P{ c( Xi) =ca }
(7)
k
and P(c a ) ex Li P{C(Xi ) = cal, P(Xi) ex Lj n(xi,Yj)? The one-sided clustering
mod el is similar to the distributional clustering model [Pereira et al. , 1993], however, there are two important differences: (i) the number of likelihood contributions
in (7) scales with the number of observations - a fact which follows from Bayes' rule
- and (ii) mixing proportions are rpissing in the original distributional clustering
model. The one-sided clustering model corresponds to an unsupervised version of
the naive Bayes' classifier, if we interpret Y as a feature space for objects Xi EX .
There are also ways to weaken the conditional independence assumption, e.g., by
utilizing a mixture of tree dependency models [Meila, Jordan, 1998] .
2.3
Two-Sided Clustering Model
The latent variable structure of the two-sided clustering model significantly reduces
the degrees of freedom in the specification of the class conditional distribution. We
Learning from Dyadic Data
469
Figure 2: Exemplary segmentation results on Aerial by one-sided clustering.
propose the following complete data model
P(S, c) =
II P(C(Xi))P(C(Yk)) [P(xi)P(Yk)1Tc(xi),c(YIc)f(x"yIc)
(8)
i,k
where 1Tc:r: cll are cluster association parameters. In this model the latent variables
in the X and Y space are coupled by the 1T-parameters. Therefore, there exists
no simple mixture model representation for P(S). Skipping some of the technical
details (cf. [Hofmann, Puzicha, 1998]) we obtain P(Xi) ex Lk n(xi,Yk), P(Yk) ex
Li n(xi' Yk) and the M-step equations
0"
"Y
L i k n(xi, Yk)P{C(Xi) = c~ /\ C(Yk) = c~}
1Tc~.c~ = [Li P{C(Xi) = ~;} Lk n(xi, Yk)] [Lk P{C(Yk) = cn Li n(xi, Yk)]
=
=
=
(9)
=
as well as P(c~)
L i P{C(Xi) c~} and P(c~) Lk P{C(Xk) cn . To preserve
tractability for the remaining problem of computing the posterior probabilities in
the E-step , we apply a factorial approximation (mean field approximation), i.e.,
P{C(Xi ) = c~ /\ C(Yk) =
~ P{C(Xi) = c~}P{C(Yk) =
This results in the
following coupled approximation equations for the marginal posterior probabilities
cO
P{ c(x;)
cn.
= c~} ex P(c~) exp [~n(x;, y,) ~ PI cry,) = c'(} log "'~"~1
(10)
and a similar equation for P {C(Yk) = c~}. The resulting approximate EM algorithm
performs updates according to the sequence (CX- post., 1T, cLpost., 1T). Intuitively
the (probabilistic) clustering in one set is optimized in alternation for a given clustering in the other space and vice versa. The two-sided clustering model can also
be shown to maximize a mutual information criterion [Hofmann, Puzicha, 1998] .
Discussion: Aspects and Clusters
To better understand the differences of the presented models it is elucidating to
systematically compare the conditional probabilities P( CO' Ixd and P( CO' IYk):
2.4
Aspect
Model
P(colxd
P(CoIYk )
P{x.ico' W{co' 2
P(x,)
P~lf.k
Ic", W( c'" 2
P(Y k)
Two-sided
Clustering
One-sided
One-sided
X Clustering
Y Clustering
P{xdcO' W{ CO' 2
P{C(Xi)
P{C(Yk) = cO'}
P{C(Yk) = c~}
P{c(xd = cO'}
P(lf.kl cO' )P(cO' 2
P(Yk)
P(x.)
= c~}
As can be seen from the above table, probabilities P(CoIXi) and P(CaIYk) correspond
to posterior probabilities of latent variables if clusters are defined in the X-and
Y-space, respectively. Otherwise, they are computed from model parameters. This
is a crucial difference as, for example, the posterior probabilities are approaching
T. Hofmann, J. Puzicha and M. I. Jordan
470
. . ...-
'''' P' '' ''' '~
I
... 1?
' ~.
.~ ,
f ..... I~
. --" ....
. . .r... ...........,.....
..
< ?? I~
h .. .. ' ? ???
"
'dol
_.ie
_ ??? 1
Il10' ' ''
t.
.,...
c r .....
~:.:'
~
I??
I''''''.'
d?,.
."....
h'"
.. , ..
t._~..
..10
..'
.,.....
? ... ~ . " ... ,
~.'''< .I ??
~:.:: ~."..
~.::;,:'::.I ~:::: ..
IT
. h .~.
.....
~....
.. . .,_.
....
I"".
... n,
??10' '''
f.:~I:.:,.,
,.1.?
~::.
~i...
"
? ..... _ .
i,.
..
.....,.
?u.
_" ....
..? .. .i'.......
f . . ..
U.tl~
.~ ~"
cl~
1&
. _11
"'7
' '''.,.
I ~ ..
'''. " ...... , .... ~ .i ul
::.7.:,.,
to.,
" ' ..tol.
Io'.,f.. ,
? ?? 1-.
::':~.,
t~?.?t
~;:.;::.'
. 1. . ....
,' " .... ,.
oa." ....,
~~?:;?i:?
~,
. .c; ?? "" ...!:~:I::'
,,,II
" . .. ..
,...
, .. h lt.
..."... , ...it,
::::,:,:;"??1::;::;.
.Iln., ...
... w
, . ..
.. ?? ? ? ? L
::,':~~;~~
~::~'
1. '10 ?
,n.,
??
".d .
D~~~ ""
I~
.1
I?? ,.
U
..,.<.
Ho,
..
..-~
"?:~::,~~:.
..,fa_.
,.1....
_ ;' ,117
I
d .....
u u ll ? ??
,,,??.U,
........
..........
...
,,<., .,,"??ua,
"
?
,
" ,
, h .... , ., " ??? , .. .. .
,~,-."
u d d . ..
y
.. .. .... hi
II
~,,"
.< ~
,."... .
. ~.c,
..
.h.
.....
?.. Ii.,
"
,e. .....
?? , ,
..?...
1111 ,
????
.~I.
1<1
~
:,~:::.
'It
b ..... ,
u ."'~'
)O ,.i.
~ ._ ..
u
:~::.
"'~h"
........... 'u..
..
.,
<? ?? .,.
<. . . .11 . . .. i . ..
"
.flu
. . , .. , . . .
..
n, ..."
,........
_.. ',.<.
:~::~
1, ? ? -
.. ... ,
? . . ,, '"
fl ????
. .... ]0
:I~;:;?.~'
",oi,
~
~~.::
",
1"
....
....'.,"10 ...
~.,
,... ,, 1. . .
.0
31
1.<0'
~
I."""'''' .. ~Ii.,<oJ
~:~::~',
..
;:~:,.
~~:::~.".~'. :::~'.~~~~'
.??;::., :::;!..'.,
Figure 3: Two-sided clustering of LOB:
7r
matrix and most probable words.
Boolean values in the infinite data limit and P(Yklxd = Lo P{C(Xi)=Co}P(Yk!co)
are converging to one of the class-conditional distributions. Yet, in the aspect model
P(Yklxd = Lo P(CoIXi)P(Yk!co) and P(CoIXi) ex: P(Co)P(Xi!co) are typically not
peaking more sharply with an increasing number of observations. In the aspect
model, conditionals P(Yk IXi) are inherently a weighted sum of the 'prototypical'
distributions P(Yk Ic o ). Cluster models in turn ultimately look for the 'best' classconditional and weights are only indirectly induced by the posterior uncertainty.
3
The Cluster-Abstraction Model
The models discussed in Section 2 all define a non-hierarchical, 'flat' latent class
structure. However, for structure discovery it is important to find hierarchical data
organizations. There are well-known architectures like the Hierarchical Mixtures
of Experts [Jordan, Jacobs , 1994] which fit hierarchical models. Yet, in the case
of dyadic data there is an alternative possibility to define a hierarchical model.
The Cluster-Abstraction Model (CAM) is a clustering model (e.g., in X) where
the conditionals P(Yk Ico) are itself xi-specific aspect mixtures, P(Yk leo, Xi) =
LII P(Yk la ll )P( alllc o , Xi) with a latent aspect mapping a. To obtain a hierarchical organization, clusters Co are identified with the terminal nodes of a hierarchy
(e.g., a complete binary tree) and aspects all with inner and terminal nodes. As
a compatibility constraint it is imposed that P( all/co, xd = 0 whenever the node
corresponding to all is not on the path to the terminal node co. Intuitively, conditioned on a 'horizontal' clustering c all observations (Xi, Yk) E Si for a particular
Xi have to be generated from one of the 'vertical' abstraction levels on the path to
c( Xi)' Since different clusters share aspects according to their topological relation ,
this favors a meaningful hierarchical organization of clusters. Moreover, aspects at
inner nodes do not simply represent averages over clusters in their subtree as they
are forced to explicitly represent what is common to all subsequent clusters.
Skipping the technical details, the E-step is given by
P{a(xi,Yk)
= all/c(xi) = co} ex: P(alllco,xi)P(Yk/all)
P{ C(Xi) = co} ex: P( co)
II L [P( alllc o , Xi)P(Yk /a ll
(11)
)r(X"Yk)
(12)
II
and the M-step formulae are P(Yk/all) ex: LiP{a(xi,Yk) = all}n(xi,Yk), P(c o ) ex:
Li P{C(Xi) = co}, and P(alllco , Xi) ex: Lk P{a(xi ' Yk) = all/c(xi) = co}n(xi, Yk)'
k
471
Learning from Dyadic Data
network
neural
function
....... :f;:::::m??????
function
: weiCht '.
.: learn \
.' error '.
.:' example \
nd
,talc
neuron
~onlinu
<lrbitr~ri
rundl.l.ln :
~idd~:r i
learn
train
problem
d .. s~ir
'r:lin
taleet
frtttur
method
~rror
cener
learn
perform
im.{
j
ctntr
ne~.~ork
model
~~rs~
,
!'in
;~=fit ~
:.:~
update-:
'_tur
.pUm
bound;
ute
cCHWtr-.in,
lin?...
It.n:d
.....-.
1
~c~.
i\, '1M1"I1
1
/~
i """ ~:!:' ~.r ./ \ ... :-e:rsion demnln
i
~;IKIU' ! . . . : :. . . .: .-: . ~;~-.//
\="
appt"e("u.
m ....
th~d
1MOt00itnn
'lUlu
;;yne~nu::=t
~:::n
~i;~
....
UrUyers
rglat
deted
,.b.1'
theta
tnnde!'
-II
~t~n ?-? ?? ?? ?-?
process
author
Infonn
dy~.~m,ic
control ;
:
/ ;?k. i ;~~:...;
~~~!~-tc
machin :
:
...... ...... -, dynamjc
:' Umt '"
.: modd \
synaptic..
f nruron ",
neuron nrtwork
actiy
nrural
spike
learn
r~pons aicorithm
i
control
neural
desi",
infonn
C);,I;\sin
1I11~'rilh~
"",wI
ptrfonn
: ntlwork:
i process ~
: data
:
j modtl :
.
,.,"'/ /
raw
.plim
m.".... ?
-;;:..
..
7?~d~_" " ~""_-::;'bll
=
t
i
~
nrtwork \
; map
\
visual memori
imae
associ
motion (papc
ctll
pattern
p~~tSS lra~
~pt
INtt:w.
j
i
~~:nl
:::;eri
map
paramll :
naI~
certic
object
'
=!:. ..... ~
,,.,....w (:...., ',....
~ni;~
:::
::~d'.;:;" ';~'';ari
thre:m.td
mKhin
Gscil
1.caI
,.,-...et
:;::~~ :~it.ri
steM,....t
?
learn.
:' a1corithm
i rult
,..oIen:
r::kl' ::!::-'
pedsJ"a.pt nop)
".ndud.
~tin
~:;::
repoa
m:
krn (...J
"tat
t,..nsfonll
Figure 4: Parts of the top levels of a hierarchical clustering solution for the Neural
document collection, aspects are represented by their 5 most probable word stems.
4
Annealed Expectation Maximization
Annealed EM is a generalization of EM based on the idea of deterministic annealing [Rose et al., 1990] that has been successfully applied as a heuristic optimization
technique to many clustering and mixture problems. Annealing reduces the sensitivity to local maxima, but, even more importantly in this context , it may also improve
the generalization performance compared to maximum likelihood estimation. 2 The
key idea in annealed EM is to introduce an (inverse temperature) parameter (3 ,
and to replace the negative (averaged) complete data log-likelihood by a substitute
known as the fre e energy (both are in fact equivalent at f3 = 1) . This effectively
results in a simple modifi cation of the E-step by taking the likelihood contribution
in Bayes ' rul e to the power of ;3. In order to determine the optimal value for f3 we
used an additional validation set in a cross validation procedure.
5
Results and Conclusions
In our experiments we have utilized the following real-world data sets: (i) Cranfield:
a standard test collection from information retrieval (N = 1400, M = 4898) , (ii)
Penn : adjective-noun co-occurrences from the Penn Treebank corpus (N = 6931 ,
M = 4995) and the LOB corpus (N = 5448, M = 6052) , (iii) Neural: a document
collection with abstracts of journal papers on neural networks (N = 1278 , M = 6065) ,
(iv) Bzble: word bigrams from the bible edition of the Gutenberg project (N = M =
12858) , (v) Aerial: Textured aerial images for segmentation (N = 128x128, M = 192).
In Fig. 1 we have visualized an aspect model fitted to the Bible bigram data. Notice
that although X = Y the role of the preceding and the subsequent words in bigrams
is quite different . Segmentation results obtained on Aerial applying the one-sided
clustering model are depicted in Fig. 2. A multi-scale Gabor filter bank (3 octaves,
4 orientations) was utilized as an image representation (cf. [Hofmann et al. , 1998]) .
In Fig. 3 a two- sided clustering solution of LOB is shown. Fig. 4 shows the top
levels of the hierarchy found by the Cluster-Abstraction Model in Neural. The
inner node distributions provide resolution-specific descriptors for the documents
in the conesponding subtree which can be utilized , e.g., in interactive browsing
for information retrieval, Fig. 5 shows typical test set perplexity curves of the
2 Moreover,
the tree topology for the CAM is heuristically grown via phase transitions.
T. Hofmann, 1. Puzicha and M 1. Jordan
472
(b)
(a)
(c)
".;----=-~---;:--~-,:;---:~
.~EJoII,"""',
....
Figure 5: Perplexity curves for annealed EM (aspect (a), (b) and one-sided clustering model (c)) on the Bible and Gran data.
Aspect
K
1
8
16
32
64
128
f3
'P
Cran
- 685
0.88 482
0.72 255
0.83 386
0.79 360
0.78 353
X-duster
f3
'P
CAM
f3
'P
X /Y-c1uster
'P
f3
Aspect
f3
X-cluster
CAM
'P
f3
'P
f3
'P
639
312
255
205
182
166
0.08
0.07
0.07
0.07
0.06
352
302
254
223
231
0.13
0.10
0.08
0.07
0.06
322
268
226
204
179
X /Y-cluster
'P
f3
Penn
0.09
0.07
0.07
0.06
0.04
527
302
452
527
663
0.18
0.10
0.12
0.11
0.10
511
268
438
422
410
0.67
0.51
0.53
OA8
OA5
0.73
0.72
0.71
0.69
0.68
615
335
506
477
462
0.55
0.51
0.46
0.44
DAD
394
335
286
272
241
Table 1: Perplexity results for different models on the Gran (predicting words conditioned on documents) and Penn data (predicting nouns conditioned on adjectives).
annealed EM algorithm for the aspect and clustering model (P = e- 1 where I is
the per-observation log-likelihood). At {J
1 (standard EM) overfitting is clearly
visible, an effect that vanishes with decreasing (J. Annealed learning performs also
better than standard EM with early stopping. Tab. 1 systematically summarizes
perplexity results for different models and data sets.
=
In conclusion mixture models for dyadic data have shown a broad application potential. Annealing yields a substantial improvement in generalization performance
compared to standard EM, in particular for clustering models, and also outperforms a complexity control via J{. In terms of perplexity, the aspect model has
the best performance. Detailed performance studies and comparisons with other
state-of-the-art techniques will appear in forthcoming papers.
References
[Dempster et al., 1977] Dempster, A.P., Laird, N.M., Rubin, D.B. (1977). Maximum likelihood from incomplete data via the EM algorithm . J. Royal Statist. Soc. B , 39 , 1-38.
[Hofmann , Puzicha, 1998] Hofmann, T., Puzicha, J. 1998. Statistical models for cooccurrence data. Tech. rept . Artifical Intelligence Laboratory Memo 1625, M.LT.
[Hofmann et al., 1998] Hofmann, T., Puzicha, J ., Buhmann, J.M. (1998). Unsupervised
texture segmentation in a deterministic annealing framework. IEEE Transactions on
Pattern Analysis and Machine Intelligence , 20(8) , 803-818.
[Jordan, Jacobs, 1994] Jordan, M.L, Jacobs, R.A. (1994). Hierarchical mixtures of experts
and the EM algorithm. Neural Computation, 6(2), 181-214.
[Meila, Jordan , 1998] Meila, M., Jordan, M. L 1998. Estimating Dependency Structure
as a Hidden Variable. In: Advances in Neural Information Processing Systems 10.
[Pereira et al., 1993] Pereira, F.e.N., Tishby, N.Z., Lee, L. 1993. Distributional clustering
of English words. Pages 189-190 of: Proceedings of the A CL.
[Rose et al., 1990] Rose, K., Gurewitz , E., Fox, G. (1990). Statistical mechanics and phase
transitions in clustering. Physical Review Letters, 65(8), 945-948.
[Saul, Pereira, 1997] Saul, 1., Pereira, F. 1997. Aggregate and mixed-order Markov models for statistical language processing. In: Proceedings of the 2nd International Conference on Empirical Methods in Natural Language Processing.
| 1503 |@word version:2 bigram:4 proportion:1 nd:2 heuristically:1 r:1 tat:1 jacob:3 loc:1 document:6 past:1 outperforms:1 thre:1 skipping:2 si:3 yet:2 written:1 visible:1 subsequent:2 partition:1 hofmann:13 update:3 intelligence:2 parameterization:1 xk:1 node:6 location:1 preference:3 x128:1 five:1 direct:1 become:1 ect:1 consists:1 fitting:1 introduce:2 behavior:1 pum:1 mechanic:1 multi:1 ol:1 discretized:1 terminal:3 decreasing:1 td:1 ua:1 increasing:1 project:1 estimating:1 underlying:1 moreover:3 what:1 unobserved:2 xd:3 interactive:1 classifier:1 partitioning:2 control:3 penn:4 appear:1 before:1 local:1 rept:1 limit:1 io:1 flu:1 path:2 equivalence:1 co:30 bi:1 averaged:1 unique:1 yj:1 lf:2 procedure:1 jan:2 area:1 empirical:2 significantly:1 gabor:1 word:9 pre:1 refers:2 get:1 cal:1 context:2 applying:1 equivalent:1 imposed:1 map:2 center:1 deterministic:2 annealed:7 straightforward:2 independently:1 formulate:1 resolution:1 identifying:2 ixd:1 rule:1 utilizing:1 importantly:1 notion:1 analogous:1 construction:1 hierarchy:3 machin:1 pt:2 ixi:1 element:1 recognition:1 utilized:3 distributional:3 observed:2 role:2 tur:1 yk:57 substantial:2 rose:3 vanishes:1 dempster:2 complexity:1 cooccurrence:1 cam:4 ultimately:1 lord:1 textured:1 joint:1 represented:1 grown:1 leo:1 train:1 infonn:2 forced:1 aggregate:2 quite:1 heuristic:1 otherwise:1 favor:1 statistic:1 itself:2 laird:1 advantage:1 sequence:1 exemplary:2 cai:1 propose:3 maximal:2 mixing:1 cluster:16 ben:1 object:4 derive:2 keywords:1 progress:1 soc:1 c:1 uu:1 filter:1 sand:1 generalization:3 biological:1 elementary:1 probable:2 im:1 ic:6 exp:1 mapping:3 early:1 estimation:2 him:1 vice:1 successfully:1 weighted:1 mit:1 clearly:1 ck:3 focus:1 improvement:1 likelihood:7 tech:1 sense:1 textbased:1 el:1 abstraction:4 stopping:1 lj:1 typically:1 hidden:1 relation:1 i1:1 germany:1 compatibility:1 ill:2 orientation:1 art:2 noun:2 mutual:1 marginal:1 field:2 f3:10 represents:1 broad:1 look:1 unsupervised:4 stimulus:1 preserve:1 individual:1 phase:2 freedom:1 organization:3 tss:1 plim:1 highly:1 mining:1 possibility:1 elucidating:1 mixture:12 nl:1 yielding:1 bible:4 xy:1 institut:1 fox:1 tree:3 iv:2 euclidean:1 incomplete:1 re:2 zation:1 weaken:1 fitted:1 modeling:6 boolean:1 cover:1 ations:1 pons:1 maximization:1 tractability:1 subset:1 sod:1 gutenberg:1 tishby:1 universitat:1 dependency:2 answer:1 my:1 fundamental:1 cll:1 sensitivity:1 ie:2 international:1 systematic:1 probabilistic:4 lee:1 discipline:1 michael:1 ym:1 lii:1 expert:2 li:5 potential:1 de:2 wilt:1 explicitly:1 vi:1 dad:1 tab:1 bayes:3 collaborative:1 contribution:2 oi:1 ir:2 ni:1 descriptor:1 bave:1 correspond:2 yield:1 raw:1 informatik:1 cation:1 umt:1 whenever:1 ed:1 synaptic:1 energy:1 naturally:1 fre:1 fiat:1 segmentation:5 cik:4 focusing:1 day:1 formulation:1 evaluated:1 just:1 cran:1 horizontal:1 defines:1 sic:1 effect:1 laboratory:1 conditionally:2 ll:3 sin:1 lob:3 iln:1 criterion:2 octave:1 complete:5 performs:2 motion:1 temperature:1 ranging:1 image:5 fi:1 ari:1 common:1 empirically:1 physical:1 association:2 discussed:1 m1:1 interpret:1 cambridge:1 versa:1 ai:1 meila:3 associ:1 language:4 dot:1 ute:1 specification:1 add:1 playa:1 nop:1 posterior:6 perplexity:5 lra:1 binary:1 alternation:1 yi:1 seen:1 additional:1 preceding:1 determine:1 maximize:1 ii:13 bast:2 multiple:1 ico:2 reduces:2 stem:2 intt:1 technical:2 characterized:1 offer:1 cross:2 retrieval:4 lin:2 post:1 converging:1 variant:1 involving:1 vision:1 expectation:2 represent:3 lulu:1 achieved:1 lea:1 conditionals:2 annealing:4 crucial:1 induced:1 mod:1 idd:1 jordan:11 obj:1 ee:1 iii:3 variety:2 independence:1 fit:3 forthcoming:1 architecture:2 approaching:1 identified:1 topology:1 inner:3 idea:2 cn:3 dyad:6 ul:1 modd:1 tol:1 byl:1 detailed:1 factorial:1 statist:1 induces:2 visualized:1 simplest:1 notice:2 estimated:1 per:1 key:2 four:2 year:1 tbe:1 sum:1 ork:1 inverse:1 letter:1 you:1 uncertainty:1 disambiguation:1 thesaurus:1 dy:1 summarizes:1 fl:1 hi:1 bound:1 topological:1 strength:1 constraint:1 sharply:1 ri:2 flat:1 bonn:2 aspect:23 lca:1 according:2 gran:2 aerial:4 across:1 slightly:1 em:13 wi:1 intuitively:2 restricted:1 peaking:1 classconditional:1 sided:16 ln:1 equation:7 turn:1 count:1 apply:1 hierarchical:10 indirectly:1 occurrence:7 alternative:1 ho:1 thomas:1 original:1 top:2 clustering:30 linguistics:2 cf:3 remaining:1 eri:1 substitute:1 unifying:1 learningfrom:1 nai:1 spike:1 usual:1 oa:1 consumption:1 me:1 seven:1 negative:1 memo:1 reliably:1 perform:1 vertical:1 observation:11 neuron:2 markov:2 finite:1 defining:1 bll:1 introduced:1 pair:1 kl:2 optimized:1 established:1 nu:1 pattern:3 adjective:2 oj:1 royal:1 power:1 natural:1 predicting:2 buhmann:1 wik:1 improve:1 iyk:1 theta:1 ne:1 lk:5 categorical:1 coupled:3 naive:1 gurewitz:1 prior:1 review:1 discovery:3 mixed:1 prototypical:1 validation:2 degree:1 rubin:1 treebank:1 bank:1 systematically:2 pi:1 cd:1 land:1 cry:1 lo:2 share:1 cranfield:1 english:1 formal:1 understand:1 saul:3 taking:1 distributed:1 curve:2 xn:1 world:1 transition:2 author:1 made:4 collection:4 counted:1 far:1 transaction:1 approximate:1 uni:1 active:1 rul:1 overfitting:1 corpus:3 summing:1 assumed:1 xi:62 alternatively:1 latent:13 decade:1 timation:1 yic:2 table:2 lip:1 learn:5 ca:7 inherently:1 symmetry:1 ioo:1 cl:4 domain:6 sp:1 edition:1 dyadic:12 fig:5 referred:1 tl:1 pereira:6 xl:1 tin:1 krn:1 formula:2 bad:1 specific:2 dol:1 ton:1 exists:1 i11:1 effectively:2 ci:2 texture:1 subtree:2 conditioned:3 sparseness:1 browsing:1 entropy:1 tc:4 cx:1 lt:2 simply:1 depicted:1 visual:1 ykt:1 scalar:1 rror:1 corresponds:2 ma:1 conditional:6 shared:1 replace:1 infinite:1 typical:1 called:3 la:1 meaningful:1 rarely:1 puzicha:10 tby:1 latter:1 arises:1 artifical:1 ex:16 |
552 | 1,504 | Example Based Image Synthesis of Articulated
Figures
Trevor Darrell
Interval Research. 1801C Page Mill Road. Palo Alto CA 94304
trevor@interval.com, http://www.interval.com/-trevor/
Abstract
We present a method for learning complex appearance mappings. such
as occur with images of articulated objects. Traditional interpolation
networks fail on this case since appearance is not necessarily a smooth
function nor a linear manifold for articulated objects. We define an appearance mapping from examples by constructing a set of independently
smooth interpolation networks; these networks can cover overlapping regions of parameter space. A set growing procedure is used to find example clusters which are well-approximated within their convex hull;
interpolation then proceeds only within these sets of examples. With this
method physically valid images are produced even in regions of parameter space where nearby examples have different appearances. We show
results generating both simulated and real arm images.
1 Introduction
Image-based view synthesis is.an important application of learning networks. offering the
ability to render realistic images without requiring detailed models of object shape and
illumination effects. To date. much attention has been given to the problem of view synthesis under varying camera pose or rigid object transformation. Several successful solutions
have been proposed in the computer graphics and vision literature. including view morphing [12], plenoptic modeling/depth recovery [8], "lightfields" [7], and recent approaches
using the trifocal tensor for view extrapolation [13].
For non-rigid view synthesis. networks for model-based interpolation and manifold learning have been used successfully in some cases [14. 2. 4. 11]. Techniques based on Radial
Basis Function (RBF) interpolation or on Principle Components Analysis (peA), have been
able to interpolate face images under varying pose. expression and identity [1.5, 6]. How-
Example-Based Image Synthesis ofArticulated Figures
769
extends the notion of example clustering to the case of coupled shape and texture appearance models.
Our basic method is to find sets of examples which can be well-approximated from their
convex hull in parameter space. We define a set growing criterion which enforces compactness and the good-interpolation property. To add a new point to an example set, we
require both that the new point must be well approximated by the previous set alone and
that all interior points in the resulting set be well interpolated from the exterior examples.
We define exterior examples to be those on the convex hull of the set in parameter space.
Given a training subset s C 0 and new point p E 0,
E(s,p)
= max(E/(s U {p}),EE(S,p))
,
with the interior and extrapolation error defined as
1ix (s) is the subset of s whose x vectors lie on the convex hull of all such vectors in s. To
add a new point, we require E < E, where E is a free parameter of the clustering method.
Given a seed example set, we look to nearest neighbors in appearance space to find the next
candidate to add. Unless we are willing to test the extrapolation error of the current model
to all points, we have to rely on precomputed non-vectorized appearance distance (e.g.,
MSE between example images). If the examples are sparse in the appearance domain, this
may not lead to effective groupings.
If examples are provided in sequence and are based on observations from an object with
realistic dynamics, then we can find effective groupings even if observations are sparse in
appearance space. We make the assumption that along the trajectory of example observations over time, the underlying object is likely to remain smooth and locally span regions of
appearance which are possible to interpolate. We thus perform set growing along examples
on their input trajectory. Specifically, in the results reported below, we select K seed points
on the trajectory to form initial clusters. At each point p we find the set s which is the
smallest interval on the example trajectory which contains p, has a non-zero interior region
(s -1i x (s)), and for which E / (s) < ?. If such set exists, we continue to expand it, growing
the set along the example trajectory until the above set growing criterion is violated. Once
we can no longer grow any set, we test whether any set is a proper subset of another, and
delete it if so. We keep the remaining sets, and use them for interpolation as described
below.
4 Synthesis using example sets
We generate new views using sets of examples: interpolation is restricted to only occur
inside the convex hull of an example set found as above for which E/(s) ::; E. Given a new
parameter vector x, we test whether it is in the convex hull of parameters in any example
set. If the point does not lie in the convex hull of any example set, we find the nearest point
on the convex hull of one of the example sets, and use that instead. This prevents erroneous
extrapolation.
If a new parameter is in the convex hull of more than one example set, we first select the
set whose median example parameter is closest to the desired example parameter. Once a
set has been selected, we interpolate a new function value from examples using the RBF
method summarized above. To enforce temporal consistency of rendered images over time,
770
T. Darrell
(b)
(c)
Figure 2: (a) Images of a real arm (from a sequence of 33 images) with changing appearance and elbow configuration. (b,c) Interpolated shape of arms tracked in previous figure.
(b) shows results using all examples in a single interpolation network; (c) shows results
using example sets algorithm. Open contours show arms example locations; filled contour shows interpolation result. Near regions of appearance singularity in parameter space
the full network method generates physically-invalid arm shapes; the example sets method
produces realistic images.
The method presented below for grouping examples into locally valid spaces is generally
applicable to both the PCA and RBF-based view synthesis techniques. However our initial
implementation, and the results reported in this paper, have been with RBF-based models.
3
Finding consistent example sets
Given examples from a complicated (non-linear, non-smooth) appearance mapping, we find
local regions of appearance which are well-behaved as smooth, possibly linear, functions.
We wish to cluster our examples into sets which can be used for successful interpolation
using our local appearance mode\.
Conceptually, this problem is similar to that faced by Bregler and Omohundro [2], who
built image manifolds using a mixture of local PCA models. Their work was limited to
modeling shape (lip outlines); they used K-means clustering of image appearance to form
the initial groupings for PCA analysis. However this approach had no model of texture and
performed clustering using a mean-squared-error distance metric in simple appearance.
Simple appearance clustering drastically over-partitions the appearance space compared to
a model that jointly represent shape and texture. Examples which are distant in simple
appearance can often be close when considered in 'vectorized' representation. Our work
771
Example-Based Image Synthesis ofArticulated Figures
",
'"
"
.. : ... .
....
. ..
.. .
...
",
"
..............
...... '.. .
'"
(b) ,-------.c'-,-' '_'
. _' ' _--,
(C)'----________
"
"
"
.
..
.. ..
.:.... ..
,
'
:': ',::'::
_ _ _ _ _ _ _ ___"
,
'"
.. ..
. . : ....
- ' ,
'
0
?
??
:::~:.:.: :-:.:
'
0
'~1.)'.
..... .
.. .
'
'-----"--'---'--_ _ _ _---' '------"--'---'--_ _--'
~L
"
"
'------"~_ _- '
'_____ _~_ _ _ _~ '-----'_ _ _ _ _ _- - '
..'
.
'-----"--'---'--_ _ _ _---'
'----~
_ _ _ _ _ _--'
Figure 1: Arm appearance interpolated from examples using approximation network. (a)
A 2DOF planar arm. Discontinuities in appearance due to workspace constraints make
this a difficult function to learn from examples; the first and last example are very close
in parameter space, but far in appearance space. (b) shows results using all examples
in a single network; (c) using the example sets algorithm described in text. Note poor
approximation on last two examples in (a); appearance discontinuities and extrapolation
cause problems for full network, but are handled well in examples sets method.
In peA-based approaches, G projects a portion of u onto a optimal linear subspace found
from D, and F projects a portion of u onto a subspace found from T [6, 5]. For example
G D (u) = PI) 59 U , where 59 is a diagonal boolean matrix which selects the texture parameters from u and PI) is a matrix containing the m-th largest principle components of D .
F warps the reconstructed texture according to the given shape: FT(u, s) = [PT 5 t u] 0 s.
While interpolation is simple using a peA approach, the parameters used in peA models
often do not have any direct physical interpretation. For the task of view synthesis, an additional mapping u = H(x) is needed to map from task parameters to peA input values;
a backpropogation neural net was used to perform this function for the task of eye gaze
analysis [10].
Using the RBF-based approach [1], the application to view synthesis is straightforward.
Both G and F are networks which compute locally-weighted regression, and parameters
are used directly (u = x) . G computes an interpolated shape, and F warps and blends the
example texture images according to that shape: G D(X) = Ei cd(x - xd, FT(X, s) =
[Ei cU(x - Xi)] os , where f is a radial basis function. The coefficients c and c' are derived
from D and T, respectively: C = D R+ , where rij = f (x i-X j) and C is the matrix of row
vectors Ci; similarly C' = T R+ [9] . We have found both vector norm and Gaussian basis
functions give good results when appearance data is from a smooth function; the results
below use f(r) = Ilrll.
772
T. Darrell
ever, these methods are limited in the types of object appearance they can accurately model.
PCA-based face analysis typically assumes images of face shape and texture fall in a linear
subspace; RBF approaches fare poorly when appearance is not a smooth function.
We want to extend non-rigid interpolation networks to handle cases where appearance is
not a linear manifold and is not a smooth function, such as with articulated bodies. The
mapping from parameter to appearance for articulated bodies is often one-to-many due to
the multiple solutions possible for a given endpoint. It will also be discontinuous when
constraints call for different solutions across a boundary in parameter space, such as the
example shown in Figure 1.
Our approach represents an appearance mapping as a set of piecewise smooth functions.
We search for sets of examples which are well approximated by the examples on the convex
hull of the set's parameter values. Once we have these 'safe' sets of examples we perform
interpolation using only the examples in a single set.
The clear advantage of this approach is that it will prevent inconsistent examples from
being combined during interpolation. It also can reduce the number of examples needed to
fully interpolate the function, as only those examples which are on the convex hull of one
or more example sets are needed. If a new example is provided and it falls within and is
well-approximated by the convex hull of an existing set, it can be safely ignored.
The remainder of this paper proceeds as follows. First, we will review methods for modeling appearance when it can be well approximated with a smooth and/or linear function.
Next, we will present a technique for clustering examples to find maximal subsets which
are well approximated in their interior. We will then detail how we select among the subsets
during interpolation, and finally show results with both synthetic and real imagery.
2 Modeling smooth and/or linear appearance functions
Traditional interpolation networks work well when object appearance can be modeled either as a linear manifold or as a smooth function over the parameters of interest (describing
pose, expression, identity, configuration, etc.). As mentioned above, both peA and RBF
approaches have been successfully applied to model facial expression.
In both approaches, a key step in modeling non-rigid shape appearance from examples is
to couple shape and texture into a single representation. Interpolation of shape has been
well studied in the computer graphics literature (e.g., splines for key-frame animation) but
does not alone render realistic images. PCA or RBF models of images without a shape
model can only represent and interpolate within a very limited range of pose or object
configuration.
In a coupled representation, texture is modeled in shape-normalized coordinates, and shape
is modeled as disparity between examples or displacement from a canonical example to all
examples. Image warping is used to generate images for a particular texture and shape.
Given a training set n = {(Yi, Xi, d i ), 0 ~ i ~ n}, where Yi is the image of example i,
Xi is the associated pose or configuration parameter, and di is a dense correspondence map
relative to a canonical pose, a set of shape-aligned texture images can be computed such
that texture ti warped with displacement di renders example image Yi: Yi = ti 0 d i [5, 1,6].
A new image is constructed using a coupled shape model G and texture model F, based on
input u:
y(n,U) = FT(GD(U),u) ,
where D, T are the matrices [dodl ... d n ], [totl ... t n ], respectively.
Example-Based Image Synthesis ofArticulated Figures
773
(c
(b)
Figure 3: Interpolated shape and texture result. (a) shows exemplar contours (open) and
interpolated shape (filled). (b) shows example texture images. (c) shows final interpolated
image.
we can use a simple additional constraint on subsequent frames. Once we have selected
an example set, we keep using it until the desired parameter value leaves the valid region
(convex hull) of that set. When this occurs, we allow transitions only to "adjacent" example
sets; adjacency is defined as those pairs of sets for which at least one example on each
convex hull are sufficiently close (11Yi - Yj II < E) in appearance space.
S Results
First we show examples using a synthetic arm with several workspace constraints. Figure
l(a) shows examples of a simple planar 2DOF ann and the inverse kinematic solution for a
variety of endpoints. Due to an artificial obstacle in the world, the ann is forced to switch
between ann-up and ann-down configurations to avoid collision.
We trained an interpolation network using a single RBF to model the appearance of the ann
as a function of endpoint location. Appearance was modeled as the vector of contour point
locations, obtained from the synthetic ann rendering function. We first trained a single RBF
network on a dense set of examples of this appearance function. Figure l(b) shows results
interpolating new ann images from these examples; results are accurate except where there
are regions of appearance discontinuity due to workspace constraints, or when the network
extrapolates erroneously.
We applied our clustering method described above to this data, yielding the results shown
in Figure 1(c). None of the problems with discontinuities or erroneous extrapolation can
be seen in these results, since our method enforces the constraint that an interpolated result
must be returned from on or within the convex hull of a valid example set.
Next we applied our method to the images of real anns shown in Figure 2(a). Ann contours
were obtained in a sequence of 33 such images using a semi-automated defonnable contour
tracker augmented with a local image distance metric [3]. Dense correspondences were interpolated from the values on the contour. Figure 2(b) shows interpolated ann shapes using
a single RBF on all examples; dramatic errors can be seen near where multiple different
774
T. Darrell
appearances exist within a small region of parameter space.
Figure 2( c) shows the results on the same points using sets of examples found using our
clustering method; physically realistic arms are generated in each case. Figure 3 shows the
final interpolated result rendered with both shape and texture.
6
Conclusion
View-based image interpolation is a powerful paradigm for generating realistic imagery
without full models of the underlying scene geometry. Current techniques for non-rigid
interpolation assume appearance is a smooth function. We apply an example clustering
approach using on-line cross validation to decompose a complex appearance mapping into
sets of examples which can be smoothly interpolated. We show results on real imagery
of human arms, with correspondences recovered from deformable contour tracking. Given
images of an arm moving on a plane with various configuration conditions (elbow up and
elbow down), and with associated parameter vectors marking the hand location, our method
is able to discover a small set of manifolds with a small number of exemplars each can
render new examples which are always physically correct. A single interpolating manifold
for this same data has errors near the boundary between different arm configurations, and
where multiple images have the same parameter value.
References
[1] D. Beymer, A. Shashua and T. Poggio, Example Based Image Analysis and Synthesis, MIT AI
Lab Memo No. 1431, MIT, 1993. also see D. Beymer and T. Poggio, Science 272:1905-1909,
1996.
[2] C. Breg1er and S. Omohundro, Nonlinear Image Interpolation using Manifold Learning, NIPS7, MIT Press, 1995.
[3] T. DarrelI, A Radial Cumulative Similarity Transform for Robust Image Correspondence, Proc.
CVPR-98. Santa Barbara, CA, IEEE CS Press, 1998.
[4] M. Jagersand, Image Based View Synthesis of Articulated Agents, Proc. CVPR-97, San Jaun,
Pureto Rico, pp. 1047-1053, IEEE CS Press, 1997.
[5] M. Jones and T Poggio, Multidimensional Morphable Models, Proc. ICCV-98, Bombay, India,
pp. 683-688, 1998.
[6] A. Lanitis, C.J. Taylor, TF. Cootes, A Unified Approach to Coding and Interpreting Fa:::e Images, Proc. ICCV-95, pp. 368-373, Cambridge, MA, 1995.
[7] M. Levoy and P. Hanrahan, Light Field Rendering, In SIGGRAPH-96, pp. 31-42,1996.
[8] L. McMillan and G. Bishop, Plenoptic Modeling: An image-based rendering system. In Proc.
SIGGRAPH-95, pp. 39-46, 1995.
[9] T Poggio and F. Girosi , A Theory of Networks for Approximation and Learning, MIT AI Lab
Memo No. 1140. 1989.
[10] T Rikert and M. Jones, Gaze Estimation using Morphable Models, Proc. IEEE Conf. Face and
Gesture Recognition '98, pp. 436-441, Nara, Japan, IEEE CS Press, 1998.
[II] L. Saul and M. Jordan, A Variational Principle for Model-based Morphing, NIPS-9, MIT Press,
1997.
[12] S. Seitz and C. Dyer, View Morphing, in Proc. SIGGRAPH-96, pp. 21-30,1996.
[13] A. Shashua and M. Werman, Trilinearity of Three Perspective Views and its Associated Tensor,
in Proc. ICCV-95, pp. 920-935, Cambridge, MA, IEEE CS Press, 1995.
[14] 1. Tenenbaum, Mapping a manifold of perceptual observations, NIPS-IO, MIT Press, 1998.
| 1504 |@word cu:1 norm:1 open:2 willing:1 seitz:1 dramatic:1 initial:3 configuration:7 contains:1 disparity:1 offering:1 existing:1 current:2 com:2 recovered:1 must:2 distant:1 subsequent:1 realistic:6 partition:1 shape:23 girosi:1 alone:2 selected:2 leaf:1 plane:1 location:4 along:3 constructed:1 direct:1 inside:1 nor:1 growing:5 elbow:3 provided:2 project:2 underlying:2 discover:1 alto:1 unified:1 finding:1 transformation:1 temporal:1 safely:1 multidimensional:1 ti:2 xd:1 local:4 io:1 interpolation:22 studied:1 limited:3 range:1 camera:1 enforces:2 yj:1 procedure:1 displacement:2 road:1 radial:3 onto:2 interior:4 close:3 www:1 map:2 straightforward:1 attention:1 independently:1 convex:15 recovery:1 handle:1 notion:1 coordinate:1 pt:1 approximated:7 recognition:1 ft:3 rij:1 region:9 mentioned:1 dynamic:1 trained:2 basis:3 siggraph:3 various:1 articulated:6 forced:1 effective:2 artificial:1 dof:2 whose:2 cvpr:2 ability:1 jointly:1 transform:1 final:2 sequence:3 advantage:1 net:1 maximal:1 remainder:1 aligned:1 date:1 poorly:1 deformable:1 cluster:3 darrell:4 produce:1 generating:2 object:9 pose:6 exemplar:2 nearest:2 c:4 safe:1 discontinuous:1 correct:1 hull:15 pea:6 human:1 adjacency:1 require:2 decompose:1 singularity:1 bregler:1 sufficiently:1 considered:1 tracker:1 seed:2 mapping:8 werman:1 smallest:1 estimation:1 proc:8 applicable:1 palo:1 largest:1 tf:1 successfully:2 weighted:1 mit:6 gaussian:1 always:1 avoid:1 varying:2 derived:1 rigid:5 typically:1 compactness:1 expand:1 selects:1 among:1 field:1 once:4 represents:1 look:1 jones:2 spline:1 piecewise:1 interpolate:5 geometry:1 interest:1 kinematic:1 mixture:1 yielding:1 light:1 accurate:1 poggio:4 facial:1 unless:1 filled:2 taylor:1 desired:2 delete:1 modeling:6 boolean:1 obstacle:1 bombay:1 cover:1 subset:5 successful:2 graphic:2 reported:2 lanitis:1 synthetic:3 combined:1 gd:1 workspace:3 gaze:2 synthesis:13 squared:1 imagery:3 containing:1 possibly:1 conf:1 warped:1 japan:1 summarized:1 coding:1 coefficient:1 performed:1 view:13 extrapolation:6 lab:2 portion:2 shashua:2 complicated:1 who:1 conceptually:1 accurately:1 produced:1 none:1 trajectory:5 anns:1 trevor:3 pp:8 associated:3 di:2 couple:1 rico:1 planar:2 until:2 hand:1 ei:2 o:1 overlapping:1 nonlinear:1 mode:1 behaved:1 effect:1 requiring:1 normalized:1 adjacent:1 during:2 criterion:2 outline:1 omohundro:2 interpreting:1 image:42 variational:1 physical:1 tracked:1 endpoint:3 extend:1 interpretation:1 fare:1 cambridge:2 backpropogation:1 ai:2 consistency:1 similarly:1 had:1 moving:1 longer:1 similarity:1 morphable:2 etc:1 add:3 closest:1 recent:1 perspective:1 barbara:1 continue:1 yi:5 hanrahan:1 seen:2 additional:2 paradigm:1 ii:2 semi:1 full:3 multiple:3 smooth:13 gesture:1 cross:1 nara:1 basic:1 regression:1 vision:1 metric:2 physically:4 represent:2 want:1 interval:4 grow:1 median:1 inconsistent:1 jordan:1 call:1 ee:1 near:3 rendering:3 variety:1 switch:1 automated:1 reduce:1 whether:2 expression:3 pca:5 handled:1 render:4 returned:1 cause:1 ignored:1 generally:1 collision:1 detailed:1 clear:1 santa:1 locally:3 tenenbaum:1 http:1 generate:2 exist:1 canonical:2 key:2 changing:1 prevent:1 inverse:1 powerful:1 cootes:1 extends:1 correspondence:4 extrapolates:1 occur:2 constraint:6 scene:1 nearby:1 interpolated:12 generates:1 erroneously:1 span:1 rendered:2 marking:1 according:2 poor:1 remain:1 across:1 restricted:1 iccv:3 describing:1 precomputed:1 fail:1 needed:3 dyer:1 apply:1 enforce:1 assumes:1 remaining:1 clustering:9 tensor:2 warping:1 occurs:1 blend:1 fa:1 traditional:2 diagonal:1 subspace:3 distance:3 simulated:1 manifold:9 modeled:4 difficult:1 memo:2 implementation:1 proper:1 perform:3 observation:4 ever:1 frame:2 mcmillan:1 pair:1 discontinuity:4 nip:2 able:2 proceeds:2 below:4 built:1 including:1 max:1 rely:1 arm:12 eye:1 coupled:3 faced:1 morphing:3 literature:2 text:1 review:1 relative:1 fully:1 validation:1 agent:1 vectorized:2 consistent:1 principle:3 pi:2 cd:1 row:1 last:2 free:1 drastically:1 allow:1 warp:2 india:1 neighbor:1 fall:2 face:4 saul:1 sparse:2 boundary:2 depth:1 valid:4 transition:1 contour:8 computes:1 world:1 cumulative:1 levoy:1 san:1 far:1 reconstructed:1 keep:2 xi:3 search:1 lip:1 learn:1 robust:1 ca:2 exterior:2 mse:1 complex:2 necessarily:1 constructing:1 domain:1 interpolating:2 dense:3 animation:1 body:2 augmented:1 wish:1 lie:2 candidate:1 perceptual:1 ix:1 down:2 erroneous:2 bishop:1 grouping:4 exists:1 ci:1 texture:16 illumination:1 jagersand:1 smoothly:1 mill:1 appearance:42 likely:1 beymer:2 prevents:1 tracking:1 ma:2 identity:2 ann:9 rbf:11 invalid:1 specifically:1 except:1 select:3 plenoptic:2 violated:1 |
553 | 1,505 | Call-based Fraud Detection in Mobile
Communication Networks using a Hierarchical
Regime-Switching Model
Jaakko Hollmen
Helsinki University of Technology
Lab. of Computer and Information Science
P.O. Box 5400, 02015 HUT, Finland
laakko.Hollmen@hut.fi
Volker Tresp
Siemens AG, Corporate Technology
Dept. Information and Communications
81730 Munich, Germany
Volker.Tresp@mchp.siemens.de
Abstract
Fraud causes substantial losses to telecommunication carriers. Detection systems which automatically detect illegal use of the network can be
used to alleviate the problem. Previous approaches worked on features
derived from the call patterns of individual users. In this paper we present
a call-based detection system based on a hierarchical regime-switching
model. The detection problem is formulated as an inference problem on
the regime probabilities. Inference is implemented by applying the junction tree algorithm to the underlying graphical model. The dynamics are
learned from data using the EM algorithm and subsequent discriminative
training. The methods are assessed using fraud data from a real mobile
communication network.
1
INTRODUCTION
Fraud is costly to a network carrier both in terms of lost income and wasted capacity. It has
been estimated that the telecommunication industry looses approximately 2-5% of its total
revenue to fraud. The true losses are expected to be even higher since telecommunication
companies are reluctant to admit fraud in their systems. A fraudulent attack causes lots of
inconveniences to the victimized subscriber which might motivate the subscriber to switch
to a competing carrier. Furthermore, potential new customers would be very reluctant to
switch to a carrier which is troubled with fraud.
Mobile communication networks -which are the focus of this work- are particularly
appealing to fraudsters as the calling from the mobile terminal is not bound to a physical
place and a subscription is easy to get. This provides means for an illegal high-profit
business requiring minimal investment and relatively low risk of getting caught. Fraud is
890
J.
Hollmen and V. Tresp
usually initiated by a mobile phone theft, by cloning the mobile phone card or by acquiring
a subscription with false identification. After intrusion the subscription can be used for
gaining free services either for the intruder himself or for his illegal customers in form of
call-selling. In the latter case, the fraudster sells calls to customers for reduced rates.
The earliest means of detecting fraud were to register overlapping calls originating from
one subscription, evidencing card cloning. While this procedure efficiently detects cloning,
it misses a large share of other fraud cases. A more advanced system is a velocity trap which
detects card cloning by using an upper speed limit at which a mobile phone user can travel.
Subsequent calls from distant places provide evidence for card cloning. Although a velocity
trap is a powerful method of detecting card cloning, it is ineffective against other types of
fraud. Therefore there is great interest in detection systems which detect fraud based on
an analysis of behavioral patterns (Barson et aI., 1996, Burge et aI., 1997, Fawcett and
Provost, 1997, Taniguchi et aI., 1998).
In an absolute analysis, a user is classified as a fraudster based on features derived from
daily statistics summarizing the call pattern such as the average number of calls. In a differential analysis, the detection is based on measures describing the changes in those features
capturing the transition from a normal use to fraud. Both approaches have the problem
of finding efficient feature representations describing normal and fraudulent behavior. As
they usually derive features as summary statistics over one day, they are plagued with a
latency time of up to a day to detect fraudulent behavior. The resulting delay in detection
can already lead to unacceptable losses and can be exploited by the fraudster. For these
reasons real-time fraud detection is seen to be the most important development in fraud
detection (Pequeno, 1997).
In this paper we present a real-time fraud detection system which is based on a stochastic
generative model. In the generative model we assume a variable victimized which indicates
if the account has been victimized by a fraudster and a second variable fraud which indicates if the fraudster is currently performing fraud. Both variables are hidden. Furthermore,
we have an observed variable call which indicates if a call being is performed or not. The
transition probabilities from no-call to call and from call to no-call are dependent on the
state of the variable fraud. Overall, we obtain a regime-switching time-series model as described by Hamilton (1994), with the modifications that first, the variables in the time series
are not continuous but binary and second, the switching variable has a hierarchical structure. The benefit of the hierarchical structure is that it allows us to model the time-series
at different time scales. At the lowest hierarchical level, we model the dynamical behavior
of the individual calls, at the next level the transition from normal behavior to fraudulent
behavior and at the highest level the transition to being victimized. To be able to model a
time-series at different temporal resolutions was also the reason for introducing a hierarchy
into a hidden Markov model for Jordan, Ghahramani and Saul (1997). Fortunately, our
hidden variables have only a small number of states such that we do not have to work with
the approximation techniques those authors have introduced.
Section 2 introduces our hierarchical regime-switching fraud model. The detection problem is formulated as an inference problem on the regime probabilities based on subscriber
data. We derive iterative algorithms for estimating the hidden variables fraud and victimized based on past and present data (filtering) or based on the complete set of observed
data (smoothing). We present EM learning rules for learning the parameters in the model
using observed data. We develop a gradient based approach for fine tuning the emission
probabilities in the non-fraud state to enhance the discrimination capability of the model.
In Section 3 we present experimental results. We show that a system which is fine-tuned on
real data can be used for detecting fraudulent behavior on-line based on the call patterns.
In Section 4 we present conclusions and discuss further applications and extensions of our
fraud model.
Fraud Detection Using a Hierarchical Regime-Switching Model
2
891
THE HIERARCHICAL REGIME-SWITCHING FRAUD
MODEL
2.1
THE GENERATIVE MODEL
The hierarchical regime-switching model consists of three variables which evolve in time
stochastically according to first-order Markov chains. The first binary variable Vt (victimized) is equal to one if the account is currently being victimized by a fraudster and zero
otherwise. The states of this variable evolve according to the state transition probabiliP(Vt = ilVt_l
j); i,j
0,1. The second binary variable St (fraud) is
ties pij
equal to one if the fraudster currently performs fraud and is equal to zero if the fraudster
is inactive. The change between actively performing fraud and intermittent silence is typical for a victimized account as is apparent from Figure 3. Note that this transient bursty
behavior of a victimized account would be difficult to capture with a pure feature based
approach. The states of this variable evolve following the state transition probabilities
pfjk = P(St
ilvt
j,St-l = k,);i,j,k
0,1. Finally, the binary variable Yt (call)
is equal to one if the mobile phone is being used and zero otherwise with state transition
matrix pfjk
P(Yt ilst j, Yt-l k); i, j, k = 0,1. Note that this corresponds to the
assumption of exponentially distributed call duration. Although not quite realistic, this is
the general assumption in telecommunications. Typically, both the frequency of calls and
the lengths of the calls are increased when fraud is executed. The joint probability of the
time series up to time T is then
=
=
=
=
=
=
= =
P(VT' ST, YT) = P(vo, So, Yo)
=
=
T
T
T
t=l
t=l
t=l
II P(Vt!Vt-l) II P(stlvt, St-l) II P(Ytlst, Yt-l)
(1)
where in the experiments we used a sampling time of one minute. Furthermore, VT
{vo, ... , VT }, ST = {so, ... , ST }, YT = {Yo, ... , YT} and P(vo, So, Yo) is the prior distribution of the initial states.
Figure 1: Dependency graph of the hierarchical regime-switching fraud model. The square
boxes denote hidden variables and the circles observed variables. The hidden variable Vt
on the top describes whether the subscriber account is victimized by fraud. The hidden
variable St indicates if fraud is currently being executed. The state of St determines the
statistics of the variable call Yt.
2.2
INFERENCE: FILTERING AND SMOOTIDNG
When using the fraud detection system, we are interested to estimate the probability that
an account is victimized or that fraud is currently occurring based on the call patterns up to
the current point in time (filtering). We can calculate the probabilities of the states of the
hidden variables by applying the following equations recursively with t = 1, ... , T.
J. Hol/men and V Tresp
892
P(Vt
= i, St-1 = k!Yt-1) = '2::prlP(Vt-1 = l, St-1 = kIYt- 1)
I
= j!Yt-1) = '2:: PjikP(Vt = i, St-1 = kIYt-1)
P(Vt = i, St
k
where c is a scaling factor. These equations can be derived from the junction tree algorithm
for the Bayesian networks (Jensen, 1996). We obtain the probability of victimization and
fraud by simple marginalization
P(Vt
= ilYt ) = L
= i, = jlYr) ; P(St = jlYd = L
P(Vt
St
P(Vt
= i, St = jlYd?
i
j
In some cases -in particular for the EM learning rules in the next section- we might
be interested in estimating the probabilities of the hidden states at some time in the past
(smoothing). In this case we can use a variation of the smoothing equations described in
Hamilton (1994) and Kim (1994). After performing the forward recursion, we can calculate
the probability of the hidden states at time tf given data up to time T > tf iterating the
following equations with t = T, T - 1, ... ,1.
P(Vt+1
"""' P(Vt+1 = k,St+1 = lIYT)
P(
_ k
-ll?') P(Vt+l
Vt+1 - ,St+1 t
.
= k,St = JIYT) = ~
I
,
?IV)
P( Vt=z,St=JI.T
2.3
.
s
= k,St = JIYt}Plkj
"""' P(Vt+1 = k,St =jIYT)p(
.
?Iv") v
P(
-k
_ 'I?,)
Vt=z,St=)I.tPki
k
Vt+1 - ,St - J t
=~
EM LEARNING RULES
Parameter estimation in the regime-switching model is conveniently formulated as an incomplete data problem, which can be solved using the EM algorithm (Hamilton, 1994).
Each iteration of the EM algorithm is guaranteed to increase the value of the marginal loglikelihood function until a fixed point is reached. This fixed point is a local optimum of the
marginal log-likelihood function.
In the M-step the model parameters are optimized using the estimates of the hidden states
using the current parameter estimates. Let 0
{prj, Pijk' P;kj} denote the current parameter estimates. The new estimates are obtained using
=
2:;=1 P( Vt
v
Pij =
s
Pijk =
Y
Pikj
=
= i, Vt-1 = jIYT; 0)
",T
(
'I ,"
L....t=l P Vt-1 = J }T; 0)
2:;-1 P(St = i, Vt = j, St-1 = kIYT; 0)
T
.
2:t=l P( Vt = J, St-1 = kIYT; 0)
2:;=l,if Yt=i andYt_l=j P(St-1 = kIYT;O)
T
2:t=l, if Yt-l=j P(St-1 kIYT; 0)
=
Fraud Detection Using a Hierarchical Regime-Switching Model
893
The E-step determines the probabilities on the right sides of the equations using the current
parameter estimates. These can be determined using the smoothing equations from the
previous section directly by marginalizing
P(Vt
= k,
St
= l, Vt+l = i, St+1 = jIYT)
where the terms on the right side are obtained from the equations in the last Section.
2.4
DISCRIMINATIVE TRAINING
In our data setting, it is not known when the fraudulent accounts were victimized by fraud.
This is why we use the EM algorithm to learn the two regimes from data in an incomplete data setting. We know, however, which accounts were victimized by fraud. After EM
learning the discrimination ability of the model was not satisfactory. We therefore used
the labeled sequences to improve the model. The reason for the poor performance was
credited to unsuitable call emission probabilities in the normal state. We therefore mini-
L:i (maXt P(v!i)I}~(i)) - t(i?)2 with regard to the parameter
mize the error function E
P;=O,j=O ,k=O' where the t(i) = {O, I} is the label for the sequence i. The error function
was minimized with Quasi-Newton procedure with numerical differentiation.
=
3
EXPERIMENTS
To test our approach we used a data set consisting of 600 accounts which were not affected
by fraud and 304 accounts which were affected by fraud. The time period for non-fraud and
fraud accounts were 49 and 92 days, respectively. We divided the data equally into training
data and test data. From the non-fraud data we estimated the parameters describing the
normal calling behavior, i.e. pr,j =O,k' Next, we fixed the probability that an account is
victimized from one time step to the next to PY=l,j=O = 10- 5 and the probability that
a victimized account becomes de-victimized as pi=O,j=l = 5 X 10- 4 ? Leaving those
parameters fixed the remaining parameters were trained using the fraudulent accounts and
the EM algorithm described in Section 2. We had to do unsupervised training since it was
known by velocity check that the accounts were affected but it was not clear when the
intrusion occurred. After unsupervised training, we further enhanced the discrimination
capability of the system which helped us reduce the amount of false alarms. The final
model parameters can be found in the Appendix.
After training, the system was tested using the test data. Unfortunately, it is not known
when the accounts were attacked by fraud, but only on per-account basis if an account was
at some point a victim of fraud. Therefore, we declare an account to be victimized if the
victimized variable at some point exceeds the threshold. Also, it is interesting to study the
results shown in Figure 3. We show data and posterior time-evolving probabilities for an
account which is known to be victimized. From the call pattern it is obvious that there are
periods of suspiciously high traffic at which the probability of victimization is recognized
to be very high. We also see that the variable fraud St follows the bursty behavior of
the fraudulent behavior correctly. Note, that for smoothing which is important both for
a retrospective analysis of call data and for learning, we achieve smoother curves for the
victimized variable.
J. Hollmen and V. Tresp
894
??
..
,- - '
,
"
00
aDI
01?
OeD
O.Dt
OIlS
0,08
om
o.oa
001
0.1
O.D1
0,(2
OeD
001
0.05
001
oar
Q.O&
001
01
Figure 2: The Receiver Operating Characteristic (ROC) curves are shown for on-line detection (left figure) and for retrospective classification (right figure). In the figures, detection
probability is plotted against the false alarm probability. The dash-dotted lines are results
before, the solid lines after discriminative training. We can see that the discriminative
training improves the model considerably.
After EM training and discriminative training, we tested the model both in on-line detection mode (filtering) and in retrospective classification (smoothing) with smoothed probabilities. The detection results are shown in Figure 2. With a fixed false alarm probability
of 0.003, the detection probabilities for the training set were found to be 0.974 and 0.934
using on-line detection mode and with smoothed probabilities, respectively. With a testing
set and a fixed false alarm probability of 0.020, we obtain the detection probabilities of
0.928 and 0.921, for the on-line detection and for retrospective classification, respectively.
4 CONCLUSIONS
We presented a call-based on-line fraud detection system which is based on a hierarchical regime-switching generative model. The inference rules are obtained from the junction
tree algorithm for the underlying graphical model. The model is trained using the EM algorithm in an incomplete data setting and is further refined with gradient-based discriminative
training, which considerably improves the results.
A few extensions are in the process of being implemented. First of all , it makes sense to
use more than one fraud model for the different fraud scenarios and several user models
to account for different user profiles. For these more complex models we might have to
rely on approximations techniques such as the ones introduced by Jordan, Ghahramani and
Saul (1997).
Appendix
The model parameters after EM training and discriminative training. Note that entering the
fraud state without first entering the victimized state is impossible.
0.9559
pY. . ,)=
. Ok
, = ( 0.3533
1.0000
pi,j=O,k = ( 0.0000
0.0441 )
0.6467
0.0000 )
1.0000
0.9292
p?t ,J=
. 1 , k = ( 0.0570
0.9979
pi,j=l ,k
0.0086
=(
0.0708 )
0.9430
0.0021 )
0.9914
895
Fraud Detection Using a Hierarchical Regime-Switching Model
_-o:t 111111111111U 11111] 111I111111[IIIIIIIIJIillIIIIIIII
0.5
1.5
~o:~:
1.5
:\J
2
2.5
1 .5
2
2.5
0.5
0.5
1
:
2
j
1
2 .5
Figure 3: The first line shows the calling data Yt from a victimized account. The second
and third lines show the states of the victimized and fraud variables, respectively. Both are
calculated with the filtering equations. The fourth and fifth lines show the same variables
using the smoothing equations. The displayed time window period is seventeen days.
References
Barson P., Field, S., Davey, N., McAskie, G., and Frank, R. (1996). The Detection of Fraud
in Mobile Phone Networks. Neural Network World, Vol. 6, No.4.
Bengio, Y. (1996). Markovian Models for Sequential Data. Technical Report # 1049,
Universite de Montreal.
Burge, P., Shawe-Taylor J., Moreau Y., Verrelst, H., Stormann C. and Gosset, P. (1997).
BRUTUS - A Hybrid Detection Tool. Proc. of ACTS Mobile Telecommunications Summit,
Aalborg, Denmark.
Fawcett, T. and Provost, F. (1997). Adaptive Fraud Detection. Journal of Data Mining and
Knowledge Discovery, , Vol. I, No.3, pp. 1-28.
Hamilton, J. D. (1994). Time Series Analysis. Princeton University Press.
Jensen, Finn V. (1996). Introduction to Bayesian Networks. UCL Press.
Jordan, M. I, Ghahramani, Z. and Saul, L. K. (1997). Hidden Markov Decision Trees, in
Advances in Neural Information Processing Systems: Proceedings of the 1996 Conference
(NIPS'9), MIT-Press, pp. 501-507.
Kim, c.-J. (1994). Dynamical linear models with Markov-switching. Journal of Econometrics, Vol. 60, pp. 1-22.
Pequeno, K. A.(1997). Real-Time fraud detection: Telecom's next big step. Telecommunications (America Edition), Vol. 31, No.5, pp. 59-60.
Taniguchi, M., Haft, M., Hollmen, J. and Tresp, V. (1998). Fraud detection in communications networks using neural and probabilistic methods. Proceedings of the 1998 IEEE Int.
Con! in Acoustics, Speech and Signal Processing (ICASSP'98). Vol. 2. pp. 1241-1244.
| 1505 |@word verrelst:1 subscriber:4 profit:1 solid:1 recursively:1 initial:1 series:6 tuned:1 past:2 current:4 distant:1 realistic:1 numerical:1 subsequent:2 discrimination:3 generative:4 inconvenience:1 provides:1 detecting:3 attack:1 unacceptable:1 differential:1 consists:1 behavioral:1 expected:1 behavior:10 terminal:1 detects:2 automatically:1 company:1 window:1 becomes:1 estimating:2 underlying:2 lowest:1 loos:1 ag:1 finding:1 differentiation:1 temporal:1 act:1 tie:1 hamilton:4 carrier:4 service:1 declare:1 local:1 before:1 haft:1 limit:1 switching:14 initiated:1 approximately:1 credited:1 might:3 testing:1 lost:1 investment:1 procedure:2 evolving:1 illegal:3 fraud:56 get:1 risk:1 applying:2 impossible:1 py:2 customer:3 yt:13 caught:1 duration:1 resolution:1 pure:1 rule:4 his:1 variation:1 enhanced:1 hierarchy:1 user:5 velocity:3 particularly:1 econometrics:1 summit:1 labeled:1 observed:4 solved:1 capture:1 calculate:2 highest:1 substantial:1 jaakko:1 dynamic:1 oed:2 hol:1 motivate:1 trained:2 basis:1 selling:1 icassp:1 joint:1 america:1 evidencing:1 refined:1 apparent:1 quite:1 victim:1 loglikelihood:1 tested:2 otherwise:2 ability:1 statistic:3 final:1 sequence:2 ucl:1 achieve:1 getting:1 optimum:1 prj:1 ilvt:1 derive:2 develop:1 montreal:1 implemented:2 stochastic:1 transient:1 alleviate:1 extension:2 hut:2 normal:5 plagued:1 great:1 bursty:2 finland:1 estimation:1 proc:1 travel:1 label:1 currently:5 tf:2 tool:1 mit:1 mobile:10 volker:2 earliest:1 derived:3 focus:1 emission:2 yo:3 indicates:4 cloning:6 likelihood:1 intrusion:2 check:1 kim:2 detect:3 summarizing:1 sense:1 inference:5 dependent:1 typically:1 hidden:12 originating:1 quasi:1 interested:2 germany:1 overall:1 classification:3 development:1 smoothing:7 marginal:2 equal:4 field:1 sampling:1 sell:1 unsupervised:2 minimized:1 report:1 few:1 aalborg:1 individual:2 consisting:1 detection:29 interest:1 mining:1 introduces:1 chain:1 daily:1 tree:4 iv:2 incomplete:3 taylor:1 circle:1 plotted:1 minimal:1 increased:1 industry:1 markovian:1 introducing:1 delay:1 dependency:1 considerably:2 st:32 probabilistic:1 enhance:1 admit:1 stochastically:1 actively:1 account:22 potential:1 de:3 int:1 register:1 performed:1 helped:1 lot:1 lab:1 traffic:1 reached:1 capability:2 om:1 square:1 characteristic:1 efficiently:1 identification:1 bayesian:2 classified:1 against:2 frequency:1 pp:5 obvious:1 universite:1 con:1 seventeen:1 reluctant:2 knowledge:1 improves:2 ok:1 higher:1 dt:1 day:4 box:2 furthermore:3 until:1 overlapping:1 mode:2 oil:1 requiring:1 true:1 entering:2 satisfactory:1 ll:1 subscription:4 complete:1 vo:3 performs:1 fi:1 physical:1 ji:1 exponentially:1 occurred:1 ai:3 tuning:1 shawe:1 had:1 operating:1 taniguchi:2 posterior:1 phone:5 scenario:1 binary:4 vt:30 exploited:1 seen:1 fortunately:1 recognized:1 period:3 signal:1 mize:1 ii:3 smoother:1 corporate:1 exceeds:1 technical:1 divided:1 equally:1 himself:1 iteration:1 fawcett:2 fine:2 leaving:1 ineffective:1 jordan:3 call:27 bengio:1 easy:1 switch:2 marginalization:1 competing:1 reduce:1 inactive:1 whether:1 retrospective:4 speech:1 cause:2 latency:1 iterating:1 clear:1 amount:1 reduced:1 dotted:1 estimated:2 per:1 correctly:1 vol:5 affected:3 threshold:1 wasted:1 graph:1 powerful:1 telecommunication:6 fraudulent:8 fourth:1 place:2 intruder:1 decision:1 appendix:2 scaling:1 capturing:1 bound:1 guaranteed:1 dash:1 pijk:2 worked:1 helsinki:1 calling:3 speed:1 performing:3 relatively:1 munich:1 according:2 poor:1 describes:1 em:12 appealing:1 modification:1 pr:1 equation:9 describing:3 discus:1 know:1 finn:1 junction:3 hierarchical:13 top:1 remaining:1 graphical:2 newton:1 unsuitable:1 ghahramani:3 already:1 costly:1 gradient:2 card:5 capacity:1 oa:1 reason:3 denmark:1 length:1 mini:1 difficult:1 executed:2 unfortunately:1 frank:1 upper:1 markov:4 attacked:1 displayed:1 communication:5 intermittent:1 provost:2 smoothed:2 introduced:2 optimized:1 acoustic:1 learned:1 nip:1 able:1 usually:2 pattern:6 dynamical:2 regime:15 gaining:1 business:1 rely:1 hybrid:1 recursion:1 advanced:1 improve:1 technology:2 tresp:6 kj:1 prior:1 discovery:1 evolve:3 marginalizing:1 loss:3 men:1 interesting:1 filtering:5 revenue:1 i111111:1 pij:2 share:1 pi:3 maxt:1 summary:1 last:1 free:1 silence:1 side:2 saul:3 absolute:1 mchp:1 fifth:1 benefit:1 distributed:1 regard:1 curve:2 calculated:1 transition:7 davey:1 world:1 moreau:1 author:1 forward:1 adaptive:1 income:1 receiver:1 discriminative:7 continuous:1 iterative:1 why:1 learn:1 adi:1 complex:1 big:1 alarm:4 profile:1 edition:1 telecom:1 roc:1 burge:2 third:1 minute:1 jensen:2 evidence:1 trap:2 false:5 sequential:1 occurring:1 conveniently:1 acquiring:1 corresponds:1 determines:2 formulated:3 change:2 typical:1 determined:1 miss:1 total:1 experimental:1 siemens:2 latter:1 assessed:1 dept:1 princeton:1 d1:1 |
554 | 1,506 | Applications of multi-resolution neural
networks to mammography
Clay D. Spence and Paul Sajda
Sarnoff Corporation
CN5300
Princeton, NJ 08543-5300
{cspence, psajda }@sarnoff.com
Abstract
We have previously presented a coarse-to-fine hierarchical pyramid/neural network (HPNN) architecture which combines multiscale image processing techniques with neural networks. In this
paper we present applications of this general architecture to two
problems in mammographic Computer-Aided Diagnosis (CAD).
The first application is the detection of microcalcifications. The
<:oarse-to-fine HPNN was designed to learn large-scale context information for detecting small objects like microcalcifications. Receiver operating characteristic (ROC) analysis suggests that the
hierarchical architecture improves detection performance of a well
established CAD system by roughly 50 %. The second application
is to detect mammographic masses directly. Since masses are large,
extended objects, the coarse-to-fine HPNN architecture is not suitable for this problem. Instead we construct a fine-to-coarse HPNN
architecture which is designed to learn small-scale detail structure
associated with the extended objects. Our initial results applying
the fine-to-coarse HPNN to mass detection are encouraging, with
detection performance improvements of about 36 %. We conclude
that the ability of the HPNN architecture to integrate information
across scales, both coarse-to-fine and fine-to-coarse, makes it well
suited for detecting objects which may have contextual clues or
detail structure occurring at scales other than the natural scale of
the object.
1
Introduction
In a previous paper [8] we presented a coarse-to-fine hierarchical pyramid/neural
network (HPNN) architecture that combines multi-scale image processing tech-
Applications of Multi-Resolution Neural Networks to Mammography
939
niques with neural networks to search for small targets in images (see figure IA).
To search an image we apply the network at a position and use its output as an
estimate of the probability that a target (an object of the class we wish to find) is
present there. We then repeat this at each position in the image. In the coarseto-fine HPNN , the hidden units of networks operating at low resolution or coarse
scale learn associated context information, since the targets themselves are difficult
to detect at low resolution. The context is then passed to networks searching at
higher resolution. The use of context can significantly improve detection performance since small objects have few distinguishing features. In the HPNN each of
the networks receives information directly from only a small part of several feature
images , and so the networks can be relatively simple. The network at the highest
resolution integrates the contextual information learned at coarser resolutions to
detect the object of interest.
The HPNN architecture can be extended by considering the implications of inverting
the information flow in the coarse-to-fine architecture. This fine-to-coarse HPNN
would have networks extracting detail structure at fine resolutions of the image
and then passing this detail information to networks operating at coarser scales
(see figure IB). For many types of objects, information about the fine structure is
important for discriminating between different classes. The fine-to-coarse HPNN is
therefore a natural architecture for exploiting fine detail information for detecting
extended objects.
In this paper , we present our experiences in applying the HPNN framework to
two problems in mammographic Computer-Aided Diagnosis (CAD); that of detecting microcalcifications in mammograms and that of detecting malignant masses in
mammograms. The coarse-to-fine HPNN architecture is well-suited for the microcalcification problem , while the fine-to-coarse HPNN is suited for mass detection.
We evaluate the performance and utility of the HPNN framework by considering
its effects on reducing false positive rates in a well characterized CAD system.
The University of Chicago (UofC) has been actively developing ' mammographic
CAD systems for micro calcification and mass detection [6] and has been evaluating
their performance clinically. A general block diagram showing the basic processing
elements of these CAD systems is shown in figure 2. First, a pre-processing step
is used to segment the breast area and increase the overall signal-to-noise levels in
the image. Regions of interest (ROIs) are defined at this stage , representing local
areas of the breast which potentially contain a cluster of micro calcifications or a
mass. The next stage typically involves feature extraction and rule-based/heuristic
analysis , in order to prune false positives. The remaining ROIs are classified as
positive or negative by a statistical classifier or neural network. The CAD system
is used as a "second reader", aiding the radiologist by pointing out spots to double
check. One of the key requirements of CAD is that false positive rates be low
enough that radiologists will not ignore the CAD system output. Therefore it is
critical to reduce false positive rates of CAD systems without significant reductions
in sensitivity. In this paper we evaluate the HPNN framework within the context
of reducing the false positive rates of the UofC CAD systems for microcalcification
and mass detection. In both cases the HPNN acts as a post-processor of the UofC
CAD system.
2
Microcalcification detection
Microcalcifications are calcium deposits in breast tissue that appear as very small
bright dots in mammograms. Clusters of microcalcifications frequently occur around
tumors. Unfortunately microcalcification clusters are sometimes missed, since they
940
C. D. Spence and P. Sqjda
P(t)
P(t)
Figure 1: Hierarchical pyramid/neural network architectures for (A) detecting microcalcifications and (B) detecting masses. In (A) context is propagated from low
to high resolution via the hidden units of low resolution networks. In (B) small
scale detail information is propagated from high to low resolution. In both cases
the output of the last integration network is an estimate of the probability that a
target is present.
Mammogram
-1
Feature extraction
Pre-processing
and rule-based/
heuristic analysis
Statistical/NN
classifier
Mass or
Cluster locations
Figure 2: Block diagram for a typical CAD detection system.
can be quite subtle and the radiologists can only spend about a minute evaluating
a patient's mammograms.
Data used for the micro calcification experiments was provided by The University of
Chicago. The first set of data consists of 50 true positive and 86 false positive ROls_
These ROIs are 99x99 pixels and digiti7,ed at 100 micron resolution. A second set
of data from the UofC clinical testing database included 47 true positives and 103
false positives, also 99x99 and sampled at 100 micron resolution.
We trained the coarse-to-fine HPNN architecture in figure 1A as a detector for individual calcifications. For each level in the pyramid a network is trained, beginning
with the network at low resolution. The network at a particular pyramid level is
applied to one pixel at a time in the image at that resolution , and so produces an
output at each pixel. All of the networks are trained to detect micro calcifications,
however, at low resolutions the micro calcifications are not directly detectable. To
achieve better than chance performance, the networks at those levels must learn
something about the context in which micro calcifications appear. To integrate
context information with the other features the outputs of hidden units from low
resolution networks are propagated hierarchically as inputs to networks operating
at higher resolutions.
Input to the neural networks come from an integrated feature pyramid (IFF) [lJ.
To construct the IFP, we used steerable filters [3J to compute local orientation
energy. The steering properties of these filters enable the direct computation of
the orientation having maximum energy. We constructed features which represent,
at each pixel location, the maximum energy (energy at 8rnax) , the energy at the
Applications of Multi-Resolution Neural Networks to Mammography
cc
Az
(7 Az
HPNN
FPF
(7FPF
Az
Chicago NN
FPF
(7 Az
TPF=l.O
1
2
3
4
5
.93
.94
.94
.93
.93
.03
.02
.03
.03
.03
.24
.21
.39
.48
.51
941
(7FPF
TPF=l.O
.11
.11
.19
.15
.06
.88
.91
.91
.90
.88
.04
.02
.03
.05
.05
.50
.43
.48
.56
.68
.11
.10
.19
.21
.21
Table 1: Comparison of HPNN and Chicago networks .
orientation perpendicular to emu;]; (ernux - 90?), and the energy at the diagonal
(energy at ernux - 45 0 ).l The resulting features are input into the coarse-to-fine
network hierarchy.
In examining the truth data for the ROI data set , we found that the experts who
specified the microcalcification positions often made errors in these positions of
up to ?2 pixels of the correct position. To take this uncertainty in position into
account , we used the following error function
Euop = -
L
pEPos
log( 1 -
IT (1 xEp
y(X))) -
L
10g(1 - y(x))
(1)
x ENe y
which we have called the Uncertain Object Position (UOP) error function [7].2 (y(x)
is the network's output when applied to position x.) It is essentially the crossentropy error, but for positive examples the probability of generating a positive
output (y( x), in this case) has been replaced by the probability of generating at
least one positive output in a region or set of pixels p in the image. In our case each
p is a five-by-five pixel square centered on the location specified by the expert. To
this we added the standard weight decay regularization term. The regularization
constant was adjusted to minimize the ten-fold cross-validation error.
The coarse-to-fine HPNN was applied to each input ROI , and an image was constructed from the output of the Level 0 network at each pixel. Each of these pixel
values is the network 's estimate of the probability that a microcalcification is present
there. Training and testing were done using as jackknife protocol [5], whereby one
half of the data (25 TPs and 43 FPs) was used for training and the other half for
testing. We used five different random splits of the data into training and test sets.
For a given ROI, the probability map produced by the network was thresholded at
a given value to produce a binary detection map. Region growing was used to count
the number of distinct detected regions. The ROI was classified as a positive if the
number of regions was greater than or equal to a certain cluster criterion.
Table 1 compares ROC results for the HPNN and another network that had been
used in the University of Chicago CAD system [9] using five different cluster criterion
(cc). Reported are the area under the ROC curve (Az), the standard deviation of
A z across the subsets of the jackknife ((7 AJ, the false posit ive fraction at a true
positive fraction of 1.0 (FPF@TPF= 1.0) and the standard deviation of the FPF
across the subsets of the jackknife ((7FPF). A z and FPF@TPF = 1.0 represent
1 We found that the energies in the two diagonal directions were nearly identical.
2Keeler et al. [4] developed a network for object recognition that had some similarities
to the UOP error. In fact the way in which the outputs of units are combined for their
error function can be shown to be an approximation to the UOP error.
942
C. D. Spence and P Sajda
the averages of the subsets of the jackknife. Note that both networks operate best
when the cluster criterion is set to two. For this case the HPNN has a higher Az
than the Chicago network while also halving the false positive rate. This difference,
between the two networks ' A z and F P F values , is statistically significant (z-test;
PAz = .0018, PFPF = .00001).
A second set of data was also tested. 150 ROls taken from a clinical prospective
study and classified as positive by the full Chicago CAD system (including the
Chicago neural network) were used to test the HPNN. Though the Chicago CAD
system classified all 150 ROls as positive, only 47 were in fact positive while 103
were negatives. We applied the HPNN trained on the entire previous data set to this
new set of ROls. The HPNN was able to reclassify 47/103 negatives as negative,
without loss in sensitivity (no false negatives were introduced).
On examining the negative examples rejected by the coarse-to-fine HPNN, we found
that many of these ROls contained linear, high-contrast structure which would
otherwise be false positives for the Chicago network. The Chicago neural network
presumably interprets the "peaks" on the linear structure as calcifications. However
because the coarse-to-fine HPNN also integrates information from low resolution it
can associate these "peaks" with the low-resolution linear structure and reject them.
3
Mass detection
Although microcalcifications are an important cue for malignant masses in mammograms, they are not visible or even present in all cases. Thus mammographic CAD
systems include algorithms to directly detect the presence of masses. We have
started to apply a fine-to-coarse HPNN architecture to detect malignant masses
in digitized mammograms. Radiologists often distinguish malignant from benign
masses based on the detailed shape of the mass border and the presence of spicules
alone the border. Thus to integrate this high resolution information to detect malignant masses, which are extended objects, we apply the fine-to-coarse HPNN of
figure lB.
As for microcalcifications, we apply the HPNN as a post-processor, but hei'e it
processes the output of the mass-detection component of UofC CAD system. The
data in our study consists of 72 positive and 100 negative ROls. These are 256-by256 pixels and are sampled at 200 micron resolution.
At each level of the fine-to-coarse HPNN several hidden units process the feature
images. The outputs of each unit at all of the positions in an image make up a
new feature image. This is reduced in resolution by the usual pyramid blur-andsubsample operation to make an input feature image for the network units at the
next lower resolution. We trained the entire fine-to-coarse HPNN as one network
instead of training a network for each level, one level at a time. This training is quite
straightforward. Back-propagating error through the network units is the same
as in conventional networks. We must also back-propagate through the pyramid
reduction operation, but this is linear and therefore quite simple. In addition we
use the same UOP error function (Equation 1) used to train the coarse-to-fine
architecture. The rationale for this application of the UOP error function is that the
truth data specifies the location of the center of the mass at the highest resolution.
However, because of the sub-sampling the center cannot be unambiguously assigned
to a particular pixel at low resolution .
The features input to the fine-to-coarse HPNN are filtered versions of the image,
with filter kernels given by
.
0/'
(r e) = (
'l/q,]1'
q!
71"(q+ lp l)!
)1 / 2r IPle-r2 / 2LIP
I(r2)etp1> in polar
q
Applications of Multi-Resolution Neural Networks to Mammography
Sensitivity
100%
95%
90%
80%
Coarse-to-Fine HPNN
Microcalcification
45%
47%
63%
69%
943
Fine-to-Coarse HPNN
Mass
32%
36%
40%
78%
Table 2: Detector Specificity (% reduction in false positive rate of UofC CAD
system) .
coordinates, with (q, p) E {(O, 1) , (1,0), (0, 2)}. These are combinations of derivatives of Gaussians, and can be written as combinations of separable filter kernels
(products of purely horizontal and vertical filters) , so they can be computed at
relatively low cost. They are also easy to steer, since this is just multiplication by
a complex phase factor. We steered these in the radial and tangential directions
relative to the tentative mass centers, and used the real and imaginary parts and
their squares and products as features. The center coordinates of the are generated
by the earlier stages of the CAD system. These features were extracted at each
level of the Gaussian pyramid representation of the mass ROI, and used as inputs
only to the network units at the same level.
The fine-to-coarse HPNN is quite similar to the convolution network proposed by
Le Cun, et al [2], however with a few notable differences. The fine-to-coarse HPNN
receives as inputs preset features extracted from the image (in this case radial
and tangential gradients) at each resolution, compared to the convolution network,
whose inputs are the original pixel values at the highest resolution. Secondly, in
the fine-to- coarse HPNN , the inputs to a hidden unit at a particular position are
the pixel values at that position in each of the feature images , one pixel value per
feature image. Thus the HPN N's hidden units do not learn linear filters, except
as linear combinations of the filters used to form the features. Finally the fine-tocoarse HPNN is trained using the UOP error function , which is not used in the Le
Cun network.
Currently our best performing fine-to-coarse HPNN system for mass detection has
two hidden units per pyramid level. This gives an ROC area of A z = 0.85 and
eliminates 36 % of the false-positives at a cost of missing 5 % of the actual positives.
To improve performance further , we are investigating different regularizers, richer
feature sets, and more complex architectures, i.e., more hidden units.
4
Conclusion
We have presented the application of multi-resolution neural network architectures
to two problems in computer-aided diagnosis , the detection of micro calcifications
in mammograms and the direct detection of malignant masses in mammograms. A
summary of the performance of these architectures is given in Table 2. In the case
of microcalcifications , the coarse-to-fine HPNN architecture successfully discovered
large-scale context information that improves the system's performance in detecting
small objects. A coarse-to-fine HPNN has been directly integrated with the UofC
CAD system for micro calcification detection and the complete system is undergoing
clinical evaluation.
In the case of malignant masses, a fine-to-coarse HPNN architecture was used to
exploit information from fine resolution detail which could be used to differentiate
C. D. Spence and P Sajda
944
malignant from benign masses. The results of this network are encouraging, but additional improvement is needed. In general, we have found that the multi-resolution
HPNNs are a useful class of network architecture for exploiting and integrating information at multiple scales.
5
Acknowledgments
This work was funded by the National Information Display Laboratory, DARPA
through ONR contract No. N00014-93-C-0202, and the Murray Foundation. We
would like to thank Drs. Robert Nishikawa and Maryellen Giger of The University
of Chicago for useful discussions and providing the data.
References
[1] Peter Burt. Smart sensing within a pyramid vision machine. Proceedings IEEE,
76(8):1006-1015, 1988. Also in Neuro- Vision Systems, Gupta and Knopf, eds.,
1994.
[2] Y. Le Cun, B. Boser, J . S. Denker, and D. Henderson. Handwritten digit recognition with a back-propagation network. In David S. Touretzky, editor, Advances in Neural Information Processing Systems 2, pages 396- 404, 2929 Campus Drive, San Mateo: CA 94403 , 1991. Morgan-Kaufmann Publishers.
[3] William T. Freeman and Edward H. Adelson. The design and use of steerable
filters. IEEE Transactions on Pattern Analysis and Machine Intelligence, PAMI13(9):891- 906, 1991.
[4] James D. Keeler, David E. Rumelhart, and Wee-Keng Leow. Integrated segmentation and recognition of hand-printed numerals. In Richard P. Lippmann,
John E. Moody, and David S. Touretzky, editors, Advances in Neural Information Processing Systems 3, pages 557- 563, 2929 Campus Drive, San Mateo, CA
94403, 1991. Morgan-Kaufmann Publishers.
[5] Charles Metz. Current problems in ROC analysis. In Proceedings of the Chest
Imaging Conference, pages 315- 33, Madison, WI, November 1988.
[6] R. M. Nishikawa, R. C. Haldemann, J. Papaioannou, M. L. Giger, P. Lu, R. A.
Schmidt, D. E. Wolverton, U. Bick, and K. DoL Initial experience with a prototype clinical intelligent mammography workstation for computer-aided diagnosis. In Murray H. Loew and Kenneth M. Hanson, editors, Medical Imaging
1995, volume 2434, pages 65- 71, P .O. Box 10, Bellingham WA 98227-0010, 1995.
SPIE.
[7] Clay D. Spence. Supervised learning of detection and classification tasks with
uncertain training data. In Image Understanding Workshop. ARPA, 1996. This
Volume.
[8] Clay D. Spence, John C. Pearson, and Jim Bergen. Coarse-to-fine image search
using neural networks. In Gerald Tesauro, David S. Touretzky, and Todd K.
Leen, editors, Advances in Neural Information Processing Systems 7, pages 981988, Massachusetts Institute of Technology, Cambridge, MA 02142, 1994. MIT
Press.
[9] W. Zhang, K. Doi, M. L. Giger, Y. Wu , R. M. Nishikawa, and R. Schmidt.
Computerized detection of clustered micro calcifications in digital mammograms
using a shift-invariant artificial neural network. Medical Physics, 21(4):517-524 ,
April 1994.
| 1506 |@word version:1 propagate:1 leow:1 reduction:3 initial:2 iple:1 imaginary:1 current:1 com:1 contextual:2 cad:21 must:2 written:1 john:2 visible:1 chicago:12 blur:1 benign:2 shape:1 designed:2 alone:1 half:2 cue:1 intelligence:1 beginning:1 filtered:1 coarse:34 detecting:8 location:4 zhang:1 five:4 constructed:2 direct:2 fps:1 consists:2 uop:6 combine:2 roughly:1 themselves:1 frequently:1 growing:1 multi:7 freeman:1 encouraging:2 actual:1 considering:2 provided:1 campus:2 mass:26 developed:1 corporation:1 nj:1 act:1 classifier:2 rols:5 unit:13 medical:2 appear:2 positive:24 local:2 todd:1 mateo:2 suggests:1 sarnoff:2 perpendicular:1 statistically:1 acknowledgment:1 testing:3 spence:6 block:2 bellingham:1 digit:1 spot:1 steerable:2 area:4 significantly:1 reject:1 printed:1 pre:2 radial:2 integrating:1 specificity:1 cannot:1 context:9 applying:2 conventional:1 map:2 center:4 missing:1 straightforward:1 resolution:33 mammography:5 rule:2 searching:1 coordinate:2 target:4 hierarchy:1 distinguishing:1 associate:1 element:1 rumelhart:1 recognition:3 cspence:1 coarser:2 database:1 region:5 highest:3 gerald:1 trained:6 segment:1 smart:1 purely:1 darpa:1 sajda:3 train:1 distinct:1 doi:1 detected:1 artificial:1 pearson:1 quite:4 heuristic:2 spend:1 ive:1 whose:1 richer:1 otherwise:1 ability:1 differentiate:1 product:2 iff:1 achieve:1 az:6 exploiting:2 cluster:7 double:1 requirement:1 produce:2 generating:2 nishikawa:3 object:14 propagating:1 edward:1 involves:1 come:1 direction:2 posit:1 correct:1 filter:8 centered:1 enable:1 numeral:1 clustered:1 secondly:1 adjusted:1 keeler:2 around:1 fpf:8 roi:8 presumably:1 pointing:1 wolverton:1 polar:1 integrates:2 coarseto:1 currently:1 successfully:1 mit:1 gaussian:1 crossentropy:1 improvement:2 check:1 tech:1 contrast:1 detect:7 bergen:1 nn:2 typically:1 integrated:3 lj:1 entire:2 hidden:8 pixel:14 overall:1 classification:1 orientation:3 integration:1 mammographic:5 equal:1 construct:2 extraction:2 having:1 sampling:1 identical:1 adelson:1 nearly:1 intelligent:1 micro:9 few:2 tangential:2 richard:1 wee:1 national:1 individual:1 tpf:4 replaced:1 phase:1 william:1 detection:19 interest:2 evaluation:1 henderson:1 regularizers:1 radiologist:4 implication:1 experience:2 uncertain:2 arpa:1 earlier:1 steer:1 cost:2 deviation:2 subset:3 examining:2 paz:1 reported:1 combined:1 peak:2 sensitivity:3 discriminating:1 contract:1 physic:1 moody:1 expert:2 derivative:1 steered:1 actively:1 account:1 notable:1 metz:1 hpn:1 psajda:1 minimize:1 bright:1 square:2 kaufmann:2 characteristic:1 who:1 handwritten:1 produced:1 lu:1 computerized:1 cc:2 drive:2 processor:2 classified:4 tissue:1 detector:2 touretzky:3 ed:2 energy:8 james:1 associated:2 spie:1 workstation:1 propagated:3 sampled:2 massachusetts:1 improves:2 segmentation:1 clay:3 subtle:1 back:3 higher:3 supervised:1 unambiguously:1 april:1 leen:1 done:1 though:1 box:1 rejected:1 stage:3 just:1 hand:1 receives:2 horizontal:1 multiscale:1 propagation:1 aj:1 effect:1 contain:1 true:3 regularization:2 assigned:1 laboratory:1 whereby:1 criterion:3 tps:1 complete:1 image:21 charles:1 ifp:1 volume:2 significant:2 cambridge:1 cn5300:1 dot:1 had:2 funded:1 similarity:1 operating:4 something:1 tesauro:1 certain:1 n00014:1 binary:1 onr:1 morgan:2 greater:1 additional:1 steering:1 prune:1 signal:1 full:1 multiple:1 hpnn:45 characterized:1 clinical:4 cross:1 post:2 halving:1 neuro:1 basic:1 breast:3 vision:2 essentially:1 patient:1 represent:2 kernel:2 sometimes:1 pyramid:11 addition:1 fine:40 diagram:2 publisher:2 operate:1 eliminates:1 flow:1 chest:1 extracting:1 presence:2 split:1 enough:1 easy:1 architecture:21 interprets:1 reduce:1 prototype:1 shift:1 utility:1 passed:1 peter:1 passing:1 useful:2 detailed:1 aiding:1 ten:1 reduced:1 specifies:1 per:2 diagnosis:4 key:1 thresholded:1 kenneth:1 imaging:2 fraction:2 micron:3 uncertainty:1 reader:1 wu:1 missed:1 distinguish:1 display:1 fold:1 occur:1 performing:1 separable:1 relatively:2 jackknife:4 developing:1 clinically:1 combination:3 across:3 wi:1 lp:1 cun:3 invariant:1 ene:1 taken:1 equation:1 previously:1 hei:1 detectable:1 malignant:8 count:1 needed:1 drs:1 operation:2 gaussians:1 apply:4 denker:1 hierarchical:4 schmidt:2 original:1 remaining:1 include:1 madison:1 exploit:1 murray:2 added:1 usual:1 diagonal:2 gradient:1 thank:1 prospective:1 providing:1 difficult:1 unfortunately:1 robert:1 potentially:1 negative:7 design:1 calcium:1 vertical:1 convolution:2 november:1 extended:5 digitized:1 jim:1 discovered:1 lb:1 burt:1 introduced:1 inverting:1 david:4 specified:2 tentative:1 hanson:1 learned:1 boser:1 established:1 emu:1 able:1 pattern:1 rnax:1 including:1 ia:1 suitable:1 critical:1 natural:2 representing:1 improve:2 technology:1 started:1 understanding:1 multiplication:1 relative:1 deposit:1 loss:1 bick:1 rationale:1 validation:1 foundation:1 integrate:3 digital:1 editor:4 summary:1 repeat:1 last:1 institute:1 curve:1 evaluating:2 made:1 clue:1 san:2 transaction:1 lippmann:1 ignore:1 investigating:1 receiver:1 conclude:1 search:3 table:4 lip:1 learn:5 ca:2 complex:2 protocol:1 hierarchically:1 keng:1 border:2 noise:1 paul:1 roc:5 sub:1 position:11 wish:1 ib:1 mammogram:10 minute:1 showing:1 undergoing:1 r2:2 decay:1 sensing:1 gupta:1 dol:1 workshop:1 false:13 niques:1 occurring:1 suited:3 contained:1 truth:2 chance:1 extracted:2 ma:1 aided:4 included:1 typical:1 except:1 reducing:2 preset:1 tumor:1 called:1 evaluate:2 princeton:1 tested:1 |
555 | 1,507 | Lazy Learning Meets
the Recursive Least Squares Algorithm
Mauro Birattari, Gianluca Bontempi, and Hugues Bersini
Iridia - Universite Libre de Bruxelles
Bruxelles, Belgium
{mbiro, gbonte, bersini} @ulb.ac.be
Abstract
Lazy learning is a memory-based technique that, once a query is received, extracts a prediction interpolating locally the neighboring examples of the query which are considered relevant according to a distance
measure. In this paper we propose a data-driven method to select on a
query-by-query basis the optimal number of neighbors to be considered
for each prediction. As an efficient way to identify and validate local
models, the recursive least squares algorithm is introduced in the context of local approximation and lazy learning. Furthermore, beside the
winner-takes-all strategy for model selection, a local combination of the
most promising models is explored. The method proposed is tested on
six different datasets and compared with a state-of-the-art approach.
1
Introduction
Lazy learning (Aha, 1997) postpones all the computation until an explicit request for a
prediction is received. The request is fulfilled interpolating locally the examples considered relevant according to a distance measure. Each prediction requires therefore a local
modeling procedure that can be seen as composed of a structural and of a parametric identification . The parametric identification consists in the optimization of the parameters of
the local approximator. On the other hand, structural identification involves, among other
things, the selection of a family of local approximators, the selection of a metric to evaluate
which examples are more relevant, and the selection of the bandwidth which indicates the
size of the region in which the data are correctly modeled by members of the chosen family
of approximators. For a comprehensive tutorial on local learning and for further references
see Atkeson et al. (1997).
As far as the problem of bandwidth selection is concerned, different approaches exist. The
choice of the bandwidth may be performed either based on some a priori assumption or
on the data themselves. A further sub-classification of data-driven approaches is of interest
M. Birattari, G. Bontempi and H. Bersini
376
here. On the one hand, a constant bandwidth may be used; in this case it is set by a global
optimization that minimizes an error criterion over the available dataset. On the other hand,
the bandwidth may be selected locally and tailored for each query point.
In the present work, we propose a method that belongs to the latter class of local data-driven
approaches. Assuming a given fixed metric and local linear approximators, the method we
introduce selects the bandwidth on a query-by-query basis by means of a localleave-oneout cross- validation. The problem of bandwidth selection is reduced to the selection of the
number k of neighboring examples which are given a non-zero weight in the local modeling
procedure. Each time a prediction is required for a specific query point, a set of local
models is identified, each including a different number of neighbors. The generalization
ability of each model is then assessed through a local cross-validation procedure. Finally,
a prediction is obtained either combining or selecting the different models on the basis of
some statistic of their cross-validation errors.
The main reason to favor a query-by-query bandwidth selection is that it allows better
adaptation to the local characteristics of the problem at hand. Moreover, this approach is
able to handle directly the case in which the database is updated on-line (Bontempi et at.,
1997). On the other hand, a globally optimized bandwidth approach would, in principle,
require the global optimization to be repeated each time the distribution of the examples
changes.
The major contribution of the paper consists in the adoption of the recursive least squares
algorithm in the context of lazy learning. This is an appealing and efficient solution to the
intrinsically incremental problem of identifying and validating a sequence of local linear
models centered in the query point, each including a growing number of neighbors. It is
worth noticing here that a leave-one-out cross-validation of each model considered does
not involve any significant computational overload, since it is obtained though the PRESS
statistic (Myers, 1990) which simply uses partial results returned by the recursive least
squares algorithm. Schaal and Atkeson (1998) used already the recursive least squares
algorithm for the incremental update of a set of local models. In the present paper, we
use for the first time this algorithm in a query-by-query perspective as an effective way to
explore the neighborhood of each query point.
As a second contribution, we propose a comparison, on a local scale, between a competitive
and a cooperative approach to model selection. On the problem of extracting a final prediction from a set of alternatives, we compared a winner-takes-all strategy with a strategy
based on the combination of estimators (Wolpert, 1992).
In Section 5 an experimental analysis of the recursive algorithm for local identification
and validation is presented. The algorithm proposed, used in conjunction with different
strategies for model selection or combination, is compared experimentally with Cubist, the
rule-based tool developed by Ross Quinlan for generating piecewise-linear models.
2
Local Weighted Regression
Given two variables x E lR m and y E lR, let us consider the mapping
only through a set of n examples {(Xi, yd} ~=l obtained as follows:
f:
lR m
--t
lR, known
(1)
where Vi, Ci is a random variable such that E[ciJ = 0 and E[ciCjJ = 0, Vj =1= i, and
such that E[ciJ = I-lm(Xi), Vm ~ 2, where I-lmO is the unknown mth moment of the
distribution of Ci and is defined as a function of Xi. In particular for m = 2, the last of
the above mentioned properties implies that no assumption of global homoscedasticity is
made.
377
Lazy Learning Meets the Recursive Least Squares Algorithm
The problem of local regression can be stated as the problem of estimating the value that the
regression function f(x) = E[Ylx] assumes for a specific query point x, using information
pertaining only to a neighborhood of x.
Given a query point x q , and under the hypothesis of a local homoscedasticity of Ci, the
parameter (3 of a local linear approximation of f (.) in a neighborhood of Xq can be obtained
solving the local polynomial regression:
(2)
where, given a metric on the space Rm, d( Xi, Xq) is the distance from the query point to the
example, K (.) is a weight function, h is the bandwidth, and where a constant value 1
has been appended to each vector Xi in order to consider a constant term in the regression.
ith
In matrix notation, the solution of the above stated weighted least squares problem is given
by:
/3
= (X'W'WX)-lX'W'Wy = (Z'Z)-lZ'V = PZ'v,
(3)
where X is a matrix whose ith row is x~, y is a vector whose ith element is Yi, W is
a diagonal matrix whose ith diagonal element is Wii = JK (d(Xi,Xq)jh), Z = WX,
v = Wy, and the matrix X'W'WX = Z'Z is assumed to be non-singular so that its
inverse P = (Z'Z)-l is defined.
Once obtained the local linear polynomial approximation, a prediction of Yq = f(x q), is
finally given by:
Yq=X~/3 .
(4)
Moreover, exploiting the linearity of the local approximator, a leave-one-out crossvalidation estimation of the error variance E[ (Yq - Yq)2] can be obtained without any
significant overload. In fact, using the PRESS statistic (Myers, 1990), it is possible to
= Yj - xj /3 _ j' without explicitly identifying the parameters /3- j
calculate the error
from the examples available with the ph removed. The formulation of the PRESS statistic
for the case at hand is the following:
er
cv _
ej
-
,A
_ Yj - xjPZ'v _ Yj - xj/3
Yj - x j {3 _ j -
1
-
'P Zj
Zj
-
1
-
where zj is the ph row of Z and therefore Zj = WjjXj, and where h jj is the
e1ementoftheHatmatrixH = ZPZ' = Z(Z'Z) - lZ' .
3
(5)
h jj '
ph
diagonal
Recursive Local Regression
In what follows, for the sake of simplicity, we will focus on linear approximator. An
extension to generic polynomial approximators of any degree is straightforward. We will
assume also that a metric on the space R m is given. All the attention will be thus centered
on the problem of bandwidth selection.
If as a weight function K(-) the indicator function
K (d(Xi'X q)) =
h
{I
0
ifd(xi,xq)::; h,
otherwise;
(6)
is adopted, the optimization of the parameter h can be conveniently reduced to the optimization of the number k of neighbors to which a unitary weight is assigned in the local
M. Birattari, G. Bontempi and H. Bersini
378
regression evaluation. In other words, we reduce the problem of bandwidth selection to a
search in the space of h( k) = d( x( k), Xq), where x( k) is the kth nearest neighbor of the
query point.
The main advantage deriving from the adoption of the weight function defined in Eq. 6,
is that, simply by updating the parameter /3(k) of the model identified using the k nearest
neighbors, it is straightforward and inexpensive to obtain /3 (k + 1). In fact, performing a
step of the standard recursive least squares algorithm (Bierman, 1977), we have:
P(k
+ 1) =
P(k) _ P(k)x(k + l)x'(k + l)P(k)
1 + x'(k + l)P(k)x(k + 1)
,(k + 1) = P(k + l)x(k
e(k + 1) =
/3(k
+ 1) =
+ 1)
y(k + 1) - x' (k + l)/3(k)
/3(k) + ,(k + l)e(k + 1)
where P(k) = (Z'Z)-l when h = h(k), and where x(k
neighbor of the query point.
(7)
+ 1)
is the (k
+ l)th
nearest
Moreover, once the matrix P(k + 1) is available, the leave-one-out cross-validation errors
can be directly calculated without the need of any further model identification:
cv
_
Yj - xj/3(k + 1)
ej (k + 1) - 1 _ xjP(k + l)x/
(8)
It will be useful in the following to define for each value of k the [k x 1] vector e CV (k) that
contains all the leave-one-out errors associated to the model {3(k).
Once an initialization /3(0) = jj and P(O) = P is given, Eq. 7 and Eq. 8 recursively
evaluate for different values of k a local approximation of the regression function f(?),
a prediction of the value of the regression function in the query point, and the vector of
leave-one-out errors from which it is possible to extract an estimate of the variance of the
prediction error. Notice that jj is an a priqri estimate of the parameter and P is the covariance matrix that reflects the reliabi!ity of f3 (Bierman, 1977). For non-reliable initialization,
the following is usually adopted: P = >'1, with>. large and where I is the identity matrix.
4
Local Model Selection and Combination
The recursive algorithm described by Eq. 7 and Eq. 8 returns for a given query point x q ,
a set of predictions Yq (k) = x~/3(k), together with a set of associated leave-one-out error
vectors e Cv (k) .
From the information available, a final prediction f)q of the value of the regression function
can be obtained in different ways. Two main paradigms deserve to be considered: the first
is based on the selection of the best approximator according to a given criterion, while the
second returns a prediction as a combination of more local models.
If the selection paradigm, frequently called winner-takes-all, is adopted, the most natural
way to extract a final prediction Yq, consists in comparing the prediction obtained for each
value of k on the basis of the classical mean square error criterion:
with k = argmin MSE(k) = argmin
A
k
k
"k
L.J'
Wi
t=l
"k
(e?CV(k))2
t
.
L.Ji=l W t
.
'
(9)
379
Lazy Learning Meets the Recursive Least Squares Algorithm
Table 1: A summary of the characteristics of the data sets considered.
Dataset
Number of
examples
Number of
regressors
I Housing I Cpu I Prices I Mpg I Servo I Ozone I
506
209
159
392
167
330
l3
6
16
7
8
8
where Wi are weights than can be conveniently used to discount each error according to the
distance from the query point to the point to which the error corresponds (Atkeson et at.,
1997).
As an alternative to the winner-takes-all paradigm, we explored also the effectiveness of
local combinations of estimates (Wolpert, 1992). Adopting also in this case the mean
square error criterion, the final prediction of the value Yq is obtained as a weighted average
of the best b models, where b is a parameter of the algorithm. Suppose the predictions il q (k)
and the error vectors e Cv (k) have been ordered creating a sequence of integers {k i } so that
MSE( ki ) ::; MSE( kj ), Vi < j. The prediction of Yq is given by
~
Yq =
L~-l (iYq(kd
",b
r.
L..-i=l ,>z
'
(10)
where the weights are the inverse of the mean square errors: (i = l/MSE(ki ). This is an
example of the generalized ensemble method (Perrone & Cooper, 1993).
5 Experiments and Results
The experimental evaluation ofthe incremental local identification and validation algorithm
was performed on six datasets. The first five, described by Quinlan (1993), were obtained
from the VCI Repository of machine learning databases (Merz & Murphy, 1998), while the
last one was provided by Leo Breiman. A summary ofthe characteristics of each dataset is
presented in Table 1.
The methods compared adopt the recursive identification and validation algorithm, combined with different strategies for model selection or combination. We considered also two
approaches in which k is selected globally:
Ibl: Local bandwidth selection for linear local models. The number of neighbors is selected on a query-by-query basis and the prediction returned is the one of the best
model according to the mean square error criterion.
IbO: Local bandwidth selection for constant local models. The algorithm for constant
models is derived directly from the recursive method described in Eq. 7 and Eq. 8.
The best model is selected according to the mean square error criterion.
IbC: Local combination of estimators. This is an example, of the method described in
Eq. 10. On the datasets proposed, for each query the best 2 linear local models
and the best 2 constant models are combined.
gbl: Global bandwidth selection for linear local models. The value of k is obtained minimizing the prediction error in 20-fold cross-validation on the dataset available.
This value is then used for all the query points.
gbO: Global bandwidth selection for constant local models. As in gbl, the value of k is
optimized globally and kept constant for all the queries.
M. Birattarl. G. Bontempi and H. Bersini
380
Table 2: Mean absolute error on unseen cases.
Method I Housing I Cpu I Prices I Mpg I Servo I Ozone
Ibl
2.21
28.38
1509
1.94
0.48
3.52
2.60
0.32
3.33
1.97
IbO
31.54
1627
IbC
2.12
26.79
1488
1.83
0.29
3.31
2.30
28.69
1492
1.92
0.52
3.46
gbl
2.59
32.19
1639
1.99
3.19
gbO
0.34
Cubist
2.17
28.37
1331
1.90
0.36
3.15
Table 3: Relative error (%) on unseen cases.
I Method I Housing I
Ibl
IbO
IbC
gb1
gbO
12.63
18.06
12.35
13.47
17.99
Cpu
9.20
20.37
9.29
9.93
21.43
Cubist
16.02
12.71
I Prices I
15.87
22.19
17.62
15.95
22.29
Mpg
12.65
12.64
11.82
12.83
13.48
11.67
12.57
I Servo I Ozone
28.66
22.04
19.72
30.46
24.30
35.25
31.11
30.28
32.58
28.21
18.53
26.59
As far as the metric is concerned, we adopted a global Euclidean metric based on the
relative influence (relevance) ofthe regressors (Friedman, 1994). We are confident that the
adoption of a local metric could improve the performance of our lazy learning method.
The results of the methods introduced are compared with those we obtained, in the same
experimental settings, with Cubist, the rule-based tool developed by Quinlan for generating
piecewise-linear models. Each approach was tested on each dataset using the same 10-fold
cross-validation strategy. Each dataset was divided randomly into 10 groups of nearly
equal size. In turn, each of these groups was used as a testing set while the remaining
ones together were providing the examples. Thus all the methods performed a prediction
on the same unseen cases, using for each of them the same set of examples. In Table 2
we present the results obtained by all the methods, and averaged on the 10 cross-validation
groups. Since the methods were compared on the same examples in exactly the same
conditions, the sensitive one-tailed paired test of significance can be used. In what follows,
by "significantly better" we mean better at least at a 5% significance level.
The first consideration about the results concerns the local combination of estimators. According to Table 2, the method IbC performs in average always better than the winnertakes-all linear and constant. On two dataset IbC is significantly better than both Ibl and
IbO; and on three dataset it is significantly better than one of the two, and better in average
than the other.
The second consideration is about the comparison between our query-by-query bandwidth
selection and a global optimization of the number of neighbors: in average Ibl and IbO
performs better than their counterparts gbl and gbO. On two datasets Ibl is significantly
better than gbl, while is about the same on the other four. On one dataset IbO is significantly
better than gbO.
As far as the comparison with Cubist is concerned, the recursive lazy identification and
validation proposed obtains results comparable with those obtained by the state-of-the-art
method implemented in Cubist. On the six datasets, IbC performs one time significantly
better than Cubist, and one time significantly worse.
Lazy Learning Meets the Recursive Least Squares Algorithm
381
The second index of performance we investigated is the relative error, defined as the mean
square error on unseen cases, normalized by the variance of the test set. The relative errors
are presented in Table 3 and show a similar picture to Table 2, although the mean square
errors considered here penalize larger absolute errors.
6
Conclusion and Future Work
The experimental results confirm that the recursive least squares algorithm can be effectively used in a local context. Despite the trivial metric adopted, the local combination
of estimators, identified and validated recursively, showed to be able to compete with a
state-of-the-art approach.
Future work will focus on the problem of local metric selection. Moreover, we will explore more sophisticated ways to combine local estimators and we will extend this work to
polynomial approximators of higher degree.
Acknowledgments
The work of Mauro Birattari was supported by the FIRST program of the Region Wallonne,
Belgium. The work of Gianluca Bontempi was supported by the European Union TMR
Grant FMBICT960692. The authors thank Ross Quinlan and gratefully acknowledge using
his software Cubist. For more details on Cubist see http://www.rulequest.com.
We also thank Leo Breiman for the dataset ozone and the UCI Repository for the other
datasets used in this paper.
References
Aha D. W. 1997. Editorial. Artificial Intelligence Review, 11(1-5), 1-6. Special Issue on
Lazy Learning.
Atkeson C. G. , Moore A. W. & Schaal S. 1997. Locally weighted learning. Artificial
Intelligence Review, 11(1-5), 11-73.
Bierman G. 1. 1977. Factorization Methodsfor Discrete Sequential Estimation. New York,
NY: Academic Press.
Bontempi G., Birattari M. & Bersini H. 1997. Lazy learning for local modeling and
control design. International Journal of Control. Accepted for publication.
Friedman 1. H. 1994. Flexible metric nearest neighbor classification. Tech. rept. Department of Statistics, Stanford University.
Merz C. J. & Murphy P. M . 1998. UCI Repository of machine learning databases.
Myers R. H. 1990. Classical and Modern Regression with Applications. Boston, MA:
PWS-KENT.
Perrone M. P. & Cooper L. N. 1993. When networks disagree: Ensemble methods for
hybrid neural networks. Pages 126-142 of Mammone R. J. (ed), Artificial Neural
Networks for Speech and Vision. Chapman and Hall.
Quinlan 1. R. 1993. Combining instance-based and model-based learning. Pages 236-243
of Machine Learning. Proceedings of the Tenth International Conference. Morgan
Kaufmann.
Schaal S. & Atkeson C. G. 1998. Constructive incremental learning from only local
information. Neural Computation, 10(8), 2047-2084.
Wolpert D. 1992. Stacked Generalization. Neural Networks, 5, 241-259.
| 1507 |@word repository:3 polynomial:4 covariance:1 kent:1 recursively:2 moment:1 contains:1 selecting:1 comparing:1 com:1 wx:3 update:1 intelligence:2 selected:4 cubist:9 ith:4 lr:4 lx:1 five:1 consists:3 combine:1 introduce:1 homoscedasticity:2 themselves:1 frequently:1 growing:1 mpg:3 globally:3 cpu:3 hugues:1 provided:1 estimating:1 moreover:4 notation:1 linearity:1 what:2 argmin:2 minimizes:1 developed:2 exactly:1 rm:1 control:2 grant:1 local:47 rept:1 despite:1 meet:4 yd:1 initialization:2 factorization:1 adoption:3 averaged:1 acknowledgment:1 yj:5 testing:1 recursive:16 union:1 procedure:3 significantly:7 word:1 selection:22 context:3 influence:1 www:1 straightforward:2 attention:1 simplicity:1 identifying:2 ozone:4 estimator:5 rule:2 deriving:1 his:1 ity:1 handle:1 updated:1 suppose:1 us:1 hypothesis:1 element:2 jk:1 updating:1 database:3 cooperative:1 calculate:1 region:2 removed:1 servo:3 mentioned:1 solving:1 basis:5 leo:2 stacked:1 effective:1 pertaining:1 query:29 artificial:3 neighborhood:3 mammone:1 whose:3 larger:1 stanford:1 otherwise:1 ability:1 statistic:5 favor:1 unseen:4 final:4 housing:3 sequence:2 myers:3 advantage:1 propose:3 adaptation:1 neighboring:2 relevant:3 combining:2 uci:2 validate:1 crossvalidation:1 exploiting:1 generating:2 incremental:4 leave:6 ac:1 nearest:4 received:2 eq:8 implemented:1 involves:1 implies:1 centered:2 require:1 generalization:2 extension:1 considered:8 hall:1 mapping:1 lm:1 major:1 adopt:1 belgium:2 estimation:2 ross:2 sensitive:1 tool:2 weighted:4 reflects:1 always:1 ej:2 breiman:2 publication:1 conjunction:1 derived:1 focus:2 validated:1 schaal:3 indicates:1 tech:1 ibl:6 mth:1 bierman:3 selects:1 issue:1 among:1 classification:2 flexible:1 priori:1 art:3 special:1 equal:1 once:4 f3:1 chapman:1 nearly:1 ibc:6 future:2 piecewise:2 modern:1 randomly:1 composed:1 comprehensive:1 murphy:2 friedman:2 interest:1 evaluation:2 bontempi:7 partial:1 aha:2 euclidean:1 rulequest:1 instance:1 modeling:3 combined:2 confident:1 international:2 vm:1 together:2 tmr:1 worse:1 creating:1 return:2 de:1 explicitly:1 vi:2 performed:3 competitive:1 contribution:2 appended:1 square:18 il:1 variance:3 characteristic:3 kaufmann:1 ensemble:2 identify:1 ofthe:3 identification:8 worth:1 ed:1 inexpensive:1 universite:1 associated:2 dataset:10 intrinsically:1 sophisticated:1 higher:1 formulation:1 though:1 furthermore:1 until:1 hand:6 vci:1 normalized:1 counterpart:1 assigned:1 moore:1 criterion:6 generalized:1 performs:3 consideration:2 ji:1 winner:4 extend:1 significant:2 cv:6 gratefully:1 l3:1 showed:1 perspective:1 belongs:1 driven:3 approximators:5 yi:1 seen:1 morgan:1 paradigm:3 gb1:1 academic:1 cross:8 divided:1 paired:1 prediction:21 regression:11 vision:1 metric:10 editorial:1 tailored:1 adopting:1 penalize:1 singular:1 gbl:5 validating:1 thing:1 member:1 effectiveness:1 birattari:5 integer:1 extracting:1 structural:2 unitary:1 concerned:3 xj:3 winnertakes:1 bandwidth:17 identified:3 reduce:1 six:3 returned:2 speech:1 york:1 jj:4 useful:1 involve:1 ylx:1 discount:1 locally:4 ph:3 reduced:2 http:1 exist:1 zj:4 tutorial:1 notice:1 fulfilled:1 correctly:1 discrete:1 group:3 four:1 tenth:1 kept:1 compete:1 inverse:2 noticing:1 family:2 comparable:1 ki:2 fold:2 software:1 sake:1 performing:1 department:1 according:7 combination:10 request:2 perrone:2 kd:1 wi:2 appealing:1 turn:1 adopted:5 available:5 wii:1 generic:1 alternative:2 assumes:1 remaining:1 quinlan:5 bersini:6 classical:2 already:1 strategy:6 ulb:1 parametric:2 diagonal:3 bruxelles:2 kth:1 distance:4 thank:2 mauro:2 trivial:1 reason:1 assuming:1 modeled:1 index:1 providing:1 minimizing:1 cij:2 oneout:1 stated:2 design:1 unknown:1 disagree:1 datasets:6 acknowledge:1 introduced:2 required:1 optimized:2 deserve:1 able:2 wy:2 usually:1 xjp:1 program:1 including:2 memory:1 reliable:1 natural:1 hybrid:1 indicator:1 improve:1 yq:9 picture:1 extract:3 kj:1 xq:5 review:2 relative:4 beside:1 approximator:4 validation:12 degree:2 principle:1 row:2 gianluca:2 summary:2 supported:2 last:2 jh:1 neighbor:10 absolute:2 calculated:1 author:1 made:1 regressors:2 atkeson:5 far:3 lz:2 obtains:1 confirm:1 global:7 assumed:1 xi:8 search:1 tailed:1 table:8 promising:1 mse:4 investigated:1 interpolating:2 european:1 vj:1 significance:2 main:3 libre:1 repeated:1 cooper:2 ny:1 sub:1 lmo:1 explicit:1 pws:1 specific:2 methodsfor:1 er:1 explored:2 pz:1 concern:1 sequential:1 effectively:1 ci:3 boston:1 wolpert:3 simply:2 explore:2 conveniently:2 lazy:12 ordered:1 corresponds:1 ma:1 identity:1 price:3 change:1 experimentally:1 called:1 accepted:1 experimental:4 merz:2 select:1 latter:1 assessed:1 relevance:1 overload:2 constructive:1 evaluate:2 tested:2 |
556 | 1,508 | Markov processes on curves for
automatic speech recognition
Lawrence Saul and Mazin Rahim
AT&T Labs - Research
Shannon Laboratory
180 Park Ave E-171
Florham Park, NJ 07932
{lsaul,rnazin}Gresearch.att.com
Abstract
We investigate a probabilistic framework for automatic speech
recognition based on the intrinsic geometric properties of curves.
In particular, we analyze the setting in which two variables-one
continuous (~), one discrete (s )-evolve jointly in time. We suppose that the vector ~ traces out a smooth multidimensional curve
and that the variable s evolves stochastically as a function of the
arc length traversed along this curve. Since arc length does not
depend on the rate at which a curve is traversed, this gives rise
to a family of Markov processes whose predictions, Pr[sl~]' are
invariant to nonlinear warpings of time. We describe the use of
such models, known as Markov processes on curves (MPCs), for
automatic speech recognition, where ~ are acoustic feature trajectories and s are phonetic transcriptions. On two tasks-recognizing
New Jersey town names and connected alpha-digits- we find that
MPCs yield lower word error rates than comparably trained hidden
Markov models.
1
Introduction
Variations in speaking rate currently present a serious challenge for automatic
speech recognition (ASR) (Siegler & Stern, 1995). It is widely observed , for example,
that fast speech is more prone to recognition errors than slow speech. A related effect, occurring at the phoneme level, is that consonants (l,re more frequently botched
than vowels. Generally speaking, consonants have short-lived, non-stationary acoustic signatures; vowels, just the opposite. Thus, at the phoneme level, we can view
the increased confusability of consonants as a consequence of locally fast speech.
L. Saul and M. Rahim
752
set) =s 1
START
t=O
x(t)
END
t='t
Figure 1: Two variables-one continuous (x), one discrete (s )- evol ve jointly in
time. The trace of s partitions the curve of x into different segments whose boundaries occur where s changes value.
In this paper , we investigate a probabilistic framework for ASR that models variations in speaking rate as arising from nonlinear warpings of time (Tishby, 1990) .
Our framework is based on the observation that acoustic feature vectors trace out
continuous trajectories (Ostendorf et aI , 1996). We view these trajectories as multidimensional curves whose intrinsic geometric properties (such as arc length or
radius) do not depend on the rate at which they are traversed (do Carmo, 1976).
We describe a probabilistic model whose predictions are based on these intrinsic
geometric properties and-as such-are invariant to nonlinear warpings of time.
The handling of this invariance distinguishes our methods from traditional hidden
Markov models (HMMs) (Rabiner & Juang, 1993).
The probabilistic models studied in this paper are known as Markov processes on
curves (MPCs). The theoretical framework for MPCs was introduced in an earlier
paper (Saul , 1997), which also discussed the problems of decoding and parameter
estimation. In the present work, we report the first experimental results for MPCs
on two difficult benchmark problems in ASR. On these problems- recognizing New
Jersey town names and connected alpha-digits- our results show that MPCs generally match or exceed the performance of comparably trained HMMs.
The organization of this paper is as follows . In section 2, we review the basic
elements of MPCs and discuss important differences between MPCs and HMMs. In
section 3, we present our experimental results and evaluate their significance.
2
Markov processes on curves
Speech recognizers take a continuous acoustic signal as input and return a sequence
of discrete labels representing phonemes, syllables, or words as output. Typically
the short-time properties of the speech signal are summarized by acoustic feature
vectors. Thus the abstract mathematical problem is to describe a multidimensional
trajectory {x(t) It E [0, T]} by a sequence of discrete labels S1 S2 . . . Sn. As shown in
figure 1, this is done by specifying consecutive time intervals such that s(t)
Sk
for t E [tk-1, tk] and attaching the labels Sk to contiguous arcs along the trajectory.
To formulate a probabilistic model of this process, we consider two variables-one
continuous (x), one discrete (s )-that evolve jointly in time. Thus the vector x
traces out a smooth multidimensional curve, to each point of which the variable s
attaches a discrete label.
=
Markov processes on curves are based on the concept of arc length. After reviewing
how to compute arc lengths along curves, we introduce a family of Markov processes
whose predictions are invariant to nonlinear warpings of time. We then consider
the ways in which these processes (and various generalizations) differ from HMMs.
Markov Processes on Curves for Automatic Speech Recognition
2.1
753
Arc length
Let g(~) define a D x D matrix-valued function over x E RP. If g(~) is everywhere
non-negative definite, then we can use it as a metric to compute distances along
curves. In particular, consider two nearby points separated by the infinitesimal
vector d~. We define the squared distance between these two points as:
(1)
Arc length along a curve is the non-decreasing function computed by integrating
these local distances. Thus, for the trajectory x(t), the arc length between the
points x(t!) and X(t2) is given by:
f=
l
t2
[~Tg(x)i:]~,
dt
(2)
tl
it
where i: =
[~(t)] denotes the time derivative of~ . Note that the arc length defined
by eq. (2) is invariant under reparameterizations of the trajectory, ~(t) -t ~(J(t)) ,
where f(t) is any smooth monotonic function of time that maps the interval [tl, t2]
into itself.
In the special case where g(~) is the identity matrix, eq. (2) reduces to the standard
definition of arc length in Euclidean space. More generally, however, eq. (1) defines
a non-Euclidean metric for computing arc lengths. Thus, for example, if the metric
g(x) varies as a function of~, then eq. (2) can assign different arc lengths to the
trajectories x(t) and x(t) + ~o, where ~o is a constant displacement.
2.2
States and lifelengths
We now return to the problem of segmentation, as illustrated in figure 1. We refer
to the possible values of s as states. MPCs are conditional random processes that
evolve the state variable s stochastically as a function of the arc length traversed
along the curve of~. In MPCs, the probability of remaining in a particular state
decays exponentially with the cumulative arc length traversed in that state. The
signature of a state is the particular way in which it computes arc length.
To formalize this idea, we associate with each state i the following quantities: (i)
a feature-dependent matrix gi (x) that can be used to compute arc lengths, as in
eq. (2); (ii) a decay parameter Ai that measures the probability per unit arc length
that s makes a transition from state i to some other state; and (iii) a set of transition
probabilities aij, where aij represents the probability that-having decayed out of
state i-the variable s makes a transition to state j . Thus, aij defines a stochastic
transition matrix with zero elements along the diagonal and rows that sum to one:
aii
0 and 2: j aij
1. A Markov process is defined by the set of differential
equations:
=
=
d
Pi
dt
=
\
-/liPi
).]:1 + ~
\
[.
L.J /ljpjaji
[X. T gi ( X X
1
1
T ( ) ? ] :I
~ 9j x ~
,
(3)
#i
where Pi(t) denotes the (forward) probability that s is in state i at time t, based
on its history up to that point in time. The right hand side of eq. (3) consists of
two competing terms. The first term computes the probability that s decays out
of state i; the second computes the probability that s decays into state i. Both
terms are proportional to measures of arc length, making the evolution of Pi along
the curve of x invariant to nonlinear warpings of time. The decay parameter, Ai,
controls the typical amount of arc length traversed in state i ; it may be viewed as
L. Saul and M. Rahim
754
an inverse lifetime or-to be more precise-an inverse lifelength. The entire process
is Markovian because the evolution of Pi depends only on quantities available at
time t.
2.3
Decoding
Given a trajectory x(t), the Markov process in eq. (3) gives rise to a conditional
probability distribution over possible segmentations, s(t). Consider the segmentation in which s(t) takes the value Sk between times tk-l and tk, and let
fSk = jtk dt [XTgsk(X) X ]%
(4)
tk-l
denote the arc length traversed in state Sk. By integrating eq. (3), one can show that
the probability of remaining in state Sk decays exponentially with the arc length f Sk '
Thus, the conditional probability of the overall segmentation is given by:
Pr[s,flx]
n
n
k=l
k=O
= II ASke->'Sklsk II aSkSk+ll
I
/
(5)
where we have used So and Sn+1 to denote the START and END states of the Markov
process. The first product in eq. (5) multiplies the probabilities that each segment
traverses exactly its observed arc length. The second product multiplies the probabilities for transitions between states Sk and Sk+l' The leading factors of ASk are
included to normalize each state's distribution over observed arc lengths.
There are many important quantities that can be computed from the distribution,
Pr[ S Ix]. Of particular interest for ASR is the most probable segmentation: s* (x) =
argmaxs,l {In Pr[s, fix]}. As described elsewhere (Saul, 1997), this maximization
can be performed by discretizing the time axis and applying a dynamic programming
procedure. The resulting algorithm is similar to the Viterbi procedure for maximum
likelihood decoding (Rabiner & Juang, 1993).
2.4
Parameter estimation
The parameters {Ai, aij, gi (x)} in MPCs are estimated from training data to maximize the log-likelihood of target segmentations. In our preliminary experiments
with MPCs, we estimated only the metric parameters, gi(X); the others were assigned the default values Ai = 1 and aij = 1/ Ii, where Ii is the fanout of state i.
The metrics gi (x) were assumed to have the parameterized form:
(6)
where (ji is a positive definite matrix with unit determinant, and cI>i (x) is a nonnegative scalar-valued function of x. For the experiments in this paper, the form of
cI>i(X) was fixed so that the MPCs reduced to HMMs as a special case, as described
in the next section. Thus the only learning problem was to estimate the matrix
parameters (ji. This was done using the reestimation formula:
(ji
~
C
J
~xT
dt. T
? 1 cI>i(x(t)),
[x (ji-1X]"2
(7)
where the integral is over all speech segments belonging to state i, and the constant
C is chosen to enforce the determinant constraint l(ji I = 1. For fixed cI>i (x), we
have shown previously (Saul, 1997) that this iterative update leads to monotonic
increases in the log-likelihood.
Markov Processes on Curves for Automatic Speech Recognition
2.5
755
Relation to HMMs and previous work
There are several important differences between HMMs and MPCs. HMMs parameterize joint distributions of the form: Pr[s, z] = Dt Pr[st+1lsd Pr[zt Isd. Thus,
in HMMs, parameter estimation is directed at learning a synthesis model, Pr[zls]'
while in MPCs, it is directed at learning a segmentation model, Pr[s,flz]. The
direction of conditioning on z is a crucial difference. MPCs do not attempt to learn
anything as ambitious as a joint distribution over acoustic feature trajectories. \
HMMs and MPCs also differ in how they weight the speech signal. In HMMs, each
state contributes an amount to the overall log-likelihood that grows in proportion
to its duration in time. In MPCs, on the other hand, each state contributes an
amount that grows in proportion to its arc length. Naturally, the weighting by arc
length attaches a more important role to short-lived but non-stationary phonemes,
such as consonants. It also guarantees the invariance to nonlinear warpings of time
(to which the predictions of HMMs are quite sensitive).
In terms of previous work,\mr motivation for MPCs resembles that of Tishby (1990),
who several years ago proposed a dynamical systems approach to speech processing.
Because MPCs exploit the continuity of acoustic feature trajectories, they also bear
some resemblance to so-called segmental HMMs (Ostendorf et aI, 1996). MPCs
nevertheless differ from segmental HMMs in two important respects: the invariance
to nonlinear warpings of time , and the emphasis on learning a segmentation model
Pr[s , flz], as opposed to a synthesis model, Pr[xls].
Finally, we note that admitting a slight generalization in the concept of arc length,
we can essentially realize HMMs as a special case of MPCs. This is done by computing arc lengths along the spacetime trajectories z(t) = {x(t),t}-that is to say,
replacing eq. (1) by dL 2 = [zTg(z) z]dt 2 , where z = {:il, 1} and g(z) is a spacetime
metric. This relaxes the invariance to nonlinear warpings of time and incorporates
both movement in acoustic feature space and duration in time as measures of phonemic evolution. Moreover, in this setting, one can mimic the predictions of HMMs
by setting the (J'i matrices to have only one non-zero element (namely, the diagonal
element for delta-time contributions to the arc length) and by defining the functions
<l>i(X) in terms of HMM emission probabilities P(xli) as:
P(zli) ]
<l>i(X) = -In [ 2::k P(xlk) .
(8)
This relation is important because it allows us to initialize the parameters of an
MPC by those of a continuous-density HMM,. This initialization was used in all the
experiments reported below.
3
Automatic speech recognition
Both HMMs and MPCs were used to, build connected speech recognizers. Training
and test data came from speaker-independent databases of telephone speech. All
data was digitized at the caller's local switch and transmitted in this form to the
receiver. For feature extraction, input telephone signals (sampled at 8 kHz and
band-limited between 100-3800 Hz) were pre-emphasized and blocked into 30ms
frames with a frame shift of 10ms. Each frame was Hamming windowed , autocorrelated, and processed by LPC cepstral analysis to produce a vector of 12 liftered
cepstral coefficients (Rabiner & Juang, 1993). The feature vector was then augmented by its normalized log energy value, as well as temporal derivatives of first
and second order. Overall, each frame of speech was described by 39 features . These
features were used diffe:.;ently by HMMs and MPCs, as described below.
L. Saul and M. Rahim
756
NJ town names
22 ~
,
Mixtures
2
4
8
16
32
64
HMM (%)
22.3
18.9
16.5
14.6
13.5
11.7
MPC ('fo)
20.9
17.5
15.1
13.3
12.3
11.4
14
- 0-
12
o
1000
2000
3000
4000
5000
parameters pe r state
Table 1: Word error rates for HMMs (dashed) and MPCs (solid) on the task of
recognizing NJ town names. The table shows the error rates versus the number of
mixture components; the graph , versus the number of parameters per hidden state.
Recognizers were evaluated on two tasks. The first task was recognizing New Jersey town names (e.g., Newark) . The training data for this task (Sachs et aI , 1994)
consisted of 12100 short phrases, spoken in the seven major dialects of American
English . These phrases, ranging from two to four words in length, were selected to
provide maximum phonetic coverage. The test data consisted of 2426 isolated utterances of 1219 New Jersey town names and was collected from nearly 100 speakers.
Note that the training and test data for this task have non-overlapping vocabularies .
Baseline recognizers were built using 43Ieft-to-right continuous-density HMMs, each
corresponding to a context-independent English phone. Phones were modeled by
three-state HMMs , with the exception of background noise , which was modeled by
a single state. State emission probabilities were computed by mixtures of Gaussians
with diagonal covariance matrices. Different sized models were trained using M
2,
4, 8, 16, 32, and 64 mixture components per hidden state; for a particular model ,
the number of mixture components was the same across all states. Parameter
estimation was handled by a Viterbi implementation of the Baum-Welch algorithm.
=
MPC recognizers were built using the same overall grammar. Each hidden state
in the MPCs was assigned a metric gi(~) = O';l<I>l(~). The functions <I>i(~) were
initialized (and fixed) by the state emission probabilities of the HMMs , as given
by eq. (8). The matrices O'i were estimated by iterating eq. (7). We computed arc
lengths along the 14 dimensional spacetime trajectories through cepstra, log-energy,
and time . Thus each O'i was a 14 x 14 symmetric matrix applied to tangent vectors
consisting of delta-cepstra, delta-log-energy, and delta-time.
The table in figure 1 shows the results of these experiments comparing MPCs to
HMMs. For various model sizes (as measured by the number of mixture components), we found the MPCs to yield consistently lower error rates than the HMMs.
The graph in figure 1 plots these word error rates versus the number of modeling parameters per hidden state. This graph shows that the MPCs are not outperforming
the HMMs merely because they have extra modeling parameters (i .e. , the O'i matrices). The beam widths for the decoding procedures in these experiments were
chosen so that corresponding recognizers activated roughly equal numbers of arcs.
The second task in our experiments involved the recognition of connected alphadigits (e.g ., N Z 3 V J 4 E 3 U 2). The training and test data consisted of
757
Markov Processes on Curves for Automatic Speech Recognition
13
12
Mixtures
2
4
8
HMM (%)
12.5
10.7
10.0
MPC (%)
10.0
8.8
8.2
..... ,
~11
~
g10
'0
CD
9
~oo
400
600
800
1000
parameters per state
1200
1400
Figure 2: Word error rates for HMMs and MPCs on the task of recognizing connected alpha-digits. The table shows the error rates versus the number of mixture
components; the graph , versus the number of parameters per hidden state.
14622 and 7255 utterances, respectively. Recognizers were built from 285 sub-word
HMMs/MPCs, each corresponding to a context-dependent English phone. The recognizers were trained and evaluated in the same way as the previous task. Results
are shown in figure 2.
While these results demonstrate the viability of MPCs for automatic speech recognition, several issues require further attention . The most important issues are feature selection-how to define meaningful acoustic trajectories from the raw speech
signal-and learning- how to parameterize and estimate the hidden state metrics
gi (~) from sampled trajectories {z (t)}. These issues and others will be studied in
future work.
References
M. P. do Carmo (1976) . Differential Geometry of Curves and Surfaces. Prentice
Hall.
M. Ostendorf, V. Digalakis, and O. Kimball (1996). From HMMs to segment models: a unified view of stochastic modeling for speech recognition. IEEE Transactions
on Acoustics, Speech and Signal Processing, 4:360-378.
L. Rabiner and B . Juang (1993) . Fundamentals of Speech Recognition. Prentice
Hall, Englewood Cliffs, NJ.
R. Sachs, M. Tikijian, and E. Roskos (1994). United States English subword speech
data. AT&T unpublished report.
L. Saul (1998) . Automatic segmentation of continuous trajectories with invariance
to nonlinear warpings of time . In Proceedings of the Fifteenth International Conference on Machine Learning, 506- 514.
M. A. Siegler and R . M. Stern (1995). On the effects of speech rate in large vocabulary speech recognition systems. In Proceedings of th e 1995 IEEE International
Conference on Acoustics, Speech, and Signal Processing, 612-615.
N. Tishby (1990). A dynamical system approach to speech processing. In Proceedings of the 1990 IEEE International Conference on Acoustics, Speech, and Signal
Processing, 365-368 .
PART VII
VISUAL PROCESSING
| 1508 |@word determinant:2 proportion:2 aske:1 covariance:1 solid:1 att:1 united:1 subword:1 com:1 comparing:1 realize:1 partition:1 plot:1 update:1 stationary:2 selected:1 short:4 traverse:1 mathematical:1 along:10 windowed:1 differential:2 consists:1 introduce:1 roughly:1 frequently:1 decreasing:1 moreover:1 spoken:1 unified:1 nj:4 guarantee:1 temporal:1 multidimensional:4 rahim:4 exactly:1 control:1 unit:2 positive:1 local:2 consequence:1 lsd:1 cliff:1 emphasis:1 initialization:1 studied:2 resembles:1 specifying:1 hmms:28 limited:1 directed:2 definite:2 digit:3 procedure:3 displacement:1 word:7 integrating:2 pre:1 selection:1 prentice:2 context:2 applying:1 map:1 baum:1 attention:1 duration:2 formulate:1 welch:1 evol:1 variation:2 caller:1 target:1 suppose:1 programming:1 associate:1 element:4 recognition:14 database:1 observed:3 role:1 parameterize:2 connected:5 movement:1 jtk:1 reparameterizations:1 flz:2 dynamic:1 signature:2 trained:4 depend:2 reviewing:1 segment:4 aii:1 joint:2 jersey:4 various:2 autocorrelated:1 dialect:1 separated:1 fast:2 describe:3 whose:5 quite:1 widely:1 valued:2 say:1 grammar:1 florham:1 gi:7 jointly:3 itself:1 sequence:2 product:2 argmaxs:1 normalize:1 juang:4 produce:1 tk:5 oo:1 measured:1 phonemic:1 eq:12 coverage:1 differ:3 direction:1 radius:1 stochastic:2 require:1 assign:1 fix:1 generalization:2 preliminary:1 probable:1 traversed:7 hall:2 lawrence:1 viterbi:2 major:1 consecutive:1 estimation:4 label:4 currently:1 sensitive:1 kimball:1 emission:3 consistently:1 likelihood:4 ave:1 baseline:1 dependent:2 typically:1 entire:1 lsaul:1 hidden:8 relation:2 overall:4 issue:3 multiplies:2 special:3 initialize:1 equal:1 asr:4 having:1 extraction:1 represents:1 park:2 nearly:1 mimic:1 future:1 t2:3 report:2 others:2 serious:1 distinguishes:1 ve:1 geometry:1 consisting:1 vowel:2 attempt:1 organization:1 interest:1 englewood:1 investigate:2 mixture:8 admitting:1 activated:1 integral:1 euclidean:2 initialized:1 re:1 isolated:1 theoretical:1 increased:1 earlier:1 modeling:3 markovian:1 contiguous:1 maximization:1 tg:1 phrase:2 recognizing:5 tishby:3 reported:1 varies:1 st:1 decayed:1 density:2 fundamental:1 international:3 probabilistic:5 decoding:4 synthesis:2 squared:1 town:6 opposed:1 stochastically:2 american:1 derivative:2 leading:1 return:2 summarized:1 coefficient:1 depends:1 performed:1 view:3 lab:1 analyze:1 start:2 contribution:1 il:1 phoneme:4 who:1 yield:2 rabiner:4 xli:1 raw:1 comparably:2 trajectory:16 ago:1 history:1 fo:1 definition:1 infinitesimal:1 energy:3 involved:1 mpc:4 naturally:1 hamming:1 sampled:2 ask:1 diffe:1 segmentation:9 formalize:1 dt:6 done:3 evaluated:2 lifetime:1 just:1 hand:2 ostendorf:3 replacing:1 nonlinear:9 overlapping:1 continuity:1 defines:2 resemblance:1 grows:2 name:6 effect:2 concept:2 normalized:1 consisted:3 evolution:3 assigned:2 symmetric:1 laboratory:1 illustrated:1 ll:1 width:1 speaker:2 anything:1 m:2 demonstrate:1 ranging:1 ji:5 conditioning:1 exponentially:2 khz:1 discussed:1 slight:1 refer:1 blocked:1 ai:7 automatic:10 recognizers:8 surface:1 segmental:2 phone:3 phonetic:2 carmo:2 discretizing:1 came:1 outperforming:1 transmitted:1 mr:1 maximize:1 signal:8 ii:5 dashed:1 reduces:1 smooth:3 match:1 prediction:5 basic:1 essentially:1 metric:8 fifteenth:1 beam:1 background:1 interval:2 crucial:1 extra:1 hz:1 incorporates:1 exceed:1 iii:1 viability:1 relaxes:1 switch:1 competing:1 opposite:1 idea:1 shift:1 handled:1 fanout:1 speech:30 speaking:3 ieft:1 generally:3 iterating:1 fsk:1 amount:3 locally:1 band:1 processed:1 reduced:1 mazin:1 sl:1 estimated:3 arising:1 per:6 delta:4 discrete:6 four:1 nevertheless:1 isd:1 graph:4 merely:1 sum:1 year:1 inverse:2 everywhere:1 parameterized:1 zli:1 flx:1 family:2 syllable:1 spacetime:3 nonnegative:1 occur:1 constraint:1 cepstra:2 nearby:1 zls:1 belonging:1 across:1 evolves:1 making:1 s1:1 invariant:5 pr:11 equation:1 previously:1 discus:1 mpcs:32 end:2 available:1 gaussians:1 enforce:1 rp:1 denotes:2 remaining:2 exploit:1 build:1 warping:9 quantity:3 traditional:1 diagonal:3 distance:3 hmm:4 seven:1 collected:1 length:30 modeled:2 difficult:1 trace:4 negative:1 rise:2 implementation:1 stern:2 lived:2 zt:1 ambitious:1 observation:1 markov:15 arc:31 benchmark:1 defining:1 precise:1 digitized:1 frame:4 introduced:1 namely:1 unpublished:1 acoustic:12 dynamical:2 below:2 lpc:1 challenge:1 built:3 confusability:1 digalakis:1 representing:1 axis:1 utterance:2 sn:2 review:1 geometric:3 tangent:1 evolve:3 bear:1 attache:2 proportional:1 versus:5 pi:4 cd:1 row:1 prone:1 elsewhere:1 english:4 aij:6 side:1 saul:8 cepstral:2 attaching:1 curve:21 boundary:1 default:1 transition:5 cumulative:1 vocabulary:2 computes:3 forward:1 transaction:1 alpha:3 transcription:1 reestimation:1 receiver:1 assumed:1 consonant:4 continuous:8 iterative:1 sk:8 table:4 learn:1 contributes:2 significance:1 sachs:2 s2:1 motivation:1 noise:1 augmented:1 tl:2 slow:1 sub:1 xl:1 pe:1 weighting:1 ix:1 formula:1 xt:1 emphasized:1 decay:6 dl:1 intrinsic:3 g10:1 ci:4 occurring:1 siegler:2 vii:1 visual:1 scalar:1 monotonic:2 conditional:3 identity:1 viewed:1 sized:1 xlk:1 change:1 included:1 typical:1 telephone:2 called:1 invariance:5 experimental:2 shannon:1 meaningful:1 exception:1 newark:1 evaluate:1 handling:1 |
557 | 1,509 | Experimental Results on Learning Stochastic
Memoryless Policies for Partially Observable
Markov Decision Processes
John K. Williams
Department of Mathematics
University of Colorado
Boulder, CO 80309-0395
jkwillia@euclid.colorado.edu
Satinder Singh
AT &T Labs-Research
180 Park Avenue
Florham Park, NJ 07932
baveja@research.att.com
Abstract
Partially Observable Markov Decision Processes (pO "MOPs) constitute
an important class of reinforcement learning problems which present
unique theoretical and computational difficulties. In the absence of the
Markov property, popular reinforcement learning algorithms such as
Q-Iearning may no longer be effective, and memory-based methods
which remove partial observability via state-estimation are notoriously
expensive. An alternative approach is to seek a stochastic memoryless
policy which for each observation of the environment prescribes a
probability distribution over available actions that maximizes the
average reward per timestep. A reinforcement learning algorithm
which learns a locally optimal stochastic memoryless policy has been
proposed by Jaakkola, Singh and Jordan, but not empirically verified.
We present a variation of this algorithm, discuss its implementation,
and demonstrate its viability using four test problems.
1 INTRODUCTION
Reinforcement learning techniques have proven quite effective in solving Markov
Decision Processes ("MOPs), control problems in which the exact state of the
environment is available to the learner and the expected result of an action depends only
on the present state [10]. Algorithms such as Q-Iearning learn optimal deterministic
policies for "MOPs----rules which for every state prescribe an action that maximizes the
expected future reward. In many important problems, however, the exact state of the
environment is either inherently unknowable or prohibitively expensive to obtain, and
only a limited, possibly stochastic observation of the environment is available. Such
1074
1. K. Williams and S. Singh
Partially Observable Markov Decision Processes (POMDPs) [3 ,6] are often much more
difficult than MDPs to solve [4]. Distinct sequences of observations and actions
preceding a given observation in a POMDP may lead to different probabilities of
occupying the underlying exact states of the MDP. If the efficacy of an action depends
on the hidden exact state of the environment, an optimal choice may require knowing
the past history as well as the current observation, and the problem is no longer Markov.
In light of this difficulty, one approach to solving POMDPs is to explore the
environment while building up a memory of past observations, actions and rewards
which allows estimation of the current hidden state [1]. Such methods produce
deterministic policies, but they are computationally expensive and may not scale well
with problem size. Furthermore, policies that require state-estimation using memory
may be complicated to implement.
Memoryless policies are particularly appropriate for problems in which the state is
expensive to obtain or inherently difficult to estimate, and they have the advantage of
being extremely simple to act upon. For a POMDP, the optimal memoryless policy is
generally a stochastic policy-one which for each observation of the environment
prescribes a probability distribution over the available actions. In fact, examples of
POMDPs can be constructed for which a stochastic policy is arbitrarily better than the
optimal deterministic policy [9] . An algorithm proposed by Jaakkola, Singh and Jordan
OSJ) [2], which we investigate here, learns memoryless stochastic policies for POMDPs.
2
POMDPs AND DIFFERENTIAL-REWARD Q-VALUES
We assume that the environment has discrete states S = {s1, S2, .. IV}, and the learner
chooses actions from a set f4. State transitions depend only on the current state s and the
action a taken (the Markov property); they occur with probabilities r(s,sl) and result in
expected rewards K'(s,s} In a POMDP, the learner cannot sense exactly the state s of
the enVironment, but rather perceives only an observation--or "message"-from a set
:M = {m 1, m 2 , .. m M } according to a conditional probability distribution P(mls). The
learner will in general not know the size of the underlying state space, its transition
probabilities, reward function, or the conditional distributions of the messages.
In MDPs, there always exists a policy which simultaneously maximizes the expected
future reward for all states, but this is not the case for POMDPs [9]. An appropriate
alternative measure of the merit of a stochastic POMDP policy 7Z{alm) is the asymptotic
average reward per timestep, R 7r, that it achieves. In seeking an optimal stochastic
policy, the JSJ algorithm makes use of Q-values determined by the infinite-horizon
differential reward for each observation-action pair (m,a). In particular, if rr denotes the
reward obtained at time t, we may define the differential-reward Q-values by
Q7r(s,a)= LE7r [Ii _R 7r I S1 =s,a 1
= a];
Q7r(m,a)= E s [Q7r(s,a)IM(s)=m](l)
1=1
where M is the observation operator. Note that E[rr] ~ R7r as t ~
so the summand
converges to ~ero. The value functions V7r(s) and V7r(m) may be defined similarly.
00,
3 POLICY IMPROVEMENT
The JSJ algorithm consists of a method for evaluating Q7r and V7r and a mechanism for
using them to improve the current policy. Roughly speaking, if Q7r(m,a) > V7r(m), then
action a realized a higher differential reward than the average for observation m, and
assigning it a slightly greater probability will increase the average reward per timestep,
R7r. We interpret the quantities ~m(a) = Q 7r(m,a) - V7r(m) as comprising a "gradient" of
R7r in policy space. Their projections onto the probability simplexes may then be written
An Algorithm which Learns Stochastic Memoryless Policies for POMDPs
1075
as 8m = Llm -<Llm,l> 11/JIl, where 1 is the one-vector (1,1, ... ,1), <, > is the inner product,
and IJIl is the number of actions, or
1
R
8mea) = Llm(a) - IAI1 LLlm (a') = Q~ (m,a) - -IAI
LQ (m, a').
a'EA
For sufficiently small
E;n,
1l'(a/m) = 1l(alm) +
(2)
a'EA
an improved policy 1l'(alm) may be obtained by the increments
E;n
8m(a) .
(3)
In practice, we also enforce 1l'(alm) ~ Pmin for all a and m to guarantee continued
exploration. The original JSJ algorithm prescribed using Llm(a) in place of 8m(a) in
equation (3), followed by renormalization [2]. Our method has the advantage that a
given value of Ll yields the same incremeiu regardless of the current value of the policy,
and it ensures that the step is in the correct direction. We also do not require the
differential-reward value estimate, yR.
4
Q-EVALUATION
As the POMDP is simulated under a fixed stochastic policy 1l, every occurrence of an
observation-action pair (m, a) begins a sequence of rewards which can be used to
estimate QR(m, a). Exploiting the fact that the QR(m, a) are defined as sums, the JSJ Qevaluation method recursively averages the estimates from all such sequences using a socalled "every-visit" Monte-Carlo method. In order to reduce the bias and variance
caused by the dependence of the evaluation sequences, a factor fJ is used to discount their
shared "tails". Specifically, at time t the learner makes observation m r , takes action ar ,
and obtains reward rr. The number of visits K(mr,ar) is incremented, the tail discount
rate rem, a) = 1- K(m, arl/4, and the following updates are performed (the indicator
function x.:Cm, a) is 1 if (m,a) = (mr,a r) and 0 otherwise).
fJ
%r(m,a)
fJ (m,a)= [1- %r(m,a)]
K(m,a) r(m,a) (m,a)+ K(m,a)
Q(m,a)=
C(m,a)=
[1- ~~::~
[1- ~f:::~
=Q(m, a) -
(4)
]Q(m, a) + fJ(m,a)[Ti - R]
(5)
]c(m,a) + fJ(m, a)
(6)
R = (1 - lIt)R + (lit) rr
Q(m, a)
(tail discount factor)
(cumulative discount effect)
(R~-estimate)
C(m, a) [R - R old ]; Rold
(7)
=R
(QR-estimate correction)
(8)
Other schedules for rem, a) are possible----see [2~and the correction provided by (8)
need not be performed at every step, but can be delayed until the Q~-estirnate is needed.
This evaluation method can be used as given for a policy-iteration type algorithm in
which independent T-step evaluations of Q~ are interspersed with policy improvements
as prescribed in section 3. However, an online version of the algorithm which performs
policy improvement after every step requires that old experience be gradually "forgotten"
so that the QR-estimate can respond to more recent experience. To achieve this, we
multiply the previous estimates of fJ, Q, and C at each timestep by a "decay" factor
a, 0 < a< 1, before they are updated via equations (4)-(6), and replace equation (7) by
R = a(l - lit) R + [1 - a(1 - lit)] r l
.
An alternative method, which also works reasonably well, is to multiply K and t by
each timestep instead.
(9)
a at
J K. Williams and S. Singh
1076
0. '
(a)
r --
-
r --
-
r --
----;,.--
- - - - ; -- - - - - ;
(b)
A
.,...
.. .. .. .. j...... .
B
10000
20000
3 0000
40000
50000
number of iterations
+1
+1
(c)
0 .8
f
\
06
~::"==_:::_-=-__~~.~=~.-....:'.. . .'-"'-" '-"'-" . ---..;...---
\..-:'
::-0
'
[0.4
0. 2
?o!:----,,;-;;c
oo=o-=o --'2::::0~
00:::0-
-'3
""0""00""'0-
--;:
40:;';,00"0
" --=
50000
number Of Iterations
Figure 1: (a) Schematic of confounded two-state POMDP, (b) evolution of the R7r_
estimate, and (c) evolution of n(A) (solid) and nCB) (dashed) for e= 0.0002, a= 0.9995 .
5
EMPIRICAL RESULTS
We present only results from single runs of our online algorithm, including the modified
policy improvement and Q-evaluation procedures described above. Results from the
policy iteration version are qualitatively similar, and statistics performed on multiple
runs verify that those shown are representative of the algorithm's behavior. To simplify
the presentation, we fix a constant learning rate, e, and decay factor, a, for each problem,
and we use Pmin = 0.02 throughout. Note, however, that appropriate schedules or online
heuristics for decreasing e and Pmin while increasing a would improve performance and
are necessary to ensure convergence. Except for the first problem, we choose the initial
policy n to be uniform. In the last two problems, values of n(alm) < 0.03 are rounded
down to zero, with renormalization, before the learned policy is evaluated.
]S]
5.1
CONFOUNDED TWO-STA TE PROBLEM
The two-state MDP diagrammed in Figure l(a) becomes a POMDP when the two states
are confounded into a single observation. The learner may take action A or B, and
receives a reward of either +1 or -1; the state transition is deterministic, as indicated in
the diagram. Note that either stationary deterministic policy results in R7r = -1 , whereas
the optimal stochastic policy assigns each action the probability 112, resulting in R7r = O.
The evolution of the R7r-estimate and policy, starting from the initial policy n(A) = 0.1
and nCB) = 0.9, is shown in Figure 1. Clearly the learned policy approaches the optimal
stochastic policy n =(112,112).
5.2
MATRIX GAME: SCISSORS-PAPER-STONE-GLASS-WATER
Scissors-Paper-Stone-Glass-Water (SPSGW), an extension of the well-known ScissorsPaper-Stone, is a symmetric zero-sum matrix game in which the learner selects a row i,
the opponent selects a column j, and the learner' s payoff is determined by the matrix
entry M(i,j). A game-theoretic solution is a stochastic (or "mixed") policy which
guarantees the learner an expected payoff of at least zero. It can be shown using linear
programming that the unique optimal strategy for SPSGW, yielding R7r = 0, is to play
stone and water with probability 1/3, and to play scissors, paper, and glass with
probability 119 [7]. Any stationary deterministic policy results in R7r = -1, since the
opponent eventually learns to anticipate the learner's choice and exploit it.
An Algorithm which Learns Stochastic Memory/ess Policiesfor POMDPs
(a)
(c)
stone
water
or:---I---\--~
paper
- 0. 4
[0 -1
1
0
-1
-1
1
1 1
M= -1
. ...
-0 5 O~--='-=OO::::OO:-----:::20=:':O=OO"-----:::300'-!:'OO:::::---:-::40::':::OOO:::-----:5;-;::'OOOO
number of iterations
scissors
(b)
1077
(d)
-1]
1 1
1 -1 -1
1
0 -1
1 0 -1
-1
1 0
0. 8
-___
~ ~.=
_______ .______________ __ _
s
__ _
%~-~1=OO~OO~~2~OO~OO~~3~OO~OO~~4~OO~OO~~50000
number of iteratio ns
Figure 2: (a) Diagram of Scissors-Paper-Stone-Glass-Water, (b) the payoff matrix,
(c) evolution of the RJr-estimate, and (d) evolution of n(stone) and n(water) (solid) and
n(scissors), n(paper), and n(glass) (dashed) for ?= 0.00005, a= 0.9995.
In formulating SPSGW as a POMDP, it is necessary to include in the state sufficient
information to allow the opponent to exploit any sub-optimal strategy. We thus choose
as states the learner's past action frequencies, multiplied at each timestep by the decay
factor, a. There is only one observation, and the learner acts by selecting the "row"
scissors, paper, stone, glass or water, producing a deterministic state transition. The
simulated opponent plays the column which maximizes its expected payoff against the
estimate of the learner's strategy obtained from the state. The learner's reward is then
obtained from the appropriate entry of the payoff matrix.
The policy n = (0.1124,0.1033,0.3350,0.1117,0.3376) learned after 50,000 iterations
(see Figure 2) is very close to the optimal policy 7i = (119, 119,113,119,1/3).
5.3
PARR AND RUSSELL'S GRID WORLD
Parr and Russell's grid world [S] consists of 11 states in a 4x3 grid with a single obstacle
as shown in Figure 3(a). The learner senses only walls to its immediate east or west and
whether it is in the goal state (upper right comer) or penalty state (directly below the
goal), resUlting in the 6 possible observations (0-3, G and P) indicated in the diagram.
The available actions are to move N, E, S, or W, but there is a probability 0.1 of slipping
to either side and only O.S of moving in the deSired direction; a movement into a wall
results in bouncing back to the original state. The learner receives a reward of + 1 for a
transition into the goal state, -1 for a transition into the penalty state, and -0.04 for all
other transitions. The goal and penalty states are connected to a cost-free absorbing
state; when the learner reaches either of them it is teleported immediately to a new start
state chosen with uniform probability.
The results are shown in Figure 3. A separate 106 -step evaluation of the final learned
policy resulted in RJr = 0.047. In contrast, the optimal deterministic policy indicated by
arrows in Figure 3(a) yields R Jr = 0.024 [5], while Parr and Russell's memory-based
SPOVA-RL algorithm achieved RJr = 0.12 after learning for 400,000 iterations [S].
5.4
MULTI-SERVER QUEUE
At each timestep, an arriving job having type 1, 2, or 3 with probability 112, 113 or 116,
respectively, must be assigned to server A, B or C; see Figure 4(a). Each server is
optimized for a particular job type which it can complete in an expected time of 2.6
J K. Williams and S. Singh
1078
0.06
(a)
(b)
0 04
~
"...
0 .0 2
;j'
~
0
t?
3
a:
- 0 .04
- 0 .06
-1
-0.0 8
(c)
t
0
~
~
2
2
1
0
40000
20000
60000
80000
100000
nurrber of itera1iorlS;
P
0
~
0
~ - 0 .02
2
2
t
~
+1
~
91
rO.
7r(alm) = 8:8i
0.02
0.21
0.34
0.02 0.43
0.52J
0.36
0.60 0.18
0.02 0.11
0.02 0.19
Figure 3: (a) Parr and Russell's grid world, with observations shown in lower right
corners and the optimal deterministic memoryless policy represented by arrows,
(b) evolution of the R7r-estimate, and (c) the resulting learned policy (observations 0-3
across columns, actions N, E, S, W down rows) for E= 0.02, a= 0.9999.
timesteps, while the other job types require 50% longer. All jobs in a server's queue are
handled in parallel, up to a capacity of 10 for each server; they finish with probability Ilf
at each timestep, where f is the product of the expected time for the job and the number
of jobs in the server's queue. The states for this POMDP are all combinations of waiting
jobs and server occupancies of the three job types, but the learner's observation is
restricted to the type of the waiting job. The state transition is obtained by removing all
jobs which have finished and adding the waiting job to the chosen server if it has space
available. The reward is + 1 if the job is successfully placed, or 0 if it is dropped.
The results are shown in Figure 4. A separate 106- step evaluation of the learned policy
obtained R7r = 0.95, corresponding to 95% success in placing jobs. In contrast, the
optimal deterministic policy, which assigns each job to the server optimized for it,
attained only 87% success. Thus the learned policy more than halves the drop rate!
6
CONCLUSION
Our online version of an algorithm proposed by Jaakkola, Singh and Jordan efficiently
learns a stochastic memoryless policy which is either provably optimal or at least
superior to any deterministic memoryless policy for each of four test problems. Many
enhancements are possible, including appropriate learning schedules to improve
performance and ensure convergence, estimation of the time between observation-action
visits to obtain better discount rates r and thereby enhance Q7r-estimate bias and variance
reduction (see [2]), and multiple starts or simulated annealing to avoid local minima. In
addition, observations could be extended to include some past history when appropriate.
Most POMDP algorithms use memory and attempt to learn an optimal deterministic
policy based on belief states. The stochastic memoryless policies learned by the JSJ
algorithm may not always be as good, but they are simpler to act upon and can adapt
smoothly in non-stationary environments. Moreover, because it searches the space of
stochastic policies, the JS] algorithm has the potential to find the optimal memoryless
policy. These considerations, along with the success of our simple implementation,
suggest that this algorithm may be a viable candidate for solving real-world POMDPs,
including distributed control or network admission and routing problems in which the
numbers of states are enormous and complete state information may be difficult to obtain
or estimate in a timely manner.
An AlgOrithm which Learns Stochastic Memoryless Policiesjor POMDPs
(a)
Server A
(b)
095
TA = (2.6,3.9,3.9)
Job
arrival
of type
1,2,or 3
1079
I
.
09
I
a:
Server B
TB = (3.9,2.6,3.9)
o. 8o'----=20~00::-:0-----,-:
4oo
~00
=---60~00-=-0--::8~00'":c00::------:-=-::'
1 00000
number of iterations
Server C
Tc
=(3.9,3.9,2.6)
(c)
[0.73 0.02 0.02]
n(alm) = 0.02 0.96 0.09
0.25 0.02 0.89
Figure 4: (a) Schematic of the multi-server queue, (b) evolution of the R71-estimate, and
(c) the resulting learned policy (observations I, 2, 3 across columns, actions A, B, C
down rows) for ?= 0.005, a= 0.9999.
Acknowledgements
We would like to thank Mike Mozer and Tim Brown for helpful discussions.
Satinder Singh was funded by NSF grant IIS-9711753.
References
[1]
Chrisman, L. (1992). Reinforcement learning with perceptual aliasing: The
perceptual distinctions approach.
In Proceedings of the Tenth National
Conference on Artificial Intelligence.
[2]
Jaakkola, T., Singh, S. P., and Jordan, M. I. (1995). Reinforcement learning
algorithm for partially observable Markov decision problems. In Advances in
Neural Information Processing Systems 7.
[3]
Littman, M., Cassandra, A., and Kaelbling, L. (1995). Learning poliCies for
partially observable environments: Scaling up. In Proceedings of the Twelfth
International Conference on Machine Learning.
[4]
Littman, M. L. (1994). Memoryless policies: Theoretical limitations and practical
results. Proceedings of the Third International Conference on Simulation of
Adaptive Behavior: From Animals to Animats.
[5]
Loch, J., and Singh, S. P. (1998). Using eligibility traces to find the best
memoryless policy in partially observable Markov decision processes. In Machine
Learning: Proceedings of the Fifteenth International Conference.
[6]
Lovejoy, W. S. (1991). A survey of algorithmic methods for partially observable
Markov decision processes. In Annals of Operations Research, 28.
[7]
Morris, P. (1994). Introduction to Game Theory. Springer-Verlag, New York.
[8]
Parr, R. and Russell, S. (1995). Approximating optimal poliCies for partially
In Proceedings of the International Joint
observable stochastic domains.
Conference on Artificial Intelligence.
[9]
Singh, S. P., Jaakkola, T., and Jordan, M. I. (1994). Learning without stateestimation in partially observable Markovian decision processes. In Machine
Learning: Proceedings of the Eleventh International Conference.
[10] Sutton, R. S. and Barto, A. G. (1998). Reinforcement Learning: An Introduction.
MIT Press.
| 1509 |@word version:3 twelfth:1 simulation:1 seek:1 thereby:1 solid:2 recursively:1 reduction:1 initial:2 att:1 efficacy:1 selecting:1 past:4 current:5 com:1 assigning:1 written:1 must:1 john:1 remove:1 drop:1 update:1 stationary:3 half:1 intelligence:2 yr:1 es:1 simpler:1 admission:1 along:1 constructed:1 differential:5 viable:1 consists:2 eleventh:1 manner:1 alm:7 expected:8 behavior:2 roughly:1 multi:2 aliasing:1 rem:2 decreasing:1 increasing:1 perceives:1 begin:1 provided:1 underlying:2 becomes:1 maximizes:4 moreover:1 cm:1 nj:1 guarantee:2 forgotten:1 every:5 act:3 ti:1 iearning:2 exactly:1 unknowable:1 prohibitively:1 ro:1 control:2 grant:1 producing:1 before:2 dropped:1 local:1 sutton:1 co:1 limited:1 unique:2 practical:1 practice:1 implement:1 x3:1 procedure:1 empirical:1 r7r:10 projection:1 suggest:1 cannot:1 onto:1 close:1 operator:1 mea:1 ilf:1 deterministic:12 williams:4 regardless:1 starting:1 pomdp:10 survey:1 assigns:2 immediately:1 rule:1 continued:1 variation:1 increment:1 updated:1 annals:1 play:3 colorado:2 exact:4 programming:1 prescribe:1 expensive:4 particularly:1 mike:1 ensures:1 connected:1 russell:5 incremented:1 movement:1 mozer:1 environment:11 reward:20 littman:2 diagrammed:1 prescribes:2 singh:11 solving:3 depend:1 upon:2 learner:18 comer:1 po:1 joint:1 represented:1 distinct:1 effective:2 q7r:6 monte:1 artificial:2 quite:1 heuristic:1 solve:1 otherwise:1 florham:1 statistic:1 final:1 online:4 sequence:4 advantage:2 rr:4 ero:1 product:2 achieve:1 qr:4 exploiting:1 convergence:2 enhancement:1 produce:1 converges:1 tim:1 oo:14 job:15 arl:1 direction:2 correct:1 f4:1 stochastic:20 exploration:1 routing:1 require:4 fix:1 wall:2 anticipate:1 im:1 c00:1 extension:1 correction:2 sufficiently:1 algorithmic:1 mop:3 parr:5 nurrber:1 achieves:1 estimation:4 occupying:1 successfully:1 mit:1 clearly:1 always:2 modified:1 rather:1 avoid:1 barto:1 jaakkola:5 improvement:4 contrast:2 sense:1 glass:6 helpful:1 lovejoy:1 hidden:2 comprising:1 selects:2 provably:1 socalled:1 animal:1 having:1 placing:1 park:2 lit:4 future:2 simplify:1 summand:1 sta:1 simultaneously:1 resulted:1 national:1 delayed:1 attempt:1 message:2 investigate:1 multiply:2 evaluation:7 yielding:1 light:1 sens:1 partial:1 necessary:2 experience:2 iv:1 old:2 desired:1 ncb:2 theoretical:2 column:4 obstacle:1 jil:1 markovian:1 ar:2 cost:1 kaelbling:1 entry:2 uniform:2 chooses:1 international:5 rounded:1 enhance:1 choose:2 possibly:1 corner:1 pmin:3 potential:1 caused:1 scissors:7 depends:2 performed:3 lab:1 start:2 complicated:1 parallel:1 timely:1 jsj:5 variance:2 efficiently:1 v7r:5 yield:2 euclid:1 carlo:1 notoriously:1 pomdps:10 history:2 reach:1 against:1 simplexes:1 frequency:1 popular:1 oooo:1 schedule:3 ea:2 back:1 higher:1 attained:1 ta:1 improved:1 iai:1 ooo:1 evaluated:1 furthermore:1 until:1 receives:2 indicated:3 mdp:2 building:1 effect:1 verify:1 brown:1 evolution:7 assigned:1 memoryless:15 symmetric:1 ll:1 game:4 eligibility:1 stone:8 theoretic:1 demonstrate:1 complete:2 performs:1 fj:6 consideration:1 superior:1 absorbing:1 empirically:1 rl:1 interspersed:1 tail:3 interpret:1 grid:4 mathematics:1 similarly:1 baveja:1 funded:1 moving:1 longer:3 j:1 recent:1 verlag:1 server:13 arbitrarily:1 success:3 slipping:1 minimum:1 greater:1 preceding:1 mr:2 dashed:2 ii:2 multiple:2 adapt:1 visit:3 schematic:2 fifteenth:1 iteration:8 achieved:1 whereas:1 addition:1 annealing:1 diagram:3 jordan:5 viability:1 finish:1 timesteps:1 observability:1 inner:1 reduce:1 avenue:1 knowing:1 whether:1 handled:1 penalty:3 queue:4 speaking:1 york:1 constitute:1 action:21 generally:1 discount:5 locally:1 morris:1 sl:1 nsf:1 per:3 discrete:1 waiting:3 four:2 enormous:1 verified:1 tenth:1 timestep:8 sum:2 run:2 respond:1 bouncing:1 place:1 throughout:1 decision:8 scaling:1 followed:1 occur:1 extremely:1 prescribed:2 formulating:1 department:1 according:1 teleported:1 combination:1 jr:1 across:2 slightly:1 s1:2 gradually:1 restricted:1 boulder:1 taken:1 computationally:1 equation:3 discus:1 eventually:1 mechanism:1 needed:1 know:1 merit:1 confounded:3 available:6 operation:1 opponent:4 multiplied:1 appropriate:6 enforce:1 occurrence:1 alternative:3 original:2 denotes:1 ensure:2 include:2 exploit:2 approximating:1 seeking:1 move:1 realized:1 quantity:1 strategy:3 dependence:1 gradient:1 separate:2 thank:1 simulated:3 capacity:1 water:7 loch:1 difficult:3 trace:1 implementation:2 policy:57 animats:1 upper:1 observation:22 markov:10 immediate:1 payoff:5 extended:1 spova:1 pair:2 optimized:2 learned:9 distinction:1 chrisman:1 below:1 tb:1 including:3 memory:6 belief:1 difficulty:2 indicator:1 occupancy:1 improve:3 mdps:2 finished:1 acknowledgement:1 asymptotic:1 mixed:1 limitation:1 proven:1 sufficient:1 row:4 placed:1 last:1 free:1 arriving:1 bias:2 allow:1 side:1 distributed:1 transition:8 evaluating:1 cumulative:1 world:4 qualitatively:1 reinforcement:7 adaptive:1 observable:9 obtains:1 satinder:2 ml:1 llm:4 search:1 osj:1 learn:2 reasonably:1 inherently:2 domain:1 arrow:2 s2:1 arrival:1 representative:1 west:1 renormalization:2 n:1 sub:1 lq:1 candidate:1 perceptual:2 third:1 learns:7 down:3 removing:1 decay:3 exists:1 adding:1 te:1 horizon:1 cassandra:1 smoothly:1 tc:1 explore:1 partially:9 springer:1 conditional:2 goal:4 presentation:1 shared:1 absence:1 replace:1 determined:2 infinite:1 specifically:1 except:1 experimental:1 east:1 |
558 | 151 | 703
WINNER-TAKE-ALL
NETWORKS OF O(N) COMPLEXITY
J. Lazzaro, S. Ryckebusch, M.A. Mahowald, and C. A. Mead
California Institute of Technology
Pasadena, CA 91125
ABSTRACT
We have designed, fabricated, and tested a series of compact CMOS
integrated circuits that realize the winner-take-all function. These
analog, continuous-time circuits use only O(n) of interconnect to
perform this function. We have also modified the winner-take-all
circuit, realizing a circuit that computes local nonlinear inhibition.
Two general types of inhibition mediate activity in neural systems: subtractive inhibition, which sets a zero level for the computation, and multiplicative (nonlinear)
inhibition, which regulates the gain of the computation. We report a physical realization of general nonlinear inhibition in its extreme form, known as winner-take-all.
We have designed and fabricated a series of compact, completely functional CMOS
integrated circuits that realize the winner-take-all function, using the full analog
nature of the medium. This circuit has been used successfully as a component
in several VLSI sensory systems that perform auditory localization (Lazzaro and
Mead, in press) and visual stereopsis (Mahowald and Delbruck, 1988). Winnertake-all circuits with over 170 inputs function correctly in these sensory systems.
We have also modified this global winner-take-all circuit, realizing a circuit that
computes local nonlinear inhibition. The circuit allows multiple winners in the network, and is well suited for use in systems that represent a feature space topographically and that process several features in parallel. We have designed, fabricated,
and tested a CMOS integrated circuit that computes locally the winner-take-all
function of spatially ordered input.
704
Lazzaro, Ryckebusch, Mahowald and Mead
THE WINNER-TAKE-ALL CmCUIT
Figure 1 is a schematic diagram of the winner-take-all circuit. A single wire, associated with the potential Vc, computes the inhibition for the entire circuit; for an
n neuron circuit, this wire is O(n) long. To compute the global inhibition, each
neuron k contributes a current onto this common wire, using transistor T2 a.' To
apply this global inhibition locally, each neuron responds to the common wire voltage Vc, using transistor Tla.' This computation is continuous in time; no clocks
are used. The circuit exhibits no hysteresis, and operates with a time constant
related to the size of the largest input. The output representation of the circuit
is not binary; the winning output encodes the logarithm of its associated input.
Figure 1. Schematic diagram of the winner-take-all circuit. Each neuron receives
a unidirectional current input 11;; the output voltages VI ?.. VB represent the result
of the winner-take-all computation. If II; = max(II ??? I B ), then VI; is a logarithmic
function of 11;; if Ii <: 11;, then Vi ~ O.
A static and dynamic ana.lysis of the two-neuron circuit illustrates these system
properties. Figure 2 shows a schematic diagram of a two-neuron winner-take-all
circuit. To understand the beha.vior of the circuit, we first consider the input
condition II = 12
1m. Transistors TIl ~d T12 have identical potentials at gate
and source, and are both sinking 1m; thus, the drain potentials VI and V2 must be
equal. Transistors T21 and T22 have identical source, drain, and gate potentials,
and therefore must sink the identical current ICI = IC2 = I c/2. In the subthreshold
region of operation, the equation 1m = 10 exp(Vc/Vo) describes transistors Til and
T 12 , where 10 is a fabrication parameter, and Vo = kT/qlt. Likewise, the equation
Ic/2 = 10 exp((Vm - Vel/Volt where Vm
VI = V2, describes transistors T21 and
T22 . Solving for Vm(Im, Ie) yields
=
=
Vm
= Voln(~:) + Voln(:;).
(1)
Winner-Take-All Networks ofO(N) Complexity
Thus, for equal input currents, the circuit produces equal output voltages; this
behavior is desirable for a winner-take-all circuit. In addition, the output voltage
Vm logarithmically encodes the magnitude of the input current 1m.
Figure 2. Schematic diagram of a two-neuron winner-take-all circuit.
The input condition II = 1m + Oi, 12 = 1m illustrates the inhibitory action of the
circuit. Transistor Til must sink 0, more current than in the previous example; as a
result, the gate voltage of Til rises. Transistors Tit and TI2 share a common gate,
howeverj thus, TI2 must also sink 1m + 0,. But only 1m is present at the drain of
T12 ? To compensate, the drain voltage of T12 , V2, must decrease. For small OiS, the
Early effect serves to decrease the current through Th , decreasing V2 linearly with
0,. For large o's, TI2 must leave saturation, driving V2 to approximately 0 volts.
As desired, the output associated with the smaller input diminishes. For large OiS,
Ie2 $!:::f 0, and Iel $!:::f Ie. The equation 1m + 0, = 10 exp(Ve/Vo) describes transistor
Til' and the equation Ie = 10 exp((VI - Vel/Yo) describes transistor T21 ? Solving
for VI yields
(2)
The winning output encodes the logarithm of the associated input. The symmetrical
circuit topology ensures similar behavior for increases in 12 relative to II.
Equation 2 predicts the winning response of the circuit; a more complex expression,
derived in (Lazzaro et.al., 1989), predicts the losing and crossover response of the
circuit. Figure 3 is a plot of this analysis, fit to experimental data. Figure 4 shows
the wide dynamic range and logarithmic properties of the circuitj the experiment in
Figure 3 is repeated for several values of 12 , ranging over four orders of magnitude.
The conductance of transistors Til and T1:a determines the losing response of the
circuit. Variants of the winner-take-all circuit shown in (Lazzaro et. aI., 1988)
achieve losing responses wider and narrower than Figure 3, using circuit and mask
layout techniques.
705
706
Lazzaro, Ryckebusch, Mahowald and Mead
WINNER-TAKE-ALL TIME RESPONSE
A good winner-take-all circuit should be stable, and should not exhibit damped
oscillations ("ringing") in response to input changes. This section explores these
dynamic properties of our winner-take-all circuit, and predicts the temporal response of the circuit. Figure 8 shows the two-neuron winner-take-all circuit, with
capacitances added to model dynamic behavior.
o
T
102
Vo
Ie
Figure 8. Schematic diagram of a two-neuron winner-take-all circuit, with capacitances added for dynamic analysis. 0 is a large MOS capacitor added to each
neuron for smoothingj 0., models the parasitic capacitance contributed by the gates
of Tu and T 12 , the drains of T21 and T22, and the interconnect.
(Lazzaro et. al., 1988) shows a small-signal analysis of this circuit. The transfer
function for the circuit has real poles, and thus the circuit is stable and does not ring,
if 10 > 41(Oe/O), where 11 RlI2 Rl 1. Figure 9 compares this bound with experimental
data.
H Ie > 41(0 0 /0), the circuit exhibits first-order behavior. The time constant OVo/I
sets the dynamics of the winning neuron, where Vo = A:T /qK. Rl 40 mV. The time
constant OVE/I sets the dynamics of the losing neuron, where VE Rl 50 v. Figure 10
compares these predictions with experimental data.
Winner-Take-All Networks ofO(N) Complexity
2.6
Vl,V,
(V)
2.4
2.2
2.0
I.S
1.6
1.4
1.2
1.0+--+--+--+--.....~I----t~--t---f
0.92 0.94 0.96 0.98 1.00 1.02 1.04 1.06 1.0S
II/I,
Figure 8. Experimental data (circles) and theory (solid lines) for a two-neuron
winner-take-all circuit. II, the input current of the first neuron, is swept about the
value of 12, the input current of the second neuron; neuron voltage outputs VI and
V2 are plotted versus normalized input current.
2.6
I.S
1.6
1.4
1.2
10- 11
10- 10
10- 0
IdA)
10- 8
Figure 4. The experiment of Figure 3 is repeated for several values of 12; experimental data of output voltage response are plotted versus absolute input current on
a log scale. The output voltage VI = V2 is highlighted with a circle for each experiment. The dashed line is a theoretical expression confirming logarithmic behavior
over four orders of magnitude (Equation 1).
707
708
Lazzaro, Ryckebusch, Mahowald and Mead
1
Figure 9. Experimental data (circles) and theoretical statements (solid line) for a
two-neuron winner-take-all circuit, showing the smallest 10 , for a given I, necessary
for a first-order response to a small-signal step input.
Figure 10. Experimental data (symbols) and theoretical statements (solid line) for
a two-neuron winner-take-all circuit, showing the time constant of the first-order
response to a small-signal step input. The winning response (filled circles) and losing
response (triangles) of a winner-take-a.ll circuit are shownj the time constants differ
by several orders of magnit ude.
Winner~Take~AlI
Networks ofO(N) Complexity
THE LOCAL NONLINEAR INHIBITION CIRCUIT
The winner-take-all circuit in Figure 1, as previously explained, locates the largest
input to the circuit. Certain applications require a gentler form of nonlinear inhibition. Sometimes, a circuit that can represent multiple intensity scales is necessary.
Without circuit modification, the winner-take-all circuit in Figure 1 can perform
this task. (Lazzaro et. al., 1988) explains this mode of operation.
Other applications require a local winner-take-all computation, with each winner
having inHuence over only a limited spatial area. Figure 12 shows a circuit that
computes the local winner-taite-all function. The circuit is identical to the original
winner-take-all circuit, except that each neuron connects to its nearest neighbors
with a nonlinear resistor circuit (Mead, in press). Each resistor conducts a current
Ir in response to a voltage ~V across it, where Ir = I.tanh(~V/(2Vo)). 1., the
saturating current of the resistor, is a controllable parameter. The current source,
10, present in the original winner-take-all circuit, is distributed between the resistors
in the local winner-take-all circuit.
Figure 11. Schematic diagram of a section of the local winner-take-all circuit.
Each neuron i receives a unidirectional current input Iii the output voltages Vi
represent the result of the local winner-take-all computation.
To understand the operation of the local winner-take-all circuit, we consider the
circuit response to a spatial impulse, defined as 1" :> 1, where 1 == h~". 1,,:> 1"-1
and 1,,:> 1"+1, so Ve:,. is much larger than Ve:,._l and Ve:lI+l' and the resistor circuits
connecting neuron 1: with neuron 1: - 1 and neuron 1: + 1 saturate. Each resistor
sinks 1. current when saturatedj transistor T2,. thus conducts 21. + Ie: current. In
the subthreshold region of operation, the equation 1" = 10 exp(Ve:,. /Vo) describes
transistor TI ,., and the equation 21. + Ie = Ioexp((V" - Ve:,.)/Vo) describes transistor
709
710
Lazzaro, Ryckebusch, Mahowald and Mead
T2,.. Solving for VA: yields
VA: = voln((2I. + 10 )/10 )
+ voln(IA:/lo).
(4)
As in the original winner-take-all circuit, the output of a winning neuron encodes
the logarithm of that neuron's associated input.
As mentioned, the resistor circuit connecting neuron Ie with neuron Ie - 1 sinks 1.
CUlTent. The current sources 10 associated with neurons Ie -1, Ie - 2, ... must supply
this current. If the current source 10 for neuron Ie - 1 supplies part of this current,
the transistor T2,._1 carries no current, and the neuron output VA:-l approaches zero.
In this way, a winning neuron inhibits its neighboring neurons.
This inhibitory action does not extend throughout the network. Neuron Ie needs
only 1. current from neurons Ie - 1, Ie - 2, .... Thus, neurons sufficiently distant
from neuron Ie maintain the service of their current source 10, and the outputs of
these distant neurons can be active. Since, for a spatial impulse, all neurons Ie - 1,
Ie - 2, ... have an equal input current I, all distant neurons have the equal output
(5)
Similar reasoning applies for neurons
Ie
+ 1, Ie + 2, ....
The relative values of 1. and 10 determine the spatial extent of the inhibitory action.
Figure 12 shows the spatial impulse response of the local winner-take-all circuit, for
different settings of 1./10 ,
o
I
2
4
I
8
10
6
Ie (Pollition)
I
12
I
14
I
16
Figure 12. Experimental data showing the spatial impulse response of the local
winner-take-all circuit, for values of 1./10 ranging over a factor of 12.7. Wider
inhibitory responses correspond to larger ratios. For clarity, the plots are vertically
displaced in 0.25 volt increments.
Winner-Take-All Networks ofO(N) Complexity
CONCLUSIONS
The circuits described in this paper use the full analog nature of MOS devices to
realize an interesting class of neural computations efficiently. The circuits exploit
the physics of the medium in many ways. The winner-take-all circuit uses a single
wire to compute and communicate inhibition for the entire circuit. Transistor TI,.
in the winner-take-all circuit uses two physical phenomena in its computation: its
exponential current function encodes the logarithm of the input, and the finite
conductance of the transistor defines the losing output response. As evolution
exploits all the physical properties of neural devices to optimize system performance,
designers of synthetic neural systems should strive to harness the full potential of
the physics of their media.
Acknow ledgments
John Platt, John Wyatt, David Feinstein, Mark Bell, and Dave Gillespie provided
mathematical insights in the analysis of the circuit. Lyn Dupre proofread the document. We thank Hewlett-Packard for computing support, and DARPA and MOSIS
for chip fabrication. This work was sponsored by the Office of Naval Research and
the System Development Foundation.
References
Lazzaro, J. P., Ryckebusch, S., Mahowald, M.A., and Mead, C.A. (1989). WinnerTake-All Networks of O(N) Oomplexity, Caltech Computer Science Department
Technical Report Caltech-CS-TR-21-88.
Lazzaro, J. P., and Mead, C.A. {in press}. Silicon Models of Auditory Localization,
Neural Oomputation.
Mahowald, M.A., and Delbruck, T.I. (1988). An Analog VLSI Implementation of
the Marr-Poggio Stereo Correspondence Algorithm, Abstracts of the First Annual
INNS Meeting, Boston, 1988, Vol. I, Supplement I, p. 392.
Mead, C. A. (in press). Analog VLSI and Neural Systems. Reading, MA: AddisonWesley.
711
| 151 |@word tr:1 solid:3 carry:1 series:2 document:1 current:26 ida:1 must:7 john:2 realize:3 distant:3 confirming:1 designed:3 plot:2 sponsored:1 device:2 realizing:2 ofo:4 mathematical:1 supply:2 mask:1 behavior:5 decreasing:1 provided:1 circuit:69 medium:3 ringing:1 fabricated:3 temporal:1 ti:2 platt:1 t1:1 service:1 local:11 vertically:1 mead:10 approximately:1 limited:1 range:1 area:1 crossover:1 bell:1 onto:1 optimize:1 cmcuit:1 layout:1 insight:1 marr:1 increment:1 losing:6 us:2 logarithmically:1 predicts:3 t12:3 region:2 ensures:1 oe:1 decrease:2 mentioned:1 complexity:5 dynamic:7 solving:3 tit:1 topographically:1 ove:1 ali:1 localization:2 completely:1 sink:5 triangle:1 darpa:1 chip:1 larger:2 highlighted:1 transistor:17 inn:1 t21:4 tu:1 neighboring:1 realization:1 achieve:1 produce:1 cmos:3 leave:1 ring:1 wider:2 nearest:1 ois:2 c:1 differ:1 vc:3 ana:1 explains:1 require:2 im:1 sufficiently:1 ic:1 exp:5 mo:2 driving:1 early:1 smallest:1 diminishes:1 tanh:1 largest:2 successfully:1 t22:3 modified:2 voltage:11 office:1 derived:1 yo:1 naval:1 interconnect:2 vl:1 integrated:3 entire:2 pasadena:1 vlsi:3 vior:1 development:1 spatial:6 equal:5 having:1 identical:4 report:2 t2:4 ve:7 connects:1 maintain:1 conductance:2 extreme:1 hewlett:1 damped:1 kt:1 necessary:2 poggio:1 addisonwesley:1 filled:1 conduct:2 logarithm:4 desired:1 circle:4 plotted:2 theoretical:3 wyatt:1 delbruck:2 mahowald:8 pole:1 fabrication:2 synthetic:1 explores:1 ie:21 vm:5 physic:2 connecting:2 strive:1 til:6 li:1 potential:5 hysteresis:1 mv:1 vi:10 multiplicative:1 parallel:1 unidirectional:2 oi:1 ir:2 qk:1 efficiently:1 likewise:1 subthreshold:2 yield:3 correspond:1 dave:1 associated:6 static:1 gain:1 auditory:2 harness:1 response:18 vel:2 clock:1 receives:2 nonlinear:7 defines:1 mode:1 impulse:4 effect:1 normalized:1 evolution:1 spatially:1 volt:3 ll:1 gentler:1 vo:8 reasoning:1 ranging:2 common:3 functional:1 physical:3 regulates:1 rl:3 winner:45 ti2:3 analog:5 sinking:1 extend:1 silicon:1 ai:1 winnertake:2 stable:2 inhibition:12 certain:1 binary:1 meeting:1 swept:1 proofread:1 caltech:2 determine:1 signal:3 ii:8 dashed:1 full:3 multiple:2 desirable:1 technical:1 long:1 compensate:1 locates:1 va:3 schematic:6 beha:1 variant:1 prediction:1 represent:4 sometimes:1 addition:1 diagram:6 source:6 capacitor:1 iii:1 ledgments:1 fit:1 topology:1 expression:2 stereo:1 lazzaro:12 action:3 locally:2 inhibitory:4 designer:1 correctly:1 vol:1 four:2 clarity:1 mosis:1 communicate:1 throughout:1 oscillation:1 vb:1 bound:1 correspondence:1 annual:1 activity:1 encodes:5 inhibits:1 magnit:1 department:1 describes:6 smaller:1 across:1 modification:1 explained:1 tla:1 equation:8 previously:1 feinstein:1 serf:1 operation:4 apply:1 v2:7 gate:5 original:3 exploit:2 capacitance:3 added:3 ude:1 ryckebusch:6 responds:1 exhibit:3 thank:1 extent:1 ratio:1 statement:2 acknow:1 rise:1 implementation:1 perform:3 contributed:1 wire:5 displaced:1 neuron:40 finite:1 lyn:1 intensity:1 david:1 california:1 reading:1 saturation:1 max:1 packard:1 gillespie:1 ia:1 technology:1 drain:5 relative:2 interesting:1 versus:2 foundation:1 ovo:1 subtractive:1 share:1 lo:1 understand:2 institute:1 wide:1 neighbor:1 absolute:1 voln:4 distributed:1 computes:5 sensory:2 ici:1 compact:2 global:3 active:1 symmetrical:1 stereopsis:1 continuous:2 nature:2 transfer:1 ca:1 controllable:1 contributes:1 complex:1 ic2:1 linearly:1 mediate:1 taite:1 repeated:2 resistor:7 winning:7 exponential:1 saturate:1 showing:3 symbol:1 supplement:1 magnitude:3 iel:1 illustrates:2 boston:1 suited:1 logarithmic:3 visual:1 ordered:1 saturating:1 applies:1 determines:1 ma:1 narrower:1 change:1 except:1 operates:1 experimental:8 parasitic:1 mark:1 support:1 tested:2 phenomenon:1 |
559 | 1,510 | The Effect of Correlations on the Fisher
Information of Population Codes
Hyoungsoo Yoon
hyoung@fiz.huji.ac.il
Haim Sompolinsky
hairn@fiz.huji.ac.il
Racah Institute of Physics and Center for Neural Computation
Hebrew University, Jerusalem 91904, Israel
Abstract
We study the effect of correlated noise on the accuracy of population coding using a model of a population of neurons that are
broadly tuned to an angle in two-dimension. The fluctuations in
the neuronal activity is modeled as a Gaussian noise with pairwise
correlations which decays exponentially with the difference between
the preferred orientations of the pair. By calculating the Fisher information of the system, we show that in the biologically relevant
regime of parameters positive correlations decrease the estimation
capability of the network relative to the uncorrelated population.
Moreover strong positive correlations result in information capacity which saturates to a finite value as the number of cells in the
population grows. In contrast, negative correlations substantially
increase the information capacity of the neuronal population.
1
Introduction
In many neural systems , information regarding sensory inputs or (intended) motor
outputs is found to be distributed throughout a localized pool of neurons. It is generally believed that one of the main characteristics of the population coding scheme is
its redundancy in representing information (Paradiso 1988; Snippe and Koenderink
1992a; Seung and Sompolinsky 1993). Hence the intrinsic neuronal noise, which
has detrimental impact on the information processing capability, is expected to be
compensated by increasing the number of neurons in a pool. Although this expectation is universally true for an ensemble of neurons whose stochastic variabilities
are statistically independent, a general theory of the efficiency of population coding
when the neuronal noise is correlated within the population, has been lacking. The
conventional wisdom has been that the correlated variability limits the information
pro cessing capacity of neuronal ensembles (Zohary, Shadlen, and Newsome 1994).
H. Yoon and H. Sompolinsky
168
However, detailed studies of simple models of a correlated population that code for
a single real-valued parameter led to apparently contradicting claims. Snippe and
Koenderink (Snippe and Koenderink 1992b) conclude that depending on the details
of the correlations, such as their spatial range, they may either increase or decrease
the information capacity relative to the un correlated one. Recently, Abbott and
Dayan (Abbott and Dayan 1998) claimed that in many cases correlated noise improves the accuracy of population code. Furthermore, even when the information
is decreased it still grows linearly with the size of the population. If true, this conclusion has an important implication on the utility of using a large population to
improve the estimation accuracy. Since cross-correlations in neuronal activity are
frequently observed in both primary sensory and motor areas (Fetz, Yoyama, and
Smith 1991 ; Lee, Port, Kruse, and Georgopoulos 1998), understanding the effect of
noise correlation in biologically relevant situations is of great importance.
In this paper we present an analytical study of the effect of noise correlations on the
population coding of a pool of cells that code for a single one-dimensional variable,
an angle on a plane, e.g. , an orientation of a visual stimulus, or the direction of
an arm movement. By assuming that the noise follows the multivariate Gaussian
distribution, we investigate analytically the effect of correlation on the Fisher information. This model is similar to that considered in (Snippe and Koenderink
1992b; Abbott and Dayan 1998). By analyzing its behavior in the biologically relevant regime of tuning width and correlation range, we derive general conclusions
about the effect of the correlations on the information capacity of the population.
2
Population Coding with Correlated Noise
We consider a population of N neurons which respond to a stimulus characterized
by an angle (), where -1r < () ~ 1r . The activity of each neuron (indexed by i) is
assumed to be Gaussian with a mean h((}) which represents its tuning curve, and a
uniform variance a . The noise is assumed to be pairwise-correlated throughout the
population. Hence the activity profile of the whole population, R = {rl, r2, .. . , r N } ,
given a stimulus () , follows the following multivariate Gaussian distribution.
P(RI(}) = Nexp
(-~ l:(ri - h((}?)Gi-/(rj -
fj((}))
(1)
t ,J
where N is a normalization constant and Cj is the correlation matrix.
G ij =
ac5ij
+ bij (1
- c5ij ).
(2)
It is assumed that the tuning curves of all the neurons are identical in form but
peaked at different angles, that is fi((}) = f((} - ?i) where the preferred angles ?i
are distributed uniformly from -1r to 1r with a lattice spacing, w , which is equal
to 21r IN. We further assume that the noise correlation between a pair of neurons
is only a function of their preferred angle difference, i.e., bij ((}) = b(ll?i - ?jll)
where lI(}l - (}211 is defined to be the relative angle between (}l and (}2, and hence
its maximum value is 1r. A decrease in the magnitude of neuronal correlations with
the dissimilarity in the preferred stimulus is often observed in cortical areas. We
model this by exponentially decaying correlations
bij
= b exp( _1I?i -
p
where p specifies the angular correlation length.
?j II)
(3)
169
Fisher Information of Correlated Population Codes
The amount of information that can be extracted from the above population will
depend on the decoding scheme. A convenient measure of the information capacitv
in the population is given by the Fisher information, which in our case is (for it
given stimulus 8)
J(8)
=L
(4)
giGi-/ gj
i ,j
where
.(e) =
gt
-
{)Iiae(8) .
(5)
The utility of this measure follows from the well known Cramer-Rao bound for the
variance of any unbiased estimators, i.e., ((8 - iJ)2) 2: 1/ J(8). For the rest of this
paper, we will concentrate on the Fisher information as a function of the noise
correlation parameters, band p, as well as the population size N.
3
Results
In the case of un correlated population (b
by (Seung and Sompolinsky 1993)
0) , the Fisher information is given
(6)
n
\vhere gn is the Fourier transform of gj, defined by
1
-
gn = N
Le .
A.
-'l.n'P]
(7)
gj.
j
The mode number n is an integer running from _N:;l to N:;l (for odd N) and
= -7f(N + 1)/N + iw, i = 1, .. . , N. Likewise, in the case of b ::j:. 0, J is given by
?i
J =
NL Ignl
G
n
2
(8)
n
where Gn are the eigenvalues of the covariance matrix,
t,]
(a _ 2&)
+ 2b 1 -
N+I
.\ cos(nuJ) - ( _ 1)11.\ -y- C'os(nw)(1 - .\)
1 - 2,\ cos(nw) + .\2
(9)
7J, .\
where w =
= e- w / p , and N is assumed to be an odd integer. Note that thE'
covariance matrix Gij remains positive definite as long as
(10)
where the lower bound holds for general N while the upper bound is valid for
large N.
To evaluate the effect of correlations in a large population it is important to specify
the appropriate scales of the system parameters. We consider here the biologically
relevant case of broadly tuned neurons that have a smoothly varying tuning curve
with a single peak. When the tuning curve is smoothly varying, Ignl 2 will be a
rapidly decaying function as n increases beyond a characteristic value which is
H. Yoon and H. Sompolinsky
170
proportional to the inverse of the tuning width, a. We further assume a broad
tuning, namely that the tuning curve spans a substantial fraction of the angular
extent. This is consistent with the observed typical values of half-width at half
height in visual and motor areas, which range from 20 to 60 degrees . Likewise , it
is reasonable to assume that the angular correlation length p spans a substantial
fraction of the entire angular range. This broad tuning of correlations with respect
to the difference in the preferred angles is commonly observed in cortex (Fetz,
Yoyama, and Smith 1991 ; Lee , Port, Kruse, and Georgopoulos 1998). To capture
these features we will consider the limit of large N while keeping the parameters p
and a constant. Note that keeping a of order 1 implies that substantial contributions
to Eq. (8) come only from n which remain of order 1 as N increases. On the
other hand, given the enormous variability in the strength of the observed crosscorrelations between pairs of neurons in cortex, we do not restrict the value of b at
this point.
Incorporating the above scaling we find that when N is large 1 is given by
N
2
p- 2 + n 2
,
1=~~19nl p-2+n2+(~)(1-(-1)ne-7I'/p) '
(11)
Inspection of the denominator in the above equation clearly shows that for all
positive values of b, 1 is smaller than 1 0 , On the other hand, when b is negative
1 is larger than 10 , To estimate the magnitude of these effects we consider below
three different regimes.
1.0 """"'----,-----r---.....,------,
D.8
.I
.10
0.6
0.4
0.2
0.0
()
1000
2000
3000
400()
N
Figure 1: Normalized Fisher information when p '" 0(1) (p
0.257r was used).
a = 1 and b = 0.1 , 0.01, and 0.001 from the bottom. We used a circular Gaussian
tuning curve, Eq. (13), with fmax = 10 and a = 0.27r.
=
Strong positive correlations: We first discuss the regime of strong positiw
correlations, by which we mean that a < b/a "" 0(1). In this case the second term
in the denominator of Eq. (11) is of order Nand Eq. (11) becomes
7rp
2
p-2 + n 2
1=b
19n1 1 _ (-1)ne-7I'/P'
(12)
L
n
This result implies that in this regime the Fisher information in the entire population does not scale linearly with the population size N but saturates to a sizeindependent finite limit . Thus, for these strong correlations, although the number
of neurons in the population may be large, the number of independent degrees of
freedom is small.
We demonstrate the above phenomenon by a numerical evaluation of 1 for the
following choice of tuning curve
f(O)
= fmax
exp ((cos(O) - 1)/a 2 )
(13)
171
Fisher Information of Correlated Population Codes
with (J = 0.211". The results are shown in Fig. 1 and Fig. 2. The results of Fig. 1
clearly show the substantial decrease in J as b increases. The reduction in J I J o
when b '" 0(1) indicates that J does not scale with N in this limit. Fig. 2 shows t.he
saturation of J when N increases. For p = 0.1 and 1 ((c) and (d)), J saturates at.
about N = 100, which means that for these parameter values the network contains
at most 100 independent degrees of freedom. When the correlation range becomes
either smaller or bigger, the saturation becomes less prominent (( a) and (b)) , which
is further explained later in the text.
40
.30
J
20
(c)
10
(ti)
200
400
600
800
N
Figure 2: Saturation of Fisher information with the correlation coefficient kept fixed:
a = 1 and b = 0.5. Both p '" 0(1) ((c) p = 0.1 and (d) p = 1) and other extreme
limits ((a) p = 0.01 and (b) p = 10) are shown. Tuning curve with fmax = 1 and
(J = 0.211" was used for all four curves .
Weak positive correlations: This regime is defined formally by positive values
In this case, while J is still smaller than .10 the
of b which scale as bla '" O(
suppressive effects of the correlations are not as strong as in the first case. This is
shown in Fig. 3 (bottom traces) for N = 1000. While J is less than J o , it is still a
substantial fraction of J o , indicating J is of order N.
-k).
2.3
2.0
.:L
J"
1.3
1.0
0.3
0
1
2
3
4
P
Figure 3: Normalized Fisher information when p '" 0(1) and bla '" O(~). N =
1000, a = 1, fmax = 10, and (J = 0.211". The top curves represent negative h
(b = -0.005 and -0.002 from the top) and the bottom ones positive b (b = 0.01
and 0.005 from the bottom).
Weak negative correlations: So far we have considered the case of positive b.
As stated above, Eq. (11) implies that when b < 0, J > J o . The lower bound of b
(Eq. (10)) means that when the correlations are negative and p is of order 1 th!'
amplitude of the c:orrelations must be small. It scales as bla = biN with b which
is of order 1 and is larger than bmin = -(11"lp)/(I- exp(-11"lp)). In this regime
(.] - .10 )IN retains a finite positive value even for large N. This enhancement call
H. Yoon and H. Sompolinsky
172
,
,
be made large if b comes close to bmin . This behavior is shown in Fig. 3 (upper
traces). Note that, for both positive and negative weak correlations, the curves
have peaks around a characteristic length scale p '" a, which is 0.211" in this figure.
Extremely long and short range correlations: Calculation with strictly uniform correlations, i. e., bij = b, shows that in this case the positive correlations
enhance the Fisher information of the system, leading to claims that this might
be a gelleri<: result (Abbott and Dayan 1998). Here we show that this behavior
is special to cases where the correlations essentially do not vary in strength. We
consider the case p '" O(N). This means that the strength of the correlations is the
same for all the neurons up to a correction of order liN. In this limit Eq. (11) is
not valid, and the Fisher information is obtained from Eq. (8) and Eq. (9),
(14)
where {} = wpl4. Note that even in this extreme regime, only for {} > 1 is 1
guaranteed to be always larger than .10' Below this value the sign of 1 -.10 depends
on the particular shape of the tuning curve and the value of b. In fact, a more
detailed analysis (Yoon and Sompolinsky 1998) shows that as soon as p? O(VN),
1 - 10 < 0, as in the case of p rv 0(1) discussed above. The crossover between these
two opposite behaviors is shown in Fig. 4. For comparison the case with p rv 0(1)
is also shown.
4.0
3.0
.J
.In
2.0
1.D
n.o
n.2
n.D
0.4
0.8
D.G
1.0
b
Figure 4: Normalized Fisher information when bla rv 0(1). N = 1000 and a = 1.
When p '" 0(1), increasing b always decreases the Fisher information (bottom curve
p = 0.2511"). However, this trend is reversed when p ,. . ., O(VN) and when p > ~N
.1 - .10 becomes always positive. From the top p = 400, 50, and 25 .
Another extreme regime is where the correlation length p scales as 1IN but the
tuning width remains of order 1. This means that a given neuron is correlated with
a small number of its immediate neighbors, which remains finite as N ~ 00. In this
limit , the Fishel' information becomes, again from Eq. (8) and Eq. (9),
_
N(>..-l_l)
2
1 - a(>..-1_1)+2bI:19nl.
(15 )
"
In this case, the behavior of 1 is similar to the cases of weak correlations discussed
above. The information remains of order N but the sign of 1 - 10 depends on the
sign of b. Thus, when the amplitude of the positive correlation function is 0(1), .]
increases linearly with N in the two opposite extremes of very large and very small
p as shown in Fig. 2 ((a) and (b)).
Fisher Information of Correlated Population Codes
4
173
Discussion
In this paper we have studied the effect of correlated variability of neuronal activity
OIl the maximum accuracy of the population coding. We have shown that the
effect of correlation on the information capacity of the population crucially depends
on the scale of correlation length. We argue that for the sensory and motor areas
which are presumed to utilize population coding, the tuning of both the correlations
and the mean response profile is broad and of the same order. This implies that
each neuron is correlated with a finite fraction of the total number of neurons,
N, and a given stimulus activates a finite fraction of N. We show that in this
regime positive correlations always decrease the information. When they are strong
enough in amplitude they reduce the number of independent degrees of freedom
to a finite number even for large population. Only in the extreme case of almost
uniform correlations the information capacity is enhanced. This is reasonable since
to overcome the positive correlations one needs to subtract the responses of different
neurons. But in general this will reduce their signal by a larger amount. When the
correlations are uniform, the reduction of the correlated noise by subtraction is
perfect and can be made in a manner that will little affect the signal component.
Acknow ledgments
H.S. acknowledges helpful discussions with Larry Abbott and Sebastian Seung. This
research is partially supported by the Fund for Basic Research of the Israeli Academy
of Science and by a grant from the United States-Israel Binational Science Foundation (BSF), Jerusalem, Israel.
References
L. F. Abbott and P. Dayan (1998). The effect of correlated variability on the
accuracy of a population code. Neural Camp., in press.
E. Fetz , K. Yoyama, and W. Smith (1991). Synaptic interactions between cortical
neurons. In A. Peters and E. G. Jones (Eds.) , Cerebral Cortex, Volume 9. New
York: Plenum Press.
D. Lee, N. L. Port, W. Kruse, and A. P. Georgopoulos (1998). Variability and
correlated noise in the discharge of neurons in motor and parietal areas of the
primate cortex. J. Neurosci. 18, 1161- 1170.
M. A. Paradiso (1988). A theory for the use of visual orientation informatioll
which exploits the columnar structure of striate cortex. BioI. Cybern . 58, 35- 49.
H. S. Seung and H. Sompolinsky (1993). Simple models for reading neuronal
population codes. Proc . Natl. Acad. Sci. USA 90, 10749- 10753.
H. P. Snippe and J. J. Koenderink (1992a). Discrimination thresholds for channelcoded Hystems. Biol. Cybern. 66, 543- 551.
H. P. Snippe and J. J. Koenderink (1992b). Information in chClnnel-coded system:
correlated receivers . Biol. Cybern. 67, 183- 190.
H. Yoon and H. Sompolinsky (1998). Population coding in neuronal systems with
correlated noise, preprint.
E. Zohary, M. N. Shadlen, and W. T. Newsome (1994). Correlated neuronal
discharge rate and its implications for psychophysical performance. Nature 370,
140- 143.
| 1510 |@word crucially:1 covariance:2 reduction:2 contains:1 united:1 tuned:2 must:1 numerical:1 shape:1 motor:5 fund:1 discrimination:1 half:2 plane:1 inspection:1 smith:3 short:1 height:1 manner:1 pairwise:2 presumed:1 expected:1 behavior:5 frequently:1 little:1 zohary:2 increasing:2 becomes:5 moreover:1 israel:3 substantially:1 ti:1 grant:1 positive:16 limit:7 acad:1 analyzing:1 fluctuation:1 might:1 studied:1 co:3 range:6 statistically:1 bi:1 bla:4 definite:1 area:5 crossover:1 convenient:1 close:1 cybern:3 conventional:1 center:1 compensated:1 jerusalem:2 estimator:1 bsf:1 racah:1 population:37 plenum:1 discharge:2 enhanced:1 trend:1 observed:5 bottom:5 yoon:6 preprint:1 capture:1 sompolinsky:9 decrease:6 movement:1 substantial:5 seung:4 depend:1 efficiency:1 whose:1 larger:4 valued:1 gi:1 transform:1 eigenvalue:1 analytical:1 interaction:1 relevant:4 rapidly:1 fmax:4 academy:1 enhancement:1 perfect:1 depending:1 derive:1 ac:2 ij:2 odd:2 eq:11 strong:6 paradiso:2 implies:4 come:2 direction:1 concentrate:1 snippe:6 stochastic:1 larry:1 bin:1 strictly:1 correction:1 hold:1 around:1 considered:2 cramer:1 exp:3 great:1 nw:2 claim:2 vary:1 estimation:2 proc:1 iw:1 clearly:2 activates:1 gaussian:5 c5ij:1 always:4 varying:2 indicates:1 contrast:1 camp:1 helpful:1 dayan:5 entire:2 nand:1 orientation:3 spatial:1 special:1 equal:1 identical:1 represents:1 broad:3 jones:1 peaked:1 stimulus:6 intended:1 fiz:2 n1:1 freedom:3 investigate:1 circular:1 evaluation:1 extreme:5 nl:3 natl:1 implication:2 indexed:1 rao:1 gn:3 newsome:2 retains:1 lattice:1 uniform:4 peak:2 huji:2 lee:3 physic:1 decoding:1 pool:3 enhance:1 again:1 koenderink:6 leading:1 li:1 coding:8 coefficient:1 depends:3 later:1 apparently:1 decaying:2 capability:2 contribution:1 il:2 accuracy:5 variance:2 characteristic:3 likewise:2 ensemble:2 wisdom:1 weak:4 sebastian:1 synaptic:1 ed:1 improves:1 cj:1 amplitude:3 specify:1 response:2 furthermore:1 angular:4 correlation:46 hand:2 o:1 jll:1 mode:1 grows:2 oil:1 effect:12 usa:1 normalized:3 true:2 unbiased:1 hence:3 analytically:1 ll:1 width:4 prominent:1 bmin:2 demonstrate:1 pro:1 fj:1 recently:1 fi:1 rl:1 binational:1 exponentially:2 cerebral:1 volume:1 discussed:2 he:1 cessing:1 tuning:15 nexp:1 cortex:5 gj:3 gt:1 multivariate:2 claimed:1 subtraction:1 kruse:3 signal:2 ii:1 rv:3 rj:1 characterized:1 calculation:1 believed:1 cross:1 long:2 lin:1 bigger:1 coded:1 impact:1 basic:1 denominator:2 essentially:1 expectation:1 normalization:1 represent:1 cell:2 spacing:1 decreased:1 suppressive:1 rest:1 integer:2 call:1 enough:1 ledgments:1 affect:1 restrict:1 opposite:2 reduce:2 regarding:1 utility:2 peter:1 york:1 generally:1 detailed:2 amount:2 band:1 crosscorrelations:1 specifies:1 sign:3 broadly:2 redundancy:1 four:1 threshold:1 enormous:1 abbott:6 kept:1 utilize:1 fraction:5 angle:8 inverse:1 respond:1 throughout:2 reasonable:2 almost:1 vn:2 scaling:1 bound:4 haim:1 guaranteed:1 activity:5 strength:3 georgopoulos:3 ri:2 fourier:1 span:2 extremely:1 remain:1 smaller:3 lp:2 biologically:4 primate:1 explained:1 equation:1 remains:4 discus:1 appropriate:1 rp:1 top:3 running:1 calculating:1 exploit:1 psychophysical:1 primary:1 striate:1 detrimental:1 reversed:1 sci:1 capacity:7 argue:1 extent:1 assuming:1 length:5 code:9 modeled:1 hebrew:1 trace:2 stated:1 negative:6 acknow:1 upper:2 neuron:18 finite:7 parietal:1 immediate:1 situation:1 saturates:3 variability:6 pair:3 namely:1 israeli:1 beyond:1 below:2 regime:10 reading:1 saturation:3 arm:1 representing:1 scheme:2 improve:1 ne:2 acknowledges:1 text:1 understanding:1 relative:3 lacking:1 proportional:1 localized:1 foundation:1 degree:4 consistent:1 shadlen:2 port:3 uncorrelated:1 supported:1 keeping:2 soon:1 institute:1 fetz:3 neighbor:1 distributed:2 curve:13 dimension:1 cortical:2 valid:2 overcome:1 sensory:3 commonly:1 made:2 universally:1 far:1 preferred:5 receiver:1 conclude:1 assumed:4 un:2 nature:1 main:1 linearly:3 neurosci:1 whole:1 noise:15 profile:2 n2:1 contradicting:1 neuronal:11 fig:8 bij:4 r2:1 decay:1 intrinsic:1 incorporating:1 importance:1 magnitude:2 dissimilarity:1 columnar:1 subtract:1 smoothly:2 led:1 visual:3 partially:1 extracted:1 bioi:1 fisher:17 typical:1 uniformly:1 total:1 gij:1 indicating:1 formally:1 vhere:1 phenomenon:1 evaluate:1 biol:2 correlated:21 |
560 | 1,511 | Learning multi-class dynamics
A. Blake, B. North and M. Isard
Department of Engineering Science, University of Oxford, Oxford OXl 3P J, UK.
Web: http://www.robots.ox.ac.uk/ ... vdg/
Abstract
Standard techniques (eg. Yule-Walker) are available for learning
Auto-Regressive process models of simple, directly observable, dynamical processes. When sensor noise means that dynamics are
observed only approximately, learning can still been achieved via
Expectation-Maximisation (EM) together with Kalman Filtering.
However, this does not handle more complex dynamics, involving
multiple classes of motion. For that problem, we show here how
EM can be combined with the CONDENSATION algorithm, which
is based on propagation of random sample-sets. Experiments have
been performed with visually observed juggling, and plausible dynamical models are found to emerge from the learning process.
1
Introduction
The paper presents a probabilistic framework for estimation (perception) and classification of complex time-varying signals, represented as temporal streams of states.
Automated learning of dynamics is of crucial importance as practical models may
be too complex for parameters to be set by hand. The framework is particularly
general, in several respects, as follows.
1. Mixed states: each state comprises a continuous and a discrete component.
The continuous component can be thought of as representing the instantaneous
position of some object in a continuum. The discrete state represents the current
class of the motion, and acts as a label, selecting the current member from a set of
dynamical models.
2. Multi-dimensionality: the continuous component of a state is, in general,
allowed to be multi-dimensional. This could represent motion in a higher dimensional continuum, for example , two-dimensional translation as in figure 1. Other
examples include multi-spectral acoustic or image signals, or multi-channel sensors
such as an electro-encephalograph.
390
A. Blake. B. North and M Isard
Figure 1: Learning the dynamics of juggling. Three motion classes, emerging
from dynamical learning, turn out to correspond accurately to ballistic motion (mid
grey), catch/throw (light grey) and carry (dark grey).
3. Arbitrary order: each dynamical system is modelled as an Auto-Regressive
Process (ARP) and allowed to have arbitrary order (the number of time-steps of
"memory" that it carries.)
4. Stochastic observations: the sequence of mixed states is "hidden" - not
observable directly, but only via observations, which may be multi-dimensional,
and are stochastically related to the continuous component of states. This aspect is
essential to represent the inherent variability of response of any real signal sensing
system.
Estimation for processes with properties 2,3,4 has been widely discussed both in
the control-theory literature as "estimation" and "Kalman filtering" (Gelb, 1974)
and in statistics as ''forecasting'' (Brockwell and Davis, 1996). Learning of models
with properties 2,3 is well understood (Gelb, 1974) and once learned can be used
to drive pattern classification procedures, as in Linear Predictive Coding (LPC) in
speech analysis (Rabiner and Bing-Hwang, 1993), or in classification of EEG signals
(Pardey et al., 1995). When property 4 is added, the learning problem becomes
harder (Ljung, 1987) because the training sets are no longer observed directly.
Mixed states (property 1) allow for combining perception with classification. Allowing properties 2,4, but restricted to a Oth order ARP (in breach of property 3), gives
391
Learning Multi-Class Dynamics
Hidden Markov Models (HMM) (Rabiner and Bing-Hwang, 1993), which have been
used effectively for visual classification (Bregler , 1997). Learning HMMs is accomplished by the "Baum-Welch" algorithm, a form of Expectation-Maximisation (EM)
(Dempster et al., 1977). Baum-Welch learning has been extended to "graphicalmodels" of quite general topology (Lauritzen, 1996). In this paper, graph topology
is a simple chain-pair as in standard HMMs, and the complexity of the problem lies
elsewhere - in the generality of the dynamical model.
Generally then, restoring non-zero order to the ARPs (property 3), there is no exact
algorithm for estimation. However the estimation problem can be solved by random
sampling algorithms , known variously as bootstrap filters (Gordon et al., 1993),
particle filters (Kitagawa, 1996), and CONDENSATION (Blake and Isard, 1997). Here
we show how such algorithms can be used, with EM, in dynamical learning theory
and experiments (figure 1).
2
Multi-class dynamics
Continuous dynamical systems can be specified in terms of a continuous state vector
Xt E nNcr. In machine vision , for example, Xt represents the parameters of a timevarying shape at time t . Multi-class dynamics are represented by appending to the
continuous state vector Xt, a discrete state component Yt to make a "mixed" state
Xt = (
~:
) ,
where Yt E Y = {I, . .. , Ny} is the discrete component of the state, drawn from
a finite set of integer labels. Each discrete state represents a class of motion , for
example "stroke", "rest" and "shade" for a hand engaged in drawing.
Corresponding to each state Yt = Y there is a dynamical model, taken to be a
Markov model of order KY that specifies Pi (Xt IXt-l, . .. Xt-KY ) . A linear-Gaussian
Markov model of order K is an Auto-Regressive Process (ARP) defined by
K
Xt
= LAkxt-k
+ d + BWt
k=1
in which each Wt is a vector of N x independent random N(O, 1) variables and
are independent for t ? t'. The dynamical parameters of the model are
Wi,
W t'
? deterministic parameters AI, A 2 , ... , AK
? stochastic parameters B, which are multipliers for the stochastic process Wt,
and determine the "coupling" of noise Wt into the vector valued process Xt.
For convenience of notation, let
Each state Y E Y has a set {AY, BY, dY} of dynamical parameters, and the goal is
to learn these from example trajectories. Note that the stochastic parameter BY
is a first-class part of a dynamical model, representing the degree and the shape
of uncertainty in motion, allowing the representation of an entire distribution of
possible motions for each state y. In addition, and independently, state transitions
are governed by the transition matrix for a 1st order Markov chain:
P(Yt = y'IYt-1 = y) = My,y"
A. Blake. B. North and M. Isard.
392
Observations Zt are assumed to be conditioned purely on the continuous part x of
the mixed state, independent of Yt, and this maintains a healthy separation between
the modelling of dynamics and of observations. Observations are also assumed to
be independent, both mutually and with respect to the dynamical process. The
observation process is defined by specifying, at each time t, the conditional density
p(ZtIXt) which is taken to be Gaussian in experiments here.
3
Maximum Likelihood learning
When observations are exact, maximum likelihood estimates (MLE) for dynamical parameters can be obtained from a training sequence Xi ... X T of mixed states.
The well known Yule-Walker formula approximates MLE (Gelb, 1974; Ljung, 1987),
but generalisations are needed to allow for short training sets (small T), to include
stochastic parameters B, to allow a non-zero offset d (this proves essential in experiments later) and to encompass multiple dynamical classes.
The resulting MLE learning rule is as follows.
AY RY
= BY0'
dY =
where (omitting the
1
(R Y _ AYRY) CY
TY _ KY 0
,
Y
= TY _1 KY
(iW
""0,0
_ AY('QY)T)
.L"O'
superscripts for clarity) C = BBT and
and the first-order moments Ri and (offset-invariant) auto correlations Ri,j, for each
class y, are given by
Rf =
where
L
y;=y
x;_i
and
RL = RL - T ~ KRfRrT,
Y
RL = L X;_iX;_j T;
Ty
= Ht
:
Y; = y} == L
1.
t:Yt=Y
Yt=Y
The MLE for the transition matrix M is constructed from relative frequencies as:
h
T y,y' = ll{t?
*
M Y,Y' = "" Ty,y'T , were
II
. Yt-l
6y'EY
4
= y, Yt* = Y'} .
Y,Y
Learning with stochastic observations
To allow for stochastic observations, direct MLE is no longer possible, but an EM
learning algorithm can be formulated. Its M-step is simply the MLE estimate of
the previous section. It might be thought that the E-step should consist simply of
computing expectations, for instance [[xtIZ[J, (where Zi = (Zl,"" Zt) denotes a
sequence of observations) and treating them as training values x;. This would be
incorrect however because the log-likelihood function I:- for the problem is not linear
in the x; but quadratic. Instead, we need expectations
Learning Multi-Class Dynamics
393
conditioned on the entire training set Z'[ of observations, given that ? is linear
in the R i , Ri,j etc. (Shumway and Stoffer, 1982). These expected values of autocorrelations and frequencies are to be used in place of actual auto correlations and
frequencies in the learning formulae of section 3. The question is, how to compute
them. In the special case y = {I} of single-class dynamics, and assuming a Gaussian
observation density, exact methods are available for computing expected moments,
using Kalman and smoothing filters (Gelb, 1974), in an "augmented state" filter
(North and Blake, 1998). For multi-class dynamics, exact computation is infeasible, but good approximations can be achieved based on propagation of sample sets,
using CONDENSATION.
Forward sampling with backward chaining
For the purposes of learning, an extended and generalised form of the CONDENSATION algorithm is required. The generalisations allow for mixed states, arbitrary order for the ARP, and backward-chaining of samples. In backward chaining,
sample-sets for successive times are built up and stored together with a complete
state history back to time t = O. The extended CONDENSATION algorithm is given
in figure 2. Note that the algorithm needs to be initialised. This requires that the
Yo and (X~~lo' k = 0, ... ,KYO - 1) be drawn from a suitable (joint) prior for the
multi-class process. One way to do this is to ensure that the training set starts in a
known state and to fix the initial sample-values accordingly. Normally, the choice
of prior is not too important as it is dominated by data.
At time t = T, when the entire training sequence has been processed, the final
sample set is
(n)} ,n -- 1 , ... , N}
{ ( X(n)
TIT'? .. , X(n?)
OIT' 7rT
represents fairly (in the limit, weakly, as N -+ 00) the posterior distribution for
the entire state sequence X O , .?? ,XT, conditioned on the entire training set Z'[
of observations. The expectations of the autocorrelation and frequency measures
required for learning can be estimated from the sample set, for example:
An alternative algorithm is a sample-set version of forward-backward propagation
(Kitagawa, 1996). Experiments have suggested that probability densities generated
by this form of smoothing converge far more quickly with respect to sample set
size N, but at the expense of computational complexity - O(N2) as opposed to
O(N log N) for the algorithm above.
5
Practical applications
Experiments are reported briefly here on learning the dynami(:s of juggling using the
EM-Condensation algorithm, as in figure 1. An offset d Y is learned for each class
in Y = {I, 2, 3}; other dynamical parameters are fixed such that that learning d Y
amounts to learning mean accelerations a Y for each class. The transition matrix is
also learned. From a more or ?less neutral starting point, learned structure emerges
as in figure 3. Around 60 iterations of EM suffice, with N = 2048, to learn dynamics
in this case. It is clear from the figure that the learned structure is an altogether
plausible model for the juggling process.
394
A. Blake, B. North and M. Isard
Iterate for t = 1, ... , T.
(n?)
(n)}
.
Construct t he sampIe-set {(X (n)
l i t " ' " X tit ,7r t
,n = 1, ... , N for time
t.
For each n:
1. Choose (with replacement) mE {I, .. . , N} with prob. 7ri~{'
2. Predict by sampling from
P
(x
t
I
vt-l 1"\.1
-
(X(m)
llt-l"'"
X(m?))
t-llt-1
to choose X~~). For multi-class ARPs this is done in two steps.
Discrete: Choose y~n) = y' E Y with probability My,y" where
y = y~~i.
Continuous: Compute
K
(n) _ ~AY (m)
x tit - ~ kXt-klt-l
k=l
+d + Bw~n),
where y = y~n) and w~n) is a vector of standard normal r.v.
3. Observation weights 7r~n) are computed from the observation
density, evaluated for the current observations Zt:
(n?)
7rt(n) = P(I
Zt Xt = x tit '
then normalised multiplicatively so that
En 7ri n )
= 1.
4. Update sample history:
X ti(n)lt
-
x(m)
tilt-I'
I
t = 1, ... , t - 1.
Figure 2: The CONDENSATION algorithm for forward propagation with backward chaining.
Acknowledgements
We are grateful for the support of the EPSRC (AB,BN) and Magdalen College
Oxford (MI).
References
Blake, A. and Isard, M. (1997) . The Condensation algorithm - conditional density propagation and applications to visual tracking. In Advances in Neural Information Processing Systems 9, pages 361-368. MIT Press.
395
Learning Multi-Class Dynamics
(:0
0.01
a = ( 0.0 )
-9.7
0.04
Ballistic
pat:::)
~
Cony
a=(-;:) ~
Catchlthrow
J
Figure 3: Learned dynamical model for juggling. The three motion classes
allowed in this experiment organise themselves into: ballistic motion (acceleration
a ~ -g),- catch/throw,- carry. As expected, life-time in the ballistic state is longest,
the transition probability of 0.95 corresponding to 20 time-steps or about 0.7 seconds. Transitions tend to be directed, as expected,- for example ballistic motion is
more likely to be followed by a catch/throw (p = 0.04) than by a carry (p = 0.01).
(Acceleration a shown here in units of m/ S2 .)
Bregler, C. (1997). Learning and recognising human dynamics in video sequences. In Proc.
Conf. Computer Vision and Pattern Recognition.
Brockwell, P. and Davis, R. (1996). Introduction to time-series and forecasting. SpringerVerlag.
Dempster, A., Laird, M., and Rubin, D. (1977) . Maximum likelihood from incomplete
data via the EM algorithm. J. Roy. Stat. Soc. B ., 39:1-38.
Gelb, A., editor (1974). Applied Optimal Estimation. MIT Press, Cambridge, MA.
Gordon, N., Salmond, D., and Smith, A. (1993). Novel approach to nonlinear/nonGaussian Bayesian state estimation. lEE Proc. F, 140(2):107- 113.
Kitagawa, G. (1996). Monte Carlo filter and smoother for non-Gaussian nonlinear state
space models. Journal of Computational and Graphical Statistics, 5(1) :1- 25 .
Lauritzen, S. (1996). Graphical models. Oxford.
Ljung, L. (1987). System identification: theory for the user. Prentice-Hall.
North, B. and Blake, A. (1998).
Learning dynamical models using expectationmaximisation. In Proc. 6th Int. Conf. on Computer Vision, pages 384-389.
Par dey, J., Roberts, S., and Tarassenko, L. (1995). A review of parametric modelling
techniques for EEG analysis. Medical Engineering Physics, 18(1):2- 1l.
Rabiner, L. and Bing-Hwang, J. (1993) . Fundamentals of speech recognition. Prentice-Hall.
Shumway, R. and Stoffer, D. (1982) . An approach to time series smoothing and forecasting
USing the EM algorithm. J. Time Series Analysis, 3:253-226 .
| 1511 |@word briefly:1 version:1 grey:3 bn:1 harder:1 carry:4 moment:2 initial:1 series:3 selecting:1 current:3 shape:2 treating:1 update:1 isard:6 accordingly:1 smith:1 short:1 bwt:1 regressive:3 successive:1 constructed:1 direct:1 incorrect:1 autocorrelation:1 expected:4 themselves:1 multi:14 ry:1 actual:1 becomes:1 notation:1 suffice:1 emerging:1 temporal:1 act:1 ti:1 uk:2 control:1 zl:1 normally:1 unit:1 medical:1 generalised:1 engineering:2 understood:1 limit:1 ak:1 oxford:4 approximately:1 might:1 specifying:1 hmms:2 directed:1 practical:2 restoring:1 maximisation:2 bootstrap:1 procedure:1 thought:2 convenience:1 prentice:2 www:1 deterministic:1 yt:9 baum:2 starting:1 independently:1 welch:2 rule:1 handle:1 user:1 exact:4 roy:1 recognition:2 particularly:1 tarassenko:1 observed:3 epsrc:1 solved:1 cy:1 dempster:2 complexity:2 dynamic:15 weakly:1 grateful:1 tit:4 predictive:1 purely:1 joint:1 represented:2 monte:1 quite:1 widely:1 plausible:2 valued:1 drawing:1 statistic:2 laird:1 superscript:1 final:1 sequence:6 kxt:1 combining:1 brockwell:2 ky:4 object:1 coupling:1 ac:1 stat:1 lauritzen:2 throw:3 soc:1 filter:5 stochastic:7 human:1 fix:1 kitagawa:3 bregler:2 around:1 hall:2 blake:8 normal:1 visually:1 predict:1 continuum:2 purpose:1 estimation:7 proc:3 label:2 ballistic:5 iw:1 healthy:1 mit:2 sensor:2 gaussian:4 varying:1 timevarying:1 yo:1 klt:1 longest:1 modelling:2 likelihood:4 entire:5 hidden:2 classification:5 smoothing:3 special:1 fairly:1 once:1 construct:1 sampling:3 represents:4 gordon:2 inherent:1 variously:1 replacement:1 bw:1 ab:1 stoffer:2 light:1 oth:1 chain:2 incomplete:1 instance:1 neutral:1 too:2 ixt:1 stored:1 reported:1 my:2 combined:1 st:1 density:5 fundamental:1 probabilistic:1 lee:1 physic:1 together:2 pardey:1 quickly:1 nongaussian:1 opposed:1 choose:3 stochastically:1 conf:2 coding:1 north:6 int:1 stream:1 performed:1 later:1 start:1 maintains:1 correspond:1 rabiner:3 bbt:1 modelled:1 bayesian:1 identification:1 accurately:1 carlo:1 trajectory:1 drive:1 history:2 stroke:1 llt:2 ty:4 frequency:4 initialised:1 mi:1 emerges:1 dimensionality:1 back:1 higher:1 response:1 done:1 ox:1 evaluated:1 generality:1 dey:1 correlation:2 hand:2 web:1 nonlinear:2 propagation:5 autocorrelations:1 hwang:3 omitting:1 multiplier:1 eg:1 ll:1 davis:2 chaining:4 ay:4 complete:1 motion:11 image:1 instantaneous:1 novel:1 rl:3 tilt:1 discussed:1 he:1 approximates:1 cambridge:1 ai:1 particle:1 robot:1 longer:2 etc:1 posterior:1 vt:1 life:1 accomplished:1 ey:1 determine:1 converge:1 signal:4 ii:1 smoother:1 multiple:2 encompass:1 mle:6 involving:1 vision:3 expectation:5 iteration:1 represent:2 achieved:2 qy:1 condensation:8 addition:1 oxl:1 walker:2 crucial:1 rest:1 tend:1 electro:1 member:1 integer:1 automated:1 iterate:1 zi:1 topology:2 oit:1 forecasting:3 juggling:5 speech:2 generally:1 clear:1 amount:1 dark:1 mid:1 cony:1 processed:1 http:1 specifies:1 estimated:1 discrete:6 drawn:2 clarity:1 ht:1 backward:5 graph:1 prob:1 uncertainty:1 place:1 separation:1 dy:2 followed:1 quadratic:1 ri:5 dominated:1 aspect:1 department:1 em:9 wi:1 restricted:1 invariant:1 taken:2 mutually:1 bing:3 turn:1 needed:1 available:2 spectral:1 appending:1 alternative:1 altogether:1 denotes:1 include:2 ensure:1 graphical:2 prof:1 added:1 question:1 parametric:1 rt:2 hmm:1 me:1 assuming:1 kalman:3 multiplicatively:1 iyt:1 robert:1 expense:1 zt:4 allowing:2 observation:16 markov:4 finite:1 gelb:5 pat:1 extended:3 variability:1 arbitrary:3 pair:1 required:2 specified:1 acoustic:1 learned:6 suggested:1 dynamical:18 perception:2 pattern:2 lpc:1 rf:1 built:1 memory:1 video:1 suitable:1 representing:2 catch:3 auto:5 breach:1 prior:2 literature:1 acknowledgement:1 review:1 relative:1 shumway:2 par:1 ljung:3 mixed:7 filtering:2 degree:1 rubin:1 editor:1 pi:1 translation:1 lo:1 elsewhere:1 infeasible:1 allow:5 normalised:1 organise:1 salmond:1 emerge:1 transition:6 forward:3 far:1 observable:2 assumed:2 xi:1 continuous:9 expectationmaximisation:1 channel:1 learn:2 eeg:2 complex:3 s2:1 noise:2 n2:1 allowed:3 augmented:1 en:1 dynami:1 ny:1 position:1 comprises:1 lie:1 governed:1 ix:1 yule:2 formula:2 shade:1 xt:10 sensing:1 offset:3 essential:2 consist:1 recognising:1 effectively:1 importance:1 conditioned:3 lt:1 simply:2 likely:1 visual:2 tracking:1 ma:1 conditional:2 goal:1 formulated:1 acceleration:3 springerverlag:1 generalisation:2 wt:3 arp:6 engaged:1 college:1 support:1 |
561 | 1,512 | Replicator Equations, Maximal Cliques,
and Graph Isomorphism
Marcello Pelillo
Dipartimento di Informatica
Universita Ca' Foscari di Venezia
Via Torino 155, 30172 Venezia Mestre, Italy
E-mail: pelillo@dsi.unive.it
Abstract
We present a new energy-minimization framework for the graph
isomorphism problem which is based on an equivalent maximum
clique formulation. The approach is centered around a fundamental
result proved by Motzkin and Straus in the mid-1960s, and recently
expanded in various ways, which allows us to formulate the maximum clique problem in terms of a standard quadratic program. To
solve the program we use "replicator" equations, a class of simple
continuous- and discrete-time dynamical systems developed in various branches of theoretical biology. We show how, despite their
inability to escape from local solutions, they nevertheless provide
experimental results which are competitive with those obtained using more elaborate mean-field annealing heuristics.
1
INTRODUCTION
The graph isomorphism problem is one of those few combinatorial optimization
problems which still resist any computational complexity characterization [6]. Despite decades of active research, no polynomial-time algorithm for it has yet been
found. At the same time, while clearly belonging to N P, no proof has beel1 provided that it is NP-complete. Indeed, there is strong evidence that this cannot be
the case for, otherwise, the polynomial hierarchy would collapse [5]. The current
belief is that the problem lies strictly between the P and NP-complete classes.
Because of its theoretical as well as practical importance, the problem has attracted
much attention in the neural network community, and various powerful heuristics have been developed [11, 18, 19, 20]. Following Hopfield and Tank's seminal
work [10], the typical approach has been to write down a (continuous) energy function whose minimizers correspond to the (discrete) solutions being sought, and then
construct a dynamical system which converges toward them. Almost invariably, all
the algorithms developed so far are based on techniques borrowed from statistical
mechanics, in particular mean field theory, which allow one to escape from poor
Replicator Equations, Maximal Cliques, and Graph Isomorphism
551
local solutions.
In this paper, we develop a new energy-minimization framework for the graph isomorphism problem which is based on the idea of reducing it to the maximum clique
problem, another well-known combinatorial optimization problem. Central to our
approach is a powerful result originally proved by Motzkin and Straus [13], and
recently extended in various ways [3, 7, 16], which allows us to formulate the maximum clique problem in terms of an indefinite quadratic program. We then present
a class of straightforward continuous- and discrete-time dynamical systems known
in mathematical biology as replicator equations, and show how, thanks to their
dynamical properties, they naturally suggest themselves as a useful heuristic for
solving the proposed graph isomorphism program. The extensive experimental results presented show that, despite their simplicity and their inherent inability to
escape from local optima, replicator dynamics are nevertheless competitive with
more sophisticated deterministic annealing algorithms. The proposed formulation
seems therefore a promising framework within which powerful continuous-based
graph matching heuristics can be developed, and is in fact being employed for solving practical computer vision problems [17J. More details on the work presented
here can be found in [15J.
2
A QUADRATIC PROGRAM FOR GRAPH
ISOMORPHISM
2.1
GRAPH ISOMORPHISM AS CLIQUE SEARCH
Let G = (V, E) be an undirected graph, where V is the set of vertices and E ~ V x V
is the set of edges. The order of G is the number of its vertices, and its size is the
number of edges. Two vertices i,j E V are said to be adjacent if (i,j) E E. The
adjacency matrix of G is the n x n symmetric matrix A = (aij) defined as follows:
aij = 1 if (i,j) E E, aij = a otherwise.
Given two graphs G' = (V', E') and Gil = (V", E") having the same order and
size, an isomorphism between them is any bijection ? : V' -t V" such that
(i,j) E E' {:} (?(i),?(j)) E E", for all i,j E V'. Two graphs are said to be
isomorphic if there exists an isomorphism between them. The graph isomorphism
problem is therefore to decide whether two graphs are isomorphic and, in the affirmative, to find an isomorphism. Barrow and Burstall [IJ introduced the notion
of an association graph as a useful auxiliary graph structure for solving general
graphjsubgraph isomorphism problems. The association graph derived from G'
and Gil is the undirected graph G = (V, E), where V = V' X V" and
E = {((i, h), (j, k)) E V x V : i:f= j, h:f= k, and (i,j) E E' {:} (h, k) E E"} .
Given an arbitrary undirected graph G = (V, E), a subset of vertices C is called a
clique if all its vertices are mutually adjacent , i.e. , for all i,j E C we have (i,j) E E.
A clique is said to be maximal if it is not contained in any larger clique, and
maximum if it is the largest clique in the graph. The clique number, denoted by
w(G), is defined as the cardinality of the maximum clique.
The following result establishes an equivalence between the graph isomorphism
problem and the maximum clique problem (see [15J for proof).
Theorem 2.1 Let G' and Gil be two graphs of order n , and let G be the corresponding association graph . Then, G' and Gil are isomorphic if and only if w(G) = n. In
this case, any maximum clique of G induces an isomorphism between G' and Gil ,
and vice versa.
552
2.2
M. Pelillo
CONTINUOUS FORMULATION OF MAX-CLIQUE
Let G = (V, E) be an arbitrary undirected graph of order n, and let Sn denote the
standard simplex of lRn :
Sn={xElR n :
Xi~O
foralli=l. .. n, and
tXi=I}.
z== 1
Given a subset of vertices C of G, we will denote by XC its characteristic vector
which is the point in Sn defined as xI = 1/ICI if i E C, xi = 0 otherwise, where ICI
denotes the cardinality of C.
Now, consider the following quadratic function:
f(x)
=
x T Ax
(1)
where "T" denotes transposition. The Motzkin-Straus theorem [13] establishes a
remarkable connection between global (local) maximizers of fin Sn and maximum
(maximal) cliques of G. Specifically, it states that a subset of vertices C of a
graph G is a maximum clique if and only if its characteristic vector XC is a global
maximizer of the function f in Sn. A similiar relationship holds between (strict)
local maximizers and maximal cliques [7, 16].
One drawback associated with the original Motzkin-Straus formulation relates to
the existence of spurious solutions, i.e., maximizers of f which are not in the form
of characteristic vectors [16]. In principle, spurious solutions represent a problem
since, while providing information about the order of the maximum clique, do not
allow us to extract the vertices comprising the clique. Fortunately, there is straightforward solution to this problem which has recently been introduced and studied
by Bomze [3]. Consider the following regularized version of function f:
j (x) = x T Ax + ~ X T X
(2)
.
The following is the spurious-free counterpart of the original Motzkin-Straus theorem (see [3] for proof).
Theorem 2.2 Let C be a subset of vertices of a graph G, and let
XC
be its charac-
teristic vector. Then the following statements hold:
(a) C is a maximum clique of G if and only if XC is a global maximizer of
the simplex Sn. Its order is then given by ICI = 1/2(1 - f(x C ) ) .
(b) C is a maximal clique of G if and only if XC is a local maximizer of
(c) All local (and hence global) maximizers of
j
j
j
over
in Sn.
over Sn are strict.
Unlike the Motzkin-Straus formulation, the previous result guarantees that all maximizers of j on Sn are strict, and are characteristic vectors of maximal/maximum
cliques in the graph. In an exact sense, therefore, a one-to-one correspondence exists between maximal cliques and local maximizers of j in Sn on the one hand, and
maximum cliques and global maximizers on the other hand.
2.3
A QUADRATIC PROGRAM FOR GRAPH ISOMORPHISM
Let G' and Gil be two arbitrary graphs of order n, and let A denote the adjacency
matrix of the corresponding association graph, whose order is assumed to be N.
The graph isomorphism problem is equivalent to the following program:
maXImIze
subject to
j(x) = x T (A
x E SN
+ ~ IN)X
(3)
553
Replicator Equations. Maximal Cliques. and Graph Isomorphism
More precisely, the following result holds, which is a straightforward consequence
of Theorems 2.1 and 2.2.
Theorem 2.3 Let G' and Gil be two graphs of order n, and let x* be a global
solution of program (3), where A is the adjacency matrix of the association graph
of G' and Gil . Then, G' and Gil are isomorphic if and only if j(x*) = 1 - 1/2n.
In this case, any global solution to (3) induces an isomorphism between G' and Gil,
and vice versa.
In [15] we discuss the analogies between our objective function and those proposed
in the literature (e.g., [18, 19]).
3
REPLICATOR EQUATIONS AND GRAPH
ISOMORPHISM
Let W be a non-negative n x n matrix, and consider the following dynamical system:
~Xi(t) = Xi(t) ("i(t) - t.X;(t)";(t))
where 7ri(t)
,
i
= 1. . . n
(4)
= 2:.7=1 WijXj(t), i = 1 . . . n , and its discrete-time counterpart:
Xi(t)7ri (t)
xi(t+1)=2:. nj = l x] ()
t 7r]. (t ) '
i = l .. . n.
(5)
It is readily seen that the simplex Sn is invariant under these dynamics, which
means that every trajectory starting in Sn will remain in Sn for all future times.
Both (4) and (5) are called replicator equations in theoretical biology, since they
are used to model evolution over time of relative frequencies of interacting, selfreplicating entities [9]. The discrete-time dynamical equations turn also out to be
a special case of a general class of dynamical systems introduced by Baum and
Eagon [2] in the context of Markov chain theory.
Theorem 3.1 If W is symmetric, then the quadratic polynomial F(x) = xTWx is
strictly increasing along any non-constant trajectory of both continuous-time (4) and
discrete-time (5) replicator equations. Furthermore, any such trajectory converges
to a (unique) stationary point. Finally, a vector x E Sn is asymptotically stable
under (4) and (5) if and only if x is a strict local maximizer of F on Sn.
The previous result is known in mathematical biology as the Fundamental Theorem
of Natural Selection [9, 21]. As far as the discrete-time model is concerned, it
can be regarded as a straightforward implication of the more general Baum-Eagon
theorem [2]. The fact that all trajectories of the replicator dynamics converge to a
stationary point is proven in [12].
Recently, there has been much interest in evolutionary game theory around the
following exponential version of replicator equations , which arises as a model of
evolution guided by imitation [8, 21]:
:t Xi (t)
= Xi(t)
(L:7~1 ~:;;;~ ..
'(t) -
1),
i
= l... n
(6)
where K, is a positive constant. As K, tends to 0, the orbits of this dynamics approach
those of the standard, first-order replicator model (4), slowed down by the factor
554
M Pelillo
K. Hofbauer [8] has recently proven that when the matrix W is symmetric, the
quadratic polynomial F defined in Theorem 3.1 is also strictly increasing, as in
the first-order case. After discussing various properties of this, and more general
dynamics, he concluded that the model behaves essentially in the same way as the
standard replicator equations, the only difference being the size of the basins of
attraction around stable equilibria. A customary way of discretizating equation (6)
is given by the following difference equations:
Xi(t
+ 1) =
xi(t)e"1l';(t)
L: n.
)=1
( )
X)? t e"1l'J
(t)'
i = l. .. n
(7)
which enjoys many of the properties of the first-order system (5), e.g., they have
the same set of equilibria.
The properties discussed above naturally suggest using replicator equations as a
useful heuristic for the graph isomorphism problem. Let G' and G" be two graphs
of order n, and let A denote the adjacency matrix of the corresponding N-vertex
association graph G. By letting
1
W = A + "2IN
we know that the replicator dynamical systems, starting from an arbitrary initial
state, will iteratively maximize the function j(x) = xT(A + !IN)x in SN, and will
eventually converge to a strict local maximizer which, by virtue of Theorem 2.2 will
then correspond to the characteristic vector of a maximal clique in the association
graph. This will in turn induce an isomorphism between two subgraphs of G' and
G" which is "maximal," in the sense that there is no other isomorphism between
subgraphs of G' and G" which includes the one found. Clearly, in theory there is no
guarantee that the converged solution will be a global maximizer of j, and therefore
that it will induce an isomorphism between the two original graphs . Previous work
done on the maximum clique problem [4, 14], and also the results presented in this
paper, however, suggest that the basins of attraction of global maximizers are quite
large, and very frequently the algorithm converges to one of them.
4
EXPERIMENTAL RESULTS
In the experiments reported here, the discrete-time replicator equation (5) and its
exponential counterpart (7) with K = 10 were used. The algorithms were started
from the barycenter of the simplex and they were stopped when either a maximal
clique was found or the distance between two successive points was smaller than a
fixed threshold, which was set to 10- 17 . In the latter case the converged vector was
randomly perturbed, and the algorithm restarted from the perturbed point . Because
of the one-to-one correspondence between local maximizers and maximal cliques,
this situation corresponds to convergence to a saddle point. All the experiments
were run on a Sparc20.
Undirected 100-vertex random graphs were generated with expected connectivities
ranging from 1% to 99%. For each connectivity value, 10'0 graphs were produced and
each of them had its vertices randomly permuted so as to obtain a pair of isomorphic
graphs. Overall, therefore, 1500 pairs of isomorphic graphs were used. Each pair
was given as input to the replicator models and, after convergence, a success was
recorded when the cardinality of the returned clique was equal to the order of the
graphs given as input (Le., 100) .1 Because of the stopping criterion employed, this
1 Due to the high computational time required, in the 1% and 99% cases the first-order
replicator algorithm (5) was tested only on 10 pairs, instead of 100.
Replicator Equations, Maximal Cliques, and Graph Isomorphism
f
.i
'00
~,
/
- -
--- .
~--.--
-- .
. -
I ---.--.. -. .-. -. . . . ?
\
- - -
\
\
75 1
t
.i
c
50
\
I
U
!
555
j/
I
25 I
'
I
i
iii
001 003 0 05
0'
0.2
0.3
0 "
as
06
aa
0.7
09
0 95 0 97 099
001 003 0 .05
01
0 .2
Expecled connectivity
03
0 "
as
06
07
08
0.9
095
a 97
0 99
Expected connecllvlty
Figure 1: Percentage of correct isomorphisms obtained using the first-order (left) and the
exponential (right) replicator equations, as a function of the expected connectivity.
100000
-- - -
(?I
?
-
-
.~~om<l~~
;;-
10000
(?21~IIIK)
!!!.
!
-
~1U7 (1)
1000
1?I""'17)
(:t294KfI)
(:t201 0)
!
I
&
~
'-'
(N ~ Ill)
'00
\~t!%)
'0
' ,-
...
02
03
t
..e
(:t226)
----.----(1<)MI
(tI07)
001 0030 05 0.1
.
I
'1<)58)
a
5
,00
'0
(:to 94)
(i069)
0"
1000
06
07
Expected connectivity
08
09 095 0 , 9 7
a
99
001 003 0 .05
a
1
0 2
03
0 "
0.5
06
0.7
08
0.9 095 097 099
Expected connectivity
Figure 2: Average computational time taken by the first-order (left) and the exponential
(right) replicator equations, as a function of the expected connectivity. The vertical axes
are in logarithmic scale, and the numbers in parentheses represent the standard deviation.
guarantees that a maximum clique, and therefore a correct isomorphism, was found.
The proportion of successes as a function of the expected connectivities for both
replicator models is plotted in Fig. 1, whereas Fig. 2 shows the average CPU time
taken by the two algorithms to converge (in logarithmic scale). Notice how the
exponential replicator system (7) is dramatically faster and also performs bet.ter
than the first-order model (5).
These results are significantly superior to those reported by Simic [20] who obtained
poor results at connectivities less than 40% even on smaller graphs (Le. , up to 75
vertices). They also compare favorably with the results obtained more recently
by Rangarajan et ai. [18] on 100-vertex random graphs for connectivities up to
50%. Specifically, at 1% and 3% connectivities they report a percentage of correct
isomorphisms of about 30% and 0%, respectively. Using our approach we obtained,
on the same kind of graphs, a percentage of success of 80% and 11%, respectively.
Rangarajan and Mjolsness [19] also ran experiments on 100-vertex random graphs
with various connectivities, using a powerful Lagrangian relaxation network. Except
for a few instances , they always obtained a correct solution. The computational
time required by their model, however, turns out to largely exceed ours. As an
example, the average time taken by their algorithm to match two 100-vertex 50%connectivity graphs was about 30 minutes on an SGI workstation. As shown in
Fig. 2, we obtained identical results in about 3 seconds.
It should be emphasized that all the algorithms mentioned above do incorporate
sophisticated annealing mechanisms to escape from poor local minima. By contrast, in the presented work no attempt was made to prevent the algorithms from
converging to such solutions.
556
M Pelillo
Acknowledgments. This work has been done while the author was visiting the Department of Computer Science at the Yale University. Funding for this research has been
provided by the Consiglio Nazionale delle Ricerche , Italy. The author would like to thank 1.
M. Bomze, A. Rangarajan, K. Siddiqi , and S. W. Zucker for many stimulating discussions.
References
[1) H. G . Barrow and R. M. Burstall, "Subgraph isomorphism, matching relational structures and maximal cliques," Inform. Process. Lett., vol. 4, no. 4, pp. 83- 84, 1976.
[2) L. E. Baum and J. A. Eagon, "An inequality with applications to statistical estimation
for probabilistic functions of Markov processes and to a model for ecology," Bull.
Amer. Math. Soc., vol. 73, pp. 360- 363, 1967.
[3) 1. M. Bomze, "Evolution towards the maximum clique," J. Global Optim., vol. 10,
pp. 143- 164, 1997.
[4) I. M. Bomze, M. Pelillo, and R. Giacomini, "Evolutionary approach to the maximum
clique problem: Empirical evidence on a larger scale," in Developments in Global
Optimization, I. M. Bomze et al., eds., Kluwer , The Netherlands, 1997, pp. 95- 108.
[5) R. B. Boppana, J. Hastad, and S. Zachos, "Does co-NP have short interactive proofs?"
Inform . Process. Lett., vol. 25, pp. 127-132, 1987.
[6) M. R. Garey and D . S. Johnson, Computers and Intractability: A Guide to the Theory
of NP-Completeness. Freeman, San Francisco , CA , 1979.
[7) L. E. Gibbons, D. W. Hearn, P. M. Pardalos, and M. V. Ramana, "Continuous
characterizations of the maximum clique problem ," Math . Oper. Res., vol. 22 , no. 3,
pp. 754- 768, 1997.
[8) J. Hofbauer, "Imitation dynamics for games," Collegium Budapest, preprint, 1995.
[9) J . Hofbauer and K. Sigmund, The Theory of Evolution and Dynamical Systems. Cambridge University Press, Cambridge, UK, 1988.
[10) J . J . Hopfield and D. W. Tank, "Neural computation of decisions in optimization
problems," Biol. Cybern ., vol. 52, pp. 141- 152, 1985.
[11) R. Kree and A. Zippelius, "Recognition of topological features of graphs and images
in neural networks ," J. Phys. A : Math . Gen., vol. 21 , pp . L813- L818 , 1988.
[12) V. Losert and E . Akin, "Dynamics of games and genes: Discrete versuS continuous
time," J . Math. Biol., vol. 17, pp . 241- 251 , 1983.
[13) T . S. Motzkin and E . G. Straus, "Maxima for graphs and a new proof of a theorem
of'I\min," Canad. J. Math., vol. 17, pp . 533- 540, 1965 .
[14) M. Pelillo, "Relaxation labeling networks for the maximum clique problem," J. Artij.
Neural Networks, vol. 2, no. 4, pp. 313- 328, 1995.
[15) M. Pelillo, "Replicator equations, maximal cliques, and graph isomorphism," Neural
Computation, to appear.
[16) M. Pelillo and A. Jagota, "Feasible and infeasible maxima in a quadratic program
for maximum clique," J. Artij. Neural N etworks, vol. 2, no. 4, pp . 411- 420, 1995.
[17) M. Pelillo, K. Siddiqi , and S. W Zucker, "Matching hierarchical structures using
association graphs," in Computer Vision - ECCV '98, Vol. II, H. Burkhardt and B.
Neumann, eds. , Springer-Verlag, Berlin, 1998, pp. 3- 16.
[18) A. Rangarajan , S. Gold, and E . Mjolsness, "A novel optimizing network architecture
with applications," Neural Computation, vol. 8, pp. 1041- 1060, 1996.
[19) A. Rangarajan and E. Mjolsness, "A Lagrangian relaxation network for graph matching," IEEE Trans. Neural Networks, vol. 7, no. 6, pp. 1365- 1381, 1996.
[20) P. D . Simir, "Constrained nets for graph matching and other quadratic assignment
problems," Neural Computation, vol. 3, pp. 268-281, 1991.
[21) J. W . Weibull , Evolutionary Game Th eory. MIT Press, Cambridge, MA, 1995.
| 1512 |@word version:2 polynomial:4 seems:1 proportion:1 initial:1 ours:1 hearn:1 current:1 optim:1 yet:1 attracted:1 readily:1 stationary:2 short:1 transposition:1 completeness:1 characterization:2 math:5 bijection:1 successive:1 mathematical:2 along:1 expected:7 indeed:1 themselves:1 frequently:1 mechanic:1 freeman:1 cpu:1 cardinality:3 increasing:2 provided:2 kind:1 affirmative:1 weibull:1 developed:4 nj:1 guarantee:3 zippelius:1 bomze:5 every:1 interactive:1 uk:1 appear:1 unive:1 positive:1 local:12 tends:1 consequence:1 despite:3 studied:1 equivalence:1 co:1 collapse:1 practical:2 unique:1 acknowledgment:1 foralli:1 empirical:1 significantly:1 matching:5 induce:2 suggest:3 cannot:1 selection:1 context:1 cybern:1 seminal:1 equivalent:2 deterministic:1 lagrangian:2 baum:3 straightforward:4 attention:1 starting:2 formulate:2 simplicity:1 subgraphs:2 attraction:2 regarded:1 notion:1 hierarchy:1 exact:1 recognition:1 preprint:1 mjolsness:3 ran:1 mentioned:1 nazionale:1 complexity:1 gibbon:1 dynamic:7 solving:3 hopfield:2 various:6 venezia:2 labeling:1 whose:2 heuristic:5 larger:2 solve:1 quite:1 otherwise:3 net:1 maximal:16 budapest:1 gen:1 subgraph:1 gold:1 sgi:1 convergence:2 optimum:1 rangarajan:5 neumann:1 converges:3 develop:1 ij:1 pelillo:10 borrowed:1 strong:1 soc:1 auxiliary:1 guided:1 drawback:1 correct:4 centered:1 pardalos:1 adjacency:4 dipartimento:1 strictly:3 hold:3 around:3 equilibrium:2 sought:1 estimation:1 combinatorial:2 largest:1 vice:2 establishes:2 minimization:2 mit:1 clearly:2 always:1 bet:1 derived:1 ax:3 contrast:1 sense:2 minimizers:1 stopping:1 spurious:3 comprising:1 tank:2 overall:1 ill:1 denoted:1 development:1 constrained:1 special:1 field:2 construct:1 equal:1 having:1 biology:4 identical:1 marcello:1 future:1 simplex:4 np:4 report:1 escape:4 few:2 inherent:1 randomly:2 attempt:1 ecology:1 invariably:1 interest:1 lrn:1 chain:1 implication:1 edge:2 orbit:1 plotted:1 re:1 theoretical:3 stopped:1 instance:1 hastad:1 delle:1 assignment:1 bull:1 vertex:16 subset:4 deviation:1 johnson:1 reported:2 perturbed:2 thanks:1 fundamental:2 probabilistic:1 mestre:1 connectivity:13 central:1 recorded:1 oper:1 includes:1 sparc20:1 competitive:2 om:1 characteristic:5 who:1 largely:1 correspond:2 produced:1 trajectory:4 converged:2 inform:2 phys:1 ed:2 energy:3 frequency:1 pp:16 straus:7 garey:1 naturally:2 proof:5 di:2 associated:1 mi:1 workstation:1 proved:2 sophisticated:2 originally:1 formulation:5 done:2 amer:1 furthermore:1 hand:2 maximizer:6 counterpart:3 evolution:4 hence:1 symmetric:3 iteratively:1 adjacent:2 game:4 simic:1 criterion:1 complete:2 txi:1 performs:1 ranging:1 image:1 sigmund:1 recently:6 funding:1 novel:1 superior:1 replicator:24 burstall:2 behaves:1 permuted:1 association:8 he:1 discussed:1 kluwer:1 versa:2 cambridge:3 ai:1 u7:1 had:1 stable:2 zucker:2 italy:2 optimizing:1 verlag:1 inequality:1 success:3 discussing:1 jagota:1 seen:1 minimum:1 fortunately:1 wijxj:1 employed:2 ricerche:1 converge:3 maximize:2 ii:1 branch:1 relates:1 match:1 faster:1 parenthesis:1 converging:1 vision:2 essentially:1 represent:2 whereas:1 annealing:3 concluded:1 unlike:1 strict:5 subject:1 undirected:5 ter:1 exceed:1 iii:1 concerned:1 architecture:1 idea:1 whether:1 isomorphism:30 akin:1 returned:1 dramatically:1 useful:3 burkhardt:1 netherlands:1 mid:1 induces:2 siddiqi:2 informatica:1 eory:1 percentage:3 notice:1 gil:10 discrete:9 write:1 vol:15 indefinite:1 nevertheless:2 threshold:1 prevent:1 graph:57 asymptotically:1 relaxation:3 run:1 powerful:4 almost:1 decide:1 decision:1 collegium:1 correspondence:2 yale:1 quadratic:9 topological:1 foscari:1 precisely:1 ri:2 min:1 expanded:1 department:1 poor:3 belonging:1 remain:1 smaller:2 slowed:1 invariant:1 taken:3 equation:19 mutually:1 etworks:1 discus:1 turn:3 eventually:1 mechanism:1 know:1 letting:1 ramana:1 hierarchical:1 eagon:3 customary:1 existence:1 original:3 denotes:2 xc:5 hofbauer:3 universita:1 objective:1 torino:1 barycenter:1 canad:1 said:3 evolutionary:3 visiting:1 distance:1 thank:1 berlin:1 entity:1 mail:1 toward:1 relationship:1 providing:1 charac:1 statement:1 favorably:1 negative:1 vertical:1 markov:2 fin:1 similiar:1 barrow:2 situation:1 extended:1 relational:1 interacting:1 arbitrary:4 community:1 introduced:3 pair:4 required:2 extensive:1 connection:1 resist:1 trans:1 dynamical:9 program:9 max:1 belief:1 natural:1 regularized:1 started:1 extract:1 sn:17 literature:1 relative:1 dsi:1 analogy:1 proven:2 remarkable:1 versus:1 basin:2 principle:1 intractability:1 eccv:1 free:1 infeasible:1 enjoys:1 aij:3 guide:1 allow:2 lett:2 author:2 made:1 ici:3 san:1 far:2 gene:1 clique:41 global:11 active:1 assumed:1 francisco:1 xi:11 imitation:2 continuous:8 search:1 decade:1 promising:1 ca:2 fig:3 elaborate:1 exponential:5 lie:1 down:2 theorem:12 minute:1 xt:1 emphasized:1 virtue:1 evidence:2 maximizers:9 exists:2 importance:1 logarithmic:2 saddle:1 motzkin:7 contained:1 restarted:1 springer:1 aa:1 corresponds:1 ma:1 stimulating:1 boppana:1 towards:1 feasible:1 typical:1 specifically:2 reducing:1 except:1 called:2 isomorphic:6 experimental:3 latter:1 inability:2 arises:1 incorporate:1 tested:1 biol:2 |
562 | 1,513 | Convergence Rates of Algorithms for
Visual Search: Detecting Visual Contours
A.L. Yuille
Smith-Kettlewell Inst .
San Francisco, CA 94115
James M. Coughlan
Smith-Kettlewell Inst.
San Francisco, CA 94115
Abstract
This paper formulates the problem of visual search as Bayesian
inference and defines a Bayesian ensemble of problem instances .
In particular, we address the problem of the detection of visual
contours in noise/clutter by optimizing a global criterion which
combines local intensity and geometry information. We analyze
the convergence rates of A * search algorithms using results from
information theory to bound the probability of rare events within
the Bayesian ensemble. This analysis determines characteristics of
the domain , which we call order parameters, that determine the
convergence rates. In particular, we present a specific admissible
A * algorithm with pruning which converges, with high probability,
with expected time O(N) in the size of the problem. In addition, we briefly summarize extensions of this work which address
fundamental limits of target contour detectability (Le. algorithm
independent results) and the use of non-admissible heuristics.
1
Introduction
Many problems in vision, such as the detection of edges and object boundaries in
noise/clutter, see figure (1), require the use of search algorithms . Though many
algorithms have been proposed, see Yuille and Coughlan (1997) for a review, none
of them are clearly optimal and it is difficult to judge their relative effectiveness. One
approach has been to compare the results of algorithms on a representative dataset
of images. This is clearly highly desirable though determining a representative
dataset is often rather subjective.
In this paper we are specifically interested in the convergence rates of A * algorithms
(Pearl 1984). It can be shown (Yuille and Coughlan 1997) that many algorithms
proposed to detect visual contours are special cases of A * . We would like to
understand what characteristics of the problem domain determine the convergence
642
A. L. Yuille and J. M Coughlan
Figure 1: The difficulty of detecting the target path in clutter depends, by our
theory (Yuille and Coughlan 1998), on the order parameter K. The larger K the
less computation required. Left, an easy detection task with K = 3.1. Middle, a
hard detection task K = 1.6. Right, an impossible task with K = -0.7.
rates.
We formulate the problem of detecting object curves in images to be one of statistical
estimation. This assumes statistical knowledge of the images and the curves, see
section (2). Such statistical knowledge has often been used in computer vision for
determining optimization criteria to be minimized. We want to go one step further
and use this statistical knowledge to determine good search strategies by defining
a Bayesian ensemble of problem instances. For this ensemble, we can prove certain
curve and boundary detection algorithms, with high probability, achieve expected
time convergence in time linear with the size of the problem. Our analysis helps
determine important characteristics of the problem, which we call order parameters,
which quantify the difficulty of the problem.
The next section (2) of this paper describes the basic statistical assumptions we
make about the domain and describes the mathematical tools used in the remaining
sections. In section (3) we specify our search algorithm and establish converEence
rates. We conclude by placing this work in a larger context and summarizing recent
extensions.
2
Statistical Background
Our approach assumes that both the intensity properties and the geometrical shapes
of the target path (i.e. the edge contour) can be determined statistically. This path
can be considered .to be a set of elementary path segments joined together. We first
consider the intensity properties along the edge and then the geometric properties.
The set of all possible paths can be represented by a tree structure, see figure (2).
The image properties at segments lying on the path are assumed to differ, in a
statistical sense, from those off the path. More precisely, we can design a filter ?(.)
with output {Yx = ?(I(x))} for a segment at point x so that:
P(Yx) = Pon(Yx), if "XII lies on the true path
P(Yx) = Poff(Yx), if "X'I lies off the true path.
(1)
For example, we can think of the {Yx} as being values of the edge strength at point
x and Pon, Poll being the probability distributions of the response of ?(.) on and
off an edge. The set of possible values of the random variable Yx is the alphabet
with alphabet size M (Le. Yx can take any of M possible values). See (Geman and
Jedynak 1996) for examples of distributions for Pon, Pol I used in computer vision
applications.
We now consider the geometry of the target contour. We require the path to be
made up of connected segments Xl, X2, ... , x N. There will be a Markov probability
distribution Pg(Xi+I!Xi) which specifies prior probabilistic knowledge of the target.
Convergence Rates ofAlgorithmsfor Visual Search: Detecting Visual Contours
643
It is convenient, in terms of the graph search algorithms we will use, to consider that
each point x has a set of Q neighbours. Following terminology from graph theory,
we refer to Q as the branching factor. We will assume that the distribution P g
depends only on the relative positions of XHI and Xi. In other words, Pg(XHllxi) =
PLlg(XHl - Xi). An important special case is when the probability distribution
is uniform for all branches (Le. PLlg(Ax) = U(Ax) = I/Q, VAx). The joint
distribution P(X, Y) of the road geometry X and filter responses Y determines the
Bayesian Ensemble.
By standard Bayesian analysis, the optimal path X* = {xi, ... , XN} maximizes the
sum of the log posterior:
(2)
where the sum i is taken over all points on the target. U(Xi+l - Xi) is the uniform
distribution and its presence merely changes the log posterior E(X) by a constant
value. It is included to make the form of the intensity and geometric terms similar,
which simplifies our later analysis.
We will refer to E(X) as the reward of the path X which is the sum of the intensity
rewards log Pon (Y(~jl) and the geometric rewards log PL:>.g (Xi+l -Xi)
Poll (Y(~i?
U(Xi+l -Xi)
It is important to emphasize that our results can be extended to higher-order
Markov chain models (provided they are shift-invariant). We can, for example,
define the x variable to represent spatial orientation and position of a small edge
segment. This will allow our theory to apply to models, such as snakes, used in
recent successful vision applications (Geman and Jedynak 1996). (It is straightforward to transform the standard energy function formulation of snakes into a Markov
chain by discretizing and replacing the derivatives by differences. The smoothness
constraints, such as membranes and thin plate terms, will transform into first and
second order Markov chain connections respectively). Recent work by Zhu (1998)
shows that Markov chain models of this type can be learnt using Minimax Entropy
Learning theory from a representative set of examples. Indeed Zhu goes further by
demonstrating that other Gestalt grouping laws can be expressed in this framework
and learnt from representative data.
Most Bayesian vision theories have stopped at this point. The statistics of the problem domain are used only to determine the optimization criterion to be minimized
and are not exploited to analyze the complexity of algorithms for performing the optimization. In this paper, we go a stage further. We use the statistics ofthe problem
domain to define a Bayesian ensemble and hence to determine the effectiveness of
algorithms for optimizing criteria such as (2). To do this requires the use of Sanov's
theorem for calculating the probability of rare events (Cover and Thomas 1991).
For the road tracking problem this can be re-expressed as the following theorem,
derived in (Yuille and Coughlan 1998):
Theorem 1. The probabilities that the spatially averaged log-likelihoods on, and
off, the true curve are above, or below, threshold T are bounded above as follows:
Pr{.!.
n
t
i=l
< T} :s; (n + I)M2-nD(PTlfPon)
(3)
> T}:S; (n+ I)M2-nD(PTIIPOI/) ,
(4)
{log Pon(y(Xi?) }on
Poff (Y(Xi?)
Pr{.!. t{lOg Pon(Y(Xi?) }off
n i=l
POff(Y(Xi?)
A. L. Yuille and J. M. Coughlan
644
where
PT(y)
on the
mined
the subscripts on and off mean that the data is generated by Pon, Po",
= p;;;>'(T) (y)P;;p jZ(T) where a ::; "\(T) ::; 1 is a scalar which depends
threshold T and Z(T) is a normalization factor. The value of "\(T) is deterby the constraint 2: y PT (y) log ;'?In}(~) = T.
In the next section, we will use Theorem 1 to determine a criterion for pruning
the search based on comparing the intensity reward to a threshold T (pruning will
also be done using the geometric reward). The choice of T involves a trade-off. If
T is large (Le. close to D(PonllPoff)) then we will rapidly reject false paths but
we might also prune out the target (true) path. Conversely, if T is small (close
to -D(PoffllPon)) then it is unlikely we will prune out the target path but we
may waste a lot of time exploring false paths. In this paper we choose T large and
write the fall-off factors (Le. the exponents in the bounds of equations (3,4)) as
D(PTllPon) = tl (T), D(PTilPoff) = D(PonilPoff) - t2(T) where tl (T), t2(T) are
positive and (tl(T),t2(T)) t-+ (0,0) as T t-+ D(PonilPoff ). We perform a similar
analysis for the geometric rewards by substituting P6.g , U for Pon , Pol I' We choose
a threshold T satisfying -D(UIIP6.g) < T < D(P6.gllU). The results of Theorem
1 apply with the obvious substitutions. In particular, the alphabet factor becomes
Q (the branching factor). Once again, in this paper, we choose T to be large and
obtain fall-off factors D(Pt'IIP6.g) = El (T), D(Pt'IIU) = D(P6.gllU) - E2(T).
3
Tree Search: A *, heuristics, and block pruning
We now consider a specific example, motivated by Geman and Jedynak (1996),
of searching for a path through a search tree. In Geman and Jedynak the path
corresponds to a road in an aerial image and they assume that they are given an
initial point and direction on the target path. They have a branching factor Q = 3
and, in their first version, the prior probability of branching is considered to be the
uniform distribution (later they consider more sophisticated priors). They assume
that no path segments overlap which means that the search space is a tree of size
QN where N is the size of the problem (Le. the longest length). The size of the
problem requires an algorithm that converges in O(N) time and they demonstrate
an algorithm which empirically performs at this speed. But no proof of convergence
rates are given in their paper. It can be shown, see (Yuille and Coughlan 1997),
that the Geman and Jedynak algorithm is a close approximation to A * which uses
pruning. (Observe that Geman and Jedynak's tree representation is a simplifying
assumption of the Bayesian model which assumes that once a path diverges from
the true path it can never recover, although we stress that the algorithm is able to
recover from false starts - for more details see Coughlan and Yuille 1998).
We consider an algorithm which uses an admissible A * heuristic and a pruning
mechanism. The idea is to examine the paths chosen by the A * heuristic. As the
length of the candidate path reaches an integer multiple of No we prune it based on
its intensity reward and its geometric reward evaluated on the previous No segments,
which we call a segment block. The reasoning is that few false paths will survive
this pruning for long but the target path will survive with high probability.
We prune on the intensity by eliminating all paths whose intensity reward, averaged
over the last No segments, is below a threshold T (recall that -D(PoffllPon) < T <
D(PonllPoff) and we will usually select T to take values close to D(PonllPoff)).
In addition, we prune on the geometry by eliminating all paths whose geometric
rewards, averaged over the last No segments, are below T (where -D(UIIP6.g) <
T < D(P6.gllU) with T typically being close to D(P6.gllU)). More precisely, we
645
Convergence Rates ofAlgOrithms for Visual Search : Detecting Visual Contours
discard a path provided (for any integer z
1 (z~o I
Pon(Yi)
~ og
No i=zNo+l
Poff(yd
< T,
~
0):
1 (z+l)No
or No
L log
i=zNo+l
PLlg(Llxi)
U(Llx.)
< T.
(5)
t
There are two important issues to address: (i) With what probability will the
algorithm converge?, (ii) How long will we expect it take to converge? The next
two subsections put bounds on these issues.
3.1
Probability of Convergence
Because of the pruning, there is a chance that there will be no paths which survive
pruning. To put a bound on this we calculate the probability that the target
(true) path survives the pruning. This gives a lower bound on the probability of
convergence (because there could be false paths which survive even if the target
path is mistakenly pruned out).
The pruning rules removes path segments for which the intensity reward r I or the
geometric reward r 9 fails the pruning test. The probability of failure by removing
a block segment of the true path, with rewards r~, r~, is Pr(r~ < T or r~ < T) ::;
Pr(r~ < T) + Pr(r~ < T) ::; (No + 1)M2- NoE1 (T) + (No + 1)Q2-NoilCT), where we
have used Theorem 1 to put bounds on the probabilities. The probability of pruning
out any No segments of the true path can therefore be made arbitrarily small by
choosing No, T, T so as to make Notl and NOtl large.
It should be emphasized that the algorithm will not necessarily converge to the
exact target path. The admissible nature of the heuristic means that the algorithm
will converge to the path with highest reward which has survived the pruning. It
is highly probable that this path is close to the target path. Our recent results
(Coughlan and Yuille 1998, Yuille and Coughlan 1998) enable us to quantify this
claim.
3.2
Bounding the Number of False Paths
Suppose we face a Q-nary tree. We can order the false paths by the stage at which
they diverge from the target (true) path, see figure (2). For example, at the first
branch point the target path lies on only one of the Q branches and there are Q - 1
false branches which generate the first set of false paths Fl' Now consider all the
Q -1 false branches at the second target branch, these generate set F 2 . As we follow
along the true path we keep generating these false sets F i . The set of all paths is
therefore the target path plus the union of the Fi (i = 1, ... , N). To determine
convergence rates we must bound the amount of time we spend searching the Fi. If
the expected time to search each Fi is constant then searching for the target path
will at most take constant? N steps.
Consider the set Fi of false paths which leave the true path at stage i. We will apply
our analysis to block segments of Fi which are completely off the true path. If (i -1)
is an integer multiple of No then all block segments of Fi will satisfy this condition.
Otherwise, we will start our analysis at the next block and make the worse case
assumption that all path segments up till this next block will be searched. Since
the distance to the next block is at most No - 1, this gives a maximum number of
QNo-l starting blocks for any branch of F i . Each Fi also has Q - 1 branches and
so this gives a generous upper bound of (Q - l)Q N o-l starting blocks for each Fi .
A. L. Yuille and J. M. Coughlan
646
Figure 2: The target path is shown as the heavy line. The false path sets are
labelled as Fl ,F2 , etc. with the numbering depending on how soon they leave the
target path. The branching factor Q = 3.
For each starting block, we wish to compute (or bound) the expected number of
blocks that are explored thereafter. This requires computing the fertility of a block,
the average number of paths in the block that survive pruning. Provided the fertility
is smaller than one, we can then apply results from the theory of branching processes
to determine the expected number of blocks searched in F i .
The fertility q is the number of paths that survive the geometric pruning times the
probability that each survives the intensity pruning. This can be bounded (using
Theorem 1) by q q where:
:s
q=
QN0(No
+ I)Q2-No{D(hgIIU)-?2(T)}(No + I)M2-No{D(PonIIPoff)-E2(T)}
= (No
+ I)Q+M 2- No {D(Pon IIPof! )-H(Pa g)-E2(T)-?2(T)},
(6)
where we used the fact that D(PLlgIIU) = 10gQ - H(PLlg).
Observe that the condition q < 1 can be satisfied provided D(PonllPolf )-H(PLlg) >
O. This condition is intuitive, it requires that the edge detector information, quantified by D(PonIIPolf )' must be greater than the uncertainty in the geometry measured by H(PLlg). In other words, the better the edge detector and the more
predictable the path geometry then the smaller q will be.
We now apply the theory of branching processes to determine the expected number
of blocks explored from a starting block in Fi 'L~o qZ = 1/(1 - q). The number
of branches of Fi is (Q - 1), the total number of segments explored per block is at
most QNo , and we explore at most QNo-l segments before reaching the first block.
The total number of Fi is N. Therefore the total number of segments wastefully
explored is at most N(Q - 1) 1~qQ2No-1. We summarize this result in a theorem:
Theorem 2. Provided q = (No + I)Q+M2- N oK < 1, where the order parameter
K = D(PonllPolf) - H(PLlg) - ?2(T) - ?2(T), then the expected number of false
segments explored is at most N(Q - 1) 1~qQ2No-1.
Comment The requirement that q < 1 is chiefly determined by the order parameter
K = D (Pon IlPolf ) - H (PLlg) - ?2 (T) - f2 (T). Our convergence proofrequires that
K > 0 and will break down if K < O. Is this a limitation of our proof? Or does it
correspond to a fundamental difficulty in solving this tracking problem?
In more recent work (Yuille and Coughlan 1998) we extend the concept of order
parameters and show that they characterize the difficulty of visual search problem
independently of the algorithm. In other words, as K 1----7 0 the problem becomes
impossible to solve by any algorithm. There will be too many false paths which
have better rewards than the target path. As K 1----7 0 there is a phase transition in
the ease of solving the problem.
Convergence Rates ofAlgorithmsfor Visual Search: Detecting Visual Contours
4
647
Conclusion
Our analysis shows it is possible to detect certain types of image contours in linear
expected time (with given starting points). We have shown how the convergence
rates depend on order parameters which characterize the problem domain. In particular, the entropy of the geometric prior and the Kullback-Leibler distance between Pon and Pof f allow us to quantify intuitions about the power of geometrical
assumptions and edge detectors to solve these tasks.
Our more recent work (Yuille and Coughlan 1998) has extended this work by showing that the order parameters can be used to specify the intrinsic (algorithm independent) difficulty of the search problem and that phase transitions occur when
these order parameters take critical values. In addition, we have proved convergence rates for A * algorithms which use inadmissible heuristics or combinations of
heuristics and pruning (Coughlan and Yuille 1998).
As shown in (Yuille and Coughlan 1997) many of the search algorithms proposed
to solve vision search problems, such as (Geman and Jedynak 1996), are special
cases of A * (or close approximations). We therefore hope that the results of this
paper will throw light on the success of the algorithms and may suggest practical
improvements and speed ups.
Acknow ledgements
We want to acknowledge funding from NSF with award number IRI-9700446, from
the Center for Imaging Sciences funded by ARO DAAH049510494, and from an
ASOSRF contract 49620-98-1-0197 to ALY. We would like to thank L. Xu, D.
Snow, S. Konishi, D. Geiger, J. Malik, and D. Forsyth for helpful discussions.
References
[1] J .M. Coughlan and A.L. Yuille. "Bayesian A * Tree Search with Expected O(N)
Convergence Rates for Road Tracking." Submitted to Artificial Intelligence.
1998.
[2] T.M. Cover and J.A. Thomas. Elements of Information Theory. Wiley
Interscience Press. New York. 1991.
[3] D. Geman. and B. Jedynak. "An active testing model for tracking roads in
satellite images". IEEE Trans. Patt. Anal. and Machine Intel. Vol. 18. No.1,
pp 1-14. January. 1996.
[4] J. Pearl. Heuristics. Addison-Wesley. 1984.
[5] A.L. Yuille and J. Coughlan. " Twenty Questions, Focus of Attention, and A *" .
In Energy Minimization Methods in Computer Vision and Pattern
Recognition. Ed. M. Pellilo and E. Hancock. Springer-Verlag. (Lecture Notes
in Computer Science 1223). 1997.
[6] A.L. Yuille and J .M~ Coughlan. "Visual Search: Fundamental Bounds, Order
Parameters, Phase Transitions, and Convergence Rates." Submitted to Pattern
Analysis and Machine Intelligence. 1998.
[7] S.C. Zhu. "Embedding Gestalt Laws in Markov Random Fields". Submitted
to IEEE Computer Society Workshop on Perceptual Organization in Computer
Vision.
| 1513 |@word middle:1 version:1 briefly:1 eliminating:2 nd:2 simplifying:1 pg:2 initial:1 substitution:1 subjective:1 comparing:1 must:2 shape:1 remove:1 intelligence:2 smith:2 coughlan:19 detecting:6 mathematical:1 along:2 kettlewell:2 prove:1 combine:1 interscience:1 indeed:1 expected:9 examine:1 becomes:2 provided:5 pof:1 bounded:2 maximizes:1 what:2 q2:2 sanov:1 positive:1 before:1 local:1 limit:1 subscript:1 path:61 yd:1 might:1 plus:1 quantified:1 conversely:1 ease:1 statistically:1 averaged:3 jedynak:8 practical:1 testing:1 union:1 block:19 survived:1 reject:1 convenient:1 ups:1 word:3 road:5 suggest:1 close:7 put:3 context:1 impossible:2 pon:12 center:1 go:3 straightforward:1 starting:5 independently:1 iri:1 attention:1 formulate:1 iiu:1 m2:5 rule:1 konishi:1 searching:3 embedding:1 target:22 pt:4 suppose:1 exact:1 us:2 pa:1 element:1 satisfying:1 recognition:1 geman:8 calculate:1 connected:1 trade:1 highest:1 intuition:1 predictable:1 complexity:1 pol:2 reward:15 depend:1 solving:2 segment:20 yuille:19 f2:2 completely:1 po:1 joint:1 represented:1 alphabet:3 hancock:1 artificial:1 choosing:1 whose:2 heuristic:8 larger:2 spend:1 solve:3 otherwise:1 statistic:2 think:1 transform:2 aro:1 gq:1 rapidly:1 till:1 achieve:1 intuitive:1 convergence:18 requirement:1 diverges:1 satellite:1 generating:1 converges:2 leave:2 object:2 help:1 depending:1 measured:1 throw:1 involves:1 judge:1 quantify:3 differ:1 direction:1 snow:1 filter:2 enable:1 require:2 probable:1 elementary:1 extension:2 pl:1 exploring:1 lying:1 considered:2 claim:1 substituting:1 generous:1 estimation:1 tool:1 survives:2 hope:1 minimization:1 clearly:2 rather:1 reaching:1 og:1 ax:2 derived:1 focus:1 longest:1 improvement:1 likelihood:1 detect:2 summarizing:1 inst:2 inference:1 sense:1 helpful:1 el:1 snake:2 unlikely:1 typically:1 interested:1 issue:2 orientation:1 exponent:1 spatial:1 special:3 field:1 once:2 never:1 placing:1 survive:6 thin:1 minimized:2 t2:3 xhi:1 few:1 neighbour:1 geometry:6 phase:3 detection:5 organization:1 highly:2 light:1 chain:4 edge:9 tree:7 re:1 stopped:1 instance:2 cover:2 formulates:1 rare:2 uniform:3 successful:1 too:1 characterize:2 learnt:2 fundamental:3 probabilistic:1 off:10 contract:1 diverge:1 together:1 again:1 satisfied:1 choose:3 worse:1 derivative:1 waste:1 forsyth:1 satisfy:1 depends:3 later:2 break:1 lot:1 analyze:2 start:2 recover:2 characteristic:3 ensemble:6 correspond:1 ofthe:1 bayesian:10 none:1 submitted:3 detector:3 reach:1 ed:1 failure:1 energy:2 pp:1 james:1 obvious:1 e2:3 proof:2 dataset:2 proved:1 recall:1 knowledge:4 subsection:1 sophisticated:1 ok:1 wesley:1 higher:1 follow:1 specify:2 response:2 formulation:1 done:1 though:2 evaluated:1 stage:3 p6:5 mistakenly:1 replacing:1 defines:1 concept:1 true:12 hence:1 spatially:1 leibler:1 branching:7 criterion:5 plate:1 stress:1 demonstrate:1 performs:1 geometrical:2 reasoning:1 image:7 fi:11 funding:1 empirically:1 jl:1 extend:1 refer:2 llx:1 smoothness:1 funded:1 etc:1 posterior:2 recent:6 optimizing:2 discard:1 verlag:1 certain:2 discretizing:1 arbitrarily:1 success:1 yi:1 exploited:1 greater:1 prune:5 determine:10 converge:4 ii:1 branch:9 multiple:2 desirable:1 long:2 award:1 basic:1 vision:8 represent:1 normalization:1 addition:3 want:2 background:1 nary:1 comment:1 effectiveness:2 call:3 integer:3 presence:1 easy:1 simplifies:1 idea:1 shift:1 motivated:1 york:1 amount:1 clutter:3 generate:2 specifies:1 nsf:1 per:1 patt:1 xii:1 write:1 detectability:1 ledgements:1 vol:1 thereafter:1 terminology:1 demonstrating:1 poll:2 threshold:5 imaging:1 graph:2 merely:1 sum:3 uncertainty:1 geiger:1 bound:10 fl:2 mined:1 strength:1 occur:1 precisely:2 constraint:2 x2:1 speed:2 pruned:1 performing:1 numbering:1 combination:1 aerial:1 membrane:1 describes:2 smaller:2 invariant:1 pr:5 taken:1 equation:1 mechanism:1 addison:1 apply:5 observe:2 thomas:2 assumes:3 remaining:1 yx:8 calculating:1 establish:1 society:1 malik:1 question:1 strategy:1 distance:2 thank:1 length:2 difficult:1 acknow:1 design:1 anal:1 twenty:1 perform:1 upper:1 markov:6 poff:4 acknowledge:1 january:1 defining:1 extended:2 intensity:11 aly:1 required:1 connection:1 pearl:2 trans:1 address:3 able:1 below:3 usually:1 pattern:2 summarize:2 power:1 event:2 overlap:1 difficulty:5 critical:1 zhu:3 minimax:1 vax:1 fertility:3 review:1 geometric:10 prior:4 determining:2 relative:2 law:2 expect:1 lecture:1 limitation:1 heavy:1 last:2 soon:1 allow:2 understand:1 fall:2 face:1 boundary:2 curve:4 xn:1 transition:3 contour:10 qn:1 made:2 san:2 gestalt:2 pruning:18 emphasize:1 kullback:1 keep:1 global:1 active:1 conclude:1 francisco:2 assumed:1 xi:15 search:21 qz:1 jz:1 nature:1 ca:2 necessarily:1 domain:6 bounding:1 noise:2 xu:1 representative:4 intel:1 tl:3 wiley:1 fails:1 position:2 wish:1 xl:1 lie:3 candidate:1 perceptual:1 admissible:4 theorem:9 removing:1 down:1 specific:2 emphasized:1 showing:1 explored:5 grouping:1 intrinsic:1 workshop:1 false:15 entropy:2 explore:1 visual:13 expressed:2 tracking:4 joined:1 scalar:1 springer:1 corresponds:1 determines:2 chance:1 chiefly:1 llxi:1 labelled:1 hard:1 change:1 included:1 specifically:1 determined:2 inadmissible:1 total:3 select:1 searched:2 |
563 | 1,514 | Coding time-varying signals using sparse,
shift-invariant representations
Terrence J. Sejnowski
terryCsalk.edu
Michael S. Lewicki*
lewickiCsalk.edu
Howard Hughes Medical Institute
Computational Neurobiology Laboratory
The Salk Institute
10010 N. Torrey Pines Rd.
La Jolla, CA 92037
Abstract
A common way to represent a time series is to divide it into shortduration blocks, each of which is then represented by a set of basis
functions. A limitation of this approach, however, is that the temporal alignment of the basis functions with the underlying structure
in the time series is arbitrary. We present an algorithm for encoding
a time series that does not require blocking the data. The algorithm
finds an efficient representation by inferring the best temporal positions for functions in a kernel basis. These can have arbitrary
temporal extent and are not constrained to be orthogonal. This
allows the model to capture structure in the signal that may occur
at arbitrary temporal positions and preserves the relative temporal
structure of underlying events. The model is shown to be equivalent
to a very sparse and highly over complete basis. Under this model,
the mapping from the data to the representation is nonlinear, but
can be computed efficiently. This form also allows the use of existing methods for adapting the basis itself to data. This approach
is applied to speech data and results in a shift invariant, spike-like
representation that resembles coding in the cochlear nerve.
1
Introduction
Time series are often encoded by first dividing the signal into a sequence of blocks.
The data within each block is then fit with a standard basis such as a Fourier or
wavelet. This has a limitation that the components of the bases are arbitrarily
aligned with respect to structure in the time series. Figure 1 shows a short segment
of speech data and the boundaries of the blocks. Although the structure in the signal
is largely periodic, each large oscillation appears in a different position within the
blocks and is sometimes split across blocks. This problem is particularly present
for acoustic events with sharp onset, such as plosives in speech. It also presents
?To whom correspondence should be addressed.
Coding Time-Varying Signals Using Sparse, Shift-Invariant Representations
731
difficulties for encoding the signal efficiently, because any basis that is adapted to
the underlying structure must represent all possible phases. This can be somewhat
circumvented by techniques such as windowing or averaging sliding blocks, but it
would be more desirable if the representation were shift invariant.
time
Figure 1: Blocking results in arbitrary phase alignment the underlying structure.
2
The Model
Our goal is to model a signal by using a small set of kernel functions that can be
placed at arbitrary time points. Ultimately, we want to find the minimal set of
functions and time points that fit the signal within a given noise level. We expect
this type of model to work well for signals composed of events whose onset can occur
at arbitrary temporal positions. Examples of these include, musical instruments
sounds with sharp attack or plosive sounds in speech.
We assume time series x(t) is modeled by
(1)
where Ti indicates the temporal position of the ith kernel function, <Pm [i) , which is
scaled by Si. The notation m[i] represents an index function that specifies which of
the M kernel functions is present at time Ti. A single kernel function can occur at
multiple times during the time series. Additive noise at time t is given by E(t).
A more general way to express (1) is to assume that the kernel functions exist
at all time points during the signal, and let the non-zero coefficients determine
the positions of the kernel functions. In this case, the model can be expressed in
convolutional form
x(t)
L / Sm(T)<Pm(t - T)dT + E(t)
L sm(t) * <Pm(t) + E(t) ,
(2)
m
(3)
m
where Sm(T) is the coefficient at time T for kernel function <Pm.
It is also helpful to express the model in matrix form using a discrete sampling of
the continuous time series:
x = As + E.
(4)
M. S. Lewicki and T. J. Sejnowski
732
The basis matrix, A, is defined by
(5)
where C(a) is an N-by-N circulant matrix parameterized by the vector a. This
matrix is constructed by replicating the kernel functions at each sample position
[~
C(a) =
al
an
a2
a3
aN-2
aN-l
aN-3
aN-2
ao
al
aN-I
a2
al
1
(6)
aN-l
ao
The kernels are zero padded to be of length N . The length of each kernel is typically
much less than the length of the signal, making A very sparse. This can be viewed as
a special case of a Toeplitz matrix. Note that the size of A is M N-by-N, and is thus
an example of an overcomplete basis, i.e. a basis with more basis functions than
dimensions in the data space (Simoncelli et al., 1992; Coifman and Wickerhauser,
1992; Mallat and Zhang, 1993; Lewicki and Sejnowski, 1998) .
3
A probabilistic formulation
The optimal coefficient values for a signal are found by maximizing the posterior
distribution
s = argmaxP(slx,A) = argmaxP(xIA,s)P(s)
8
(7)
8
where s is the most probable representation of the signal. Note that omission of the
normalizing constant P(xIA) does not change the location of the maximum. This
formulation of the problem offers the advantage that the model can fit more general
types of distributions and naturally "denoises" the signal. Note that the mapping
from x to s is nonlinear with non-zero additive noise and an overcomplete basis
(Chen et al., 1996; Lewicki and Sejnowski, 1998). Optimizing (7) essentially selects
out the subset of basis functions that best account for the data.
To define a probabilistic model, we follow previous conventions for linear generative models with additive noise (Cardoso, 1997; Lewicki and Sejnowski, 1998). We
assume the noise, to, to have a Gaussian distribution which yields a data likelihood
for a given representation of
1
logP(xIA,s) ex - 2u 2 (x - As)2.
(8)
The function P(s) describes the a priori distribution of the coefficients. Under the
assumption that P(s) is sparse (highly -peaked around zero), maximizing (7) results
in very few nonzero coefficients. A compact representation of s is to describe the
values of the non-zero coefficients and their temporal positions
M
P(s) =
n",
II P(Um,Tm) = II II P(Um,i)P(Tm ,i),
m
(9)
m=l i =l
where the prior for the non-zero coefficient values, Um,i, is assumed to be Laplacian,
and the prior for the temporal positions (or intervals), Tm,i, is assumed to be a
gamma distribution.
Coding Time-Varying Signals Using Sparse, Shift-Invariant Representations
4
733
Finding the best encoding
A difficult challenge presented by the proposed model is finding a computationally
tractable method for fitting it to the data. The brute-force approach of generating
the basis matrix A generates an intractable number basis functions for signals of
any reasonable length, so we need to look for ways of making the optimization of
(7) more efficient. The gradient of the log posterior is given by
a
as 10gP(sIA,x) oc AT(x - As)
+ z(s) ,
(10)
where z(s) = (logP(s)),. A basic operation required is v = AT u. We saw that
x = As can be computed efficiently using convolution (2). Because AT is also block
circulant
AT = [
C.(~.D
1
(11)
C(?'u )
where ?'(1 : N) = ?(N : -1 : 1). Thus, terms involving AT can also be computed
efficiently using convolution
v = AT U = [ ?1 (-~~ ~ u(t)
?M( -t) * u(t)
1
(12)
Obtaining an initial representation
An alternative approach to optimizing (7) is to make use of the fact that if the
kernel functions are short enough in length, direct multiplication is faster than
convolution, and that, for this highly overcomplete basis, most of the coefficients
will be zero after being fit to the data. The central problem in encoding the signal
then is to determine which coefficients are non-zero, ideally finding a description of
the time series with the minimal number of non-zero coefficients. This is equivalent
to determining the best set of temporal positions for each of the kernel functions
(1).
A crucial step in this approach is to obtain a good initial estimate of the coefficients.
One way to do this is to consider the projection of the signal onto each of the basis
functions, i.e. AT x. This estimate will be exact (i.e. zero residual error) in the
case of zero noise and A orthogonal. For the non-orthogonal, overcomplete case the
solution will be approximate, but for certain choices of the basis matrix, an exact
representation can still be obtained efficiently (Daubechies, 1990; Simoncelli et aI.,
1992).
Figure 2 shows examples of convolving two different kernel functions with data. One
disadvantage with this initial solution is that the coefficient functions s~(t) are not
sparse. For example, even though the signal in figure 2a is composed of only three
instances of the kernel function, the convolution is mostly non-zero.
A simple procedure for obtaining a better initial estimate of the most probable
coefficients is to select the time locations of the maxima (or extrema) in the convolutions. These are positions where the kernel functions capture the greatest amount
of signal structure and where the optimal coefficients are likely to be non-zero. This
generates a large number of positions, but their number can be reduced further by
selecting only those that contribute significantly, i.e. where the average power is
greater than some fraction of the noise level. From these, a basis for the entire signal
is constructed by replicating the kernel functions at the appropriate time positions.
734
M. S. Lewicki and T J Sejnowski
~Z'C7'C71
V1
I
fJVSNSM
~
I
Figure 2: Convolution using the fast Fourier transform is an efficient way to select an
initial solution for the temporal positions of the kernel functions. (a) The convolution of
a sawtooth-shaped kernel function, ?J(t), with a sawtooth waveform, x(t). (b) A single
period sine-wave kernel function convolved with a speech segment.
Once an initial estimate and basis are formed, the most probable coefficient values are estimated using a modified conjugate gradient procedure. The size of the
generated basis does not pose a problem for optimization, because it is has very
few non-zero elements (the number of which is roughly constant per unit time).
This arises because each column is non-zero only around the position of the kernel
function, which is typically much shorter in duration than the data waveform. This
structure affords the use of sparse matrix routines for all the key computations in
the conjugate gradient routine. After the initial fit, there typically are a large number of basis functions that give a very small contribution. These can be pruned
to yield, after refitting, a more probable representation that has significantly fewer
coefficients.
5
Properties of the representation
Figure 3 shows the results of fitting a segment of speech with a sine wave kernel.
The 64 kernel functions were constructed using a single period of a sine function
whose log frequencies were evenly distributed between 0 and Nyquist (4 kHz), which
yielded kernel functions that were minimally correlated (they are not orthogonal
because each has only one cycle and is zero elsewhere). The kernel function lengths
varied between 2 and 64 samples. The plots show the positions of the non-zero
coefficients superimposed on the waveform. The residual errors curves from the
fitted waveforms are shown offset, below each waveform. The right axes indicate
the kernel function number which increase with frequency. The dots show the
starting position of the kernels with non-zero coefficients, with the dot size scaled
according to the mean power contribution. This plot is essentially a time/frequency
analysis, similar to a wavelet decomposition, but on a finer temporal scale.
Figure 3a shows that the structure in the coefficients repeats for each oscillation in
the waveform. Adding a delay leaves the relative temporal structure of the nonzero coefficients mostly unchanged (figure 3b). The small variations between the
two sets of coefficients are due to variations in the fitting of the small-magnitude
coefficients. Representing the signal in figure 3b with a standard complete basis
would result in a very different representation.
735
Coding Time- Varying Signals Using Sparse, Shift-Invariant Representations
a
.. .
.
.:
...
.: .
0.2
: ?
e? ? ?
:
:
0.1
:
53
o
14
~. 1
o
20
40
60
60
100
120
time
14
~. 1
o
20
40
60
80
100
120
time
Figure 3: Fitting a shift-invariant model to a segment of speech, x(t). Dots indicate
positions of kernels (right axis) with size scaled by the mean power contribution. Fitting
error is plotted below speech signal.
M S. Lewicki and T. J. Sejnowski
736
6
Discussion
The model presented here can be viewed as an extension of the shiftable transforms
of Simoncelli et al. (1992). One difference is that here no constraints are placed on
the kernel functions. Furthermore, this model accounts for additive noise, which
yields automatic signal denoising and provides sensible criteria for selecting significant coefficients. An important unresolved issue is how well the algorithm works
for increasingly non-orthogonal kernels.
One interesting property of this representation is that it results in a spike-like representation . In the resulting set of non-zero coefficients, not only is their value important for representing the signal, but also their relative temporal position, which
indicate when an underlying event has occurred. This shares many properties with
cochlear models. The model described here also has capacity to have an over complete representation at any given timepoint, e.g. a kernel basis with an arbitrarily
large number of frequencies. These properties make this model potentially useful
for binaural signal processing applications.
The effectiveness of this method for efficient coding remains to be proved. A trivial
example of a shift-invariant basis is a delta-function model. For a model to encode
information efficiently, the representation should be non-redundant. Each basis
function should "grab" as much structure in the data as possible and achieve the
same level of coding efficiency for arbitrary shifts of the data. The matrix form
of the model (4) suggests that it is possible to achieve this optimum by adapting
the kernel functions themselves using the methods of Lewicki and Sejnowski (1998).
Initial results suggest that this approach is promising. Beyond this, it is evident that
modeling the higher-order structure in the coefficients themselves will be necessary
both to achieve an efficient representation and to capture structure that is relevant
to such tasks as speech recognition or auditory stream segmentation. These results
are a step toward these goals.
Acknowledgments. We thank Tony Bell, Bruno Olshausen, and David Donoho
for helpful discussions.
References
Cardoso, J.-F. (1997). Infomax and maximum likelihood for blind source separation.
IEEE Signal Processing Letters, 4:109- 11 I.
Chen, S., Donoho, D. L., and Saunders, M. A. (1996). Atomic decomposition by
basis pursuit. Technical report, Dept. Stat., Stanford Univ., Stanford, CA.
Coifman, R. R. and Wickerhauser, M. V. (1992). Entropy-based algorithms for best
basis selection. IEEE Transactions on Information Theory, 38(2) :713- 718.
Daubechies, I. (1990). The wavelet transform, time-frequency localization, and
signal analysis. IEEE Transactions on Information Theory, 36(5):961- 1004.
Lewicki, M. S. and Sejnowski, T. J. (1998). Learning overcomplete representations.
Neural Computation. submitted.
Mallat, S. G. and Zhang, Z. F. (1993). Matching pursuits with time-frequency
dictionaries. IEEE Transactions on Signal Processing, 41(12):3397-3415.
Simoncelli, E. P., Freeman, W. T., Adelson, E. H., and J ., H. D. (1992). Shiftable
multiscale transforms. IEEE Trans . Info . Theory, 38:587- 607.
| 1514 |@word decomposition:2 initial:8 series:9 selecting:2 existing:1 si:1 must:1 additive:4 plot:2 generative:1 fewer:1 leaf:1 ith:1 short:2 provides:1 contribute:1 location:2 attack:1 zhang:2 constructed:3 direct:1 fitting:5 coifman:2 roughly:1 themselves:2 freeman:1 underlying:5 notation:1 finding:3 extremum:1 temporal:14 ti:2 um:3 scaled:3 brute:1 unit:1 medical:1 encoding:4 minimally:1 resembles:1 suggests:1 slx:1 acknowledgment:1 atomic:1 hughes:1 block:8 procedure:2 bell:1 adapting:2 significantly:2 projection:1 matching:1 suggest:1 onto:1 selection:1 equivalent:2 maximizing:2 starting:1 duration:1 variation:2 mallat:2 exact:2 element:1 recognition:1 particularly:1 blocking:2 capture:3 cycle:1 ideally:1 ultimately:1 segment:4 localization:1 efficiency:1 basis:28 binaural:1 represented:1 univ:1 fast:1 describe:1 sejnowski:9 saunders:1 whose:2 encoded:1 stanford:2 toeplitz:1 torrey:1 gp:1 itself:1 transform:2 sequence:1 advantage:1 unresolved:1 aligned:1 relevant:1 achieve:3 description:1 optimum:1 argmaxp:2 generating:1 stat:1 pose:1 ex:1 dividing:1 indicate:3 convention:1 waveform:6 require:1 ao:2 timepoint:1 probable:4 extension:1 around:2 mapping:2 pine:1 dictionary:1 a2:2 saw:1 gaussian:1 modified:1 varying:4 encode:1 ax:1 indicates:1 likelihood:2 superimposed:1 helpful:2 typically:3 entire:1 selects:1 issue:1 priori:1 constrained:1 special:1 once:1 shaped:1 sampling:1 represents:1 look:1 adelson:1 peaked:1 report:1 few:2 composed:2 preserve:1 gamma:1 phase:2 highly:3 alignment:2 necessary:1 shorter:1 orthogonal:5 divide:1 plotted:1 overcomplete:5 minimal:2 fitted:1 instance:1 column:1 modeling:1 disadvantage:1 logp:2 subset:1 shiftable:2 delay:1 periodic:1 refitting:1 probabilistic:2 terrence:1 infomax:1 michael:1 daubechies:2 central:1 convolving:1 denoises:1 account:2 coding:7 coefficient:25 onset:2 stream:1 blind:1 sine:3 wave:2 contribution:3 formed:1 convolutional:1 musical:1 largely:1 efficiently:6 yield:3 finer:1 submitted:1 c7:1 frequency:6 naturally:1 auditory:1 proved:1 segmentation:1 routine:2 nerve:1 appears:1 higher:1 dt:1 follow:1 formulation:2 though:1 furthermore:1 nonlinear:2 multiscale:1 olshausen:1 laboratory:1 nonzero:2 during:2 oc:1 criterion:1 evident:1 complete:3 common:1 khz:1 occurred:1 significant:1 ai:1 rd:1 automatic:1 pm:4 bruno:1 replicating:2 dot:3 base:1 posterior:2 optimizing:2 jolla:1 certain:1 arbitrarily:2 greater:1 somewhat:1 determine:2 period:2 redundant:1 signal:30 ii:3 sliding:1 windowing:1 desirable:1 sound:2 multiple:1 simoncelli:4 technical:1 faster:1 offer:1 laplacian:1 involving:1 basic:1 essentially:2 represent:2 kernel:32 sometimes:1 want:1 addressed:1 interval:1 source:1 crucial:1 effectiveness:1 split:1 enough:1 fit:5 tm:3 shift:9 nyquist:1 speech:9 useful:1 cardoso:2 amount:1 transforms:2 reduced:1 specifies:1 exist:1 affords:1 estimated:1 delta:1 per:1 discrete:1 express:2 key:1 v1:1 grab:1 padded:1 wickerhauser:2 fraction:1 parameterized:1 letter:1 reasonable:1 separation:1 oscillation:2 correspondence:1 yielded:1 adapted:1 occur:3 constraint:1 generates:2 fourier:2 pruned:1 circumvented:1 according:1 conjugate:2 across:1 describes:1 increasingly:1 making:2 invariant:8 computationally:1 remains:1 tractable:1 instrument:1 pursuit:2 operation:1 appropriate:1 alternative:1 convolved:1 include:1 tony:1 unchanged:1 spike:2 gradient:3 thank:1 capacity:1 sensible:1 evenly:1 cochlear:2 whom:1 extent:1 trivial:1 toward:1 length:6 sia:1 modeled:1 index:1 difficult:1 mostly:2 potentially:1 info:1 convolution:7 sm:3 howard:1 neurobiology:1 varied:1 arbitrary:7 sharp:2 omission:1 david:1 required:1 acoustic:1 trans:1 beyond:1 below:2 challenge:1 greatest:1 event:4 power:3 difficulty:1 force:1 residual:2 representing:2 axis:1 prior:2 multiplication:1 determining:1 relative:3 sawtooth:2 expect:1 interesting:1 limitation:2 share:1 elsewhere:1 placed:2 repeat:1 institute:2 circulant:2 sparse:9 distributed:1 boundary:1 dimension:1 xia:3 curve:1 transaction:3 approximate:1 compact:1 assumed:2 continuous:1 promising:1 ca:2 obtaining:2 plosive:2 noise:8 salk:1 inferring:1 position:19 wavelet:3 offset:1 a3:1 normalizing:1 intractable:1 adding:1 magnitude:1 chen:2 entropy:1 likely:1 expressed:1 lewicki:9 goal:2 viewed:2 donoho:2 change:1 averaging:1 denoising:1 la:1 select:2 arises:1 dept:1 correlated:1 |
564 | 1,515 | Almost Linear VC Dimension Bounds for
Piecewise Polynomial Networks
Peter L. Bartlett
Department of System Engineering
Australian National University
Canberra, ACT 0200
Australia
Peter.Bartlett@anu.edu.au
Vitaly Maiorov
Department of Mathematics
Technion, Haifa 32000
Israel
Ron Meir
Department of Electrical Engineering
Technion, Haifa 32000
Israel
rmeir@dumbo.technion.ac.il
Abstract
We compute upper and lower bounds on the VC dimension of
feedforward networks of units with piecewise polynomial activation functions. We show that if the number of layers is fixed, then
the VC dimension grows as W log W, where W is the number of
parameters in the network. This result stands in opposition to the
case where the number of layers is unbounded, in which case the
VC dimension grows as W 2 ?
1
MOTIVATION
The VC dimension is an important measure of the complexity of a class of binaryvalued functions, since it characterizes the amount of data required for learning in
the PAC setting (see [BEHW89, Vap82]). In this paper, we establish upper and
lower bounds on the VC dimension of a specific class of multi-layered feedforward
neural networks. Let F be the class of binary-valued functions computed by a
feed forward neural network with W weights and k computational (non-input) units,
each with a piecewise polynomial activation function. Goldberg and Jerrum [GJ95]
Cl(W 2 + Wk) = O(W2), where Cl is a constant.
have shown that VCdim(F)
Moreover, Koiran and Sontag [KS97] have demonstrated such a network that has
VCdim(F) ~ C2 W 2 = O(W2), which would lead one to conclude that the bounds
:s
Almost Linear VC Dimension Bounds for Piecewise Polynomial Networks
191
are in fact tight up to a constant. However, the proof used in [KS97] to establish
the lower bound made use of the fact that the number of layers can grow with W.
In practical applications, this number is often a small constant. Thus, the question
remains as to whether it is possible to obtain a better bound in the realistic scenario
where the number of layers is fixed.
The contribution of this work is the proof of upper and lower bounds on the VC
dimension of piecewise polynomial nets. The upper bound behaves as O(W L2 +
W L log W L), where L is the number of layers. If L is fixed, this is O(W log W),
which is superior to the previous best result which behaves as O(W2). Moreover,
using ideas from [KS97] and [GJ95] we are able to derive a lower bound on the VC
dimension which is O(WL) for L = O(W). Maass [Maa94] shows that three-layer
networks with threshold activation functions and binary inputs have VC dimension
O(W log W), and Sakurai [Sak93] shows that this is also true for two-layer networks
with threshold activation functions and real inputs. It is easy to show that these
results imply similar lower bounds if the threshold activation function is replaced by
any piecewise polynomial activation function f that has bounded and distinct limits
limx-t - oo f(x) and limx-too f(x). We thus conclude that if the number oflayers L is
fixed, the VC dimension of piecewise polynomial networks with L ~ 2 layers and real
inputs, and of piecewise polynomial networks with L ~ 3 layers and binary inputs,
grows as W log W. We note that for the piecewise polynomial networks considered
in this work, it is easy to show that the VC dimension and pseudo-dimension are
closely related (see e.g. [Vid96]), so that similar bounds (with different constants)
hold for the pseudo-dimension. Independently, Sakurai has obtained similar upper
bounds and improved lower bounds on the VC dimension of piecewise polynomial
networks (see [Sak99]).
2
UPPER BOUNDS
We begin the technical discussion with precise definitions of the VC-dimension and
the class of networks considered in this work.
Definition 1 Let X be a set, and A a system of subsets of X. A set S =
{ Xl, . .. ,xn} is shattered by A if, for every subset B ~ S, there exists a set A E A
such that SnA = B. The VC-dimension of A, denoted by VCdim(A), is the largest
integer n such that there exists a set of cardinality n that is shattered by A.
Intuitively, the VC dimension measures the size, n, of the largest set of points for
which all possible 2n labelings may be achieved by sets A E A. It is often convenient
to talk about the VC dimension of classes of indicator functions F. In this case we
simply identify the sets of points X E X for which f(x) = 1 with the subsets of A,
and use the notation VCdim(F).
A feedforward multi-layer network is a directed acyclic graph that represents a
parametrized real-valued function of d real inputs. Each node is called either an
input unit or a computation unit. The computation units are arranged in L layers.
Edges are allowed from input units to computation units. There can also be an
edge from a computation unit to another computation unit, but only if the first
unit is in a lower layer than the second. There is a single unit in the final layer,
called the output unit. Each input unit has an associated real value, which is One
of the components of the input vector x E Rd. Each computation unit has an
associated real value, called the unit's output value. Each edge has an associated
real parameter, as does each computation unit. The output of a computation unit
is given by (7 CEe weze + wo), where the sum ranges over the set of edges leading to
192
P L. Bartlett, V. Maiorov and R. Meir
the unit, We is the parameter (weight) associated with edge e, Ze is the output value
of the unit from which edge e emerges, Wo is the parameter (bias) associated with
the unit, and a : R -t R is called the activation function of the unit. The argument
of a is called the net input of the unit. We suppose that in each unit except the
output unit, the activation function is a fixed piecewise polynomial function of the
form
for i = 1, ... ,p+ 1 (and set to = -00 and tp+1 = 00), where each cPi is a polynomial
of degree no more than l. We say that a has p break-points, and degree l. The
activation function in the output unit is the identity function. Let k i denote the
number of computational units in layer i and suppose there is a total of W parameters (weights and biases) and k computational units (k = k1 + k2 + ... + k L - 1 + 1).
For input x and parameter vector a E A = R w, let f(x, a) denote the output of
this network, and let F = {x f-t f(x,a) : a E RW} denote the class of functions
computed by such an architecture, as we vary the W parameters. We first discuss the computation of the VC dimension, and thus consider the class of functions
sgn(F) = {x f-t sgn(f(x, a)) : a E RW}.
Before giving the main theorem of this section, we present the following result,
which is a slight improvement of a result due to Warren (see [ABar], Chapter 8).
Lemma 2.1 Suppose II (.), h (.), .. , ,fm (-) are fixed polynomials of degree at
most 1 in n ~ m variables.
Then the number of distinct sign vectors
{sgn(Jl (a)), ... ,sgn(Jm (a))} that can be generated by varying a ERn is at most
2(2eml/n)n.
We then have our main result:
Theorem 2.1 For any positive integers W, k ~ W, L ~ W, l, and p, consider a
network with real inputs, up to W parameters, up to k computational units arranged
in L layers, a single output unit with the identity activation function, and all other
computation units with piecewise polynomial activation functions of degree 1 and
with p break-points. Let F be the class of real-valued functions computed by this
network. Then
VCdim(sgn(F)) ~ 2WLlog(2eWLpk)
+ 2WL2log(1 + 1) + 2L.
Since Land k are O(W), for fixed 1 and p this implies that
VCdim(sgn(F)) = O(WLlogW
+ WL2).
Before presenting the proof, we outline the main idea in the construction. For
any fixed input x, the output of the network f(x, a) corresponds to a piecewise
polynomial function in the parameters a, of degree no larger than (l + I)L-1 (recall
that the last layer is linear). Thus, the parameter domain A = R W can be split
into regions, in each of which the function f(x,?) is polynomial. From Lemma 2.1,
it is possible to obtain an upper bound on the number of sign assignments that can
be attained by varying the parameters of a set of polynomials. The theorem will be
established by combining this bound with a bound on the number of regions.
PROOF OF THEOREM
2.1 For an arbitrary choice of m points Xl, X2, ..? ,xm , we
wish to bound
K
= I{(sgn(f(Xl ,a)), . ..
,sgn(J(xm, a))) : a E A }I.
Almost Linear VC Dimension Bounds for Piecewise Polynomial Networks
193
Fix these m points, and consider a partition {SI, S2, ... , S N} of the parameter
domain A. Clearly
N
K ~
L I{(sgn(J(xl , a?, ... , sgn(J(xm, a?) : a ESdi?
i=1
We choose the partition so that within each region Si, f (Xl, .), ... ,f (x m, .) are all
fixed polynomials of degree no more than (1 + I)L-1. Then, by Lemma 2.1, each
term in the sum above is no more than
2
(2em(1;' I)L - l) W
(1)
The only remaining point is to construct the partition and determine an upper
bound on its size. The partition is constructed recursively, using the following
procedure. Let 51 be a partition of A such that, for all S E 51, there are constants
bh,i,j E {0,1} for which
for all a E S,
where j E {I, ... ,m}, h E {I, ... ,kd and i E {1, ... ,pl. Here ti are the breakpoints of the piecewise polynomial activation functions, and Ph,x) is the affine function describing the net input to the h-th unit in the first layer, in response to X j.
That is,
where ah E R d, ah,O E R are the weights of the h-th unit in the first layer. Note
that the partition 51 is determined solely by the parameters corresponding to the
first hidden layer, as the input to this layer is unaffected by the other parameters.
Clearly, for a E S, the output of any first layer unit in response to an Xj is a fixed
polynomial in a.
Now, let WI, ... , W L be the number of variables used in computing the unit outputs
up to layer 1, ... , L respectively (so W L = W), and let k l , . .. , kL be the number of
computation units in layer 1, ... , L respectively (recall that kL = 1). Then we can
choose 51 so that 151 1is no more than the number of sign assignments possible with
mk l P affine functions in WI variables. Lemma 2.1 shows that
151 ~ 2 (2e~~IP) WI
1
Now, we define 5 n (for n > 1) as follows. Assume that for all S in 5 n - 1 and all
Xj, the net input of every unit in layer n in response to Xj is a fixed polynomial
function of a E S, of degree no more than (1 + l)n-1 . Let 5 n be a partition of A
that is a refinement of 5 n- 1 (that is, for all S E 5 n , there is an S' E 5 n- 1 with
S ~ S'), such that for all S E 5 n there are constants bh,i,j E {O, I} such that
sgn(Ph,x) (a)
- ti ) =
bh,i,j
for all a E S,
(2)
where Ph ,x) is the polynomial function describing the net input of the h-th unit in
the n-th layer, in response to Xj, when a E S. Since S ~ S' for some S' E 5 n- 1 , (2)
implies that the output of each n-th layer unit in response to an X j is a fixed
polynomial in a of degree no more than l (l + 1) n-l, for all a E S.
Finally, we can choose 5 n such that, for all S' E 5 n- 1 we have I{S E 5 n : S ~
S'}I is no more than the number of sign assignments of mknP polynomials in Wn
variables of degree no more than (l + 1)n- l, and by Lemma 2.1 this is no more than
2 (2emkn~n+lr-I ) Wn
.
Notice also that the net input of every unit in layer n + 1 in
P. L. Bartlett, V Maiorov and R. Meir
194
response to
(l + l)n.
Xj
is a fixed polynomial function of a ESE Sn of degree no more than
Proceeding in this way we get a partition SL-l of A such that for S E SL-l the
network output in response to any Xj is a fixed polynomial of a E S of degree no
more than l(l + 1)L-2. Furthermore,
JSL-d
< 2
<
Ce;:,P)
W,
TI eemk'p~,+ 1)'-')
2
TI 2CemkiP~,+ 1)'-')
W,
W;
Multiplying by the bound (1) gives the result
~
K
IT
2 (2emkip(l .+ l)i-l) W.
i=l
Wt
Since the points Xl, ... ,X m were chosen arbitrarily, this .gives a bound on the maximal number of dichotomies induced by a E A on m points. An upper bound on
the VC-dimension is then obtained by computing the largest value of m for which
this number is at least 2 m , yielding
m
< L+
t.
w, log Cempk'~,+ 1)i-1 )
< L [1 + (L - l)W log(l + 1) + W log(2empk)] ,
where all logarithms are to the base 2. We conclude (see for example [Vid96] Lemma
4.4) that
VCdim(F) ~ 2L [(L -l)W log(l
+ 1) + W
log (2eWLpk)
+ 1].
?
We briefly mention the application of this result to the problem of learning a regression function E[YIX = x], from n input/output pairs {(Xi, Yi)}i=l' drawn
independently at random from an unknown distribution P(X, Y). In the case of
quadratic loss, L(f) = E(Y - f(X))2, one can show that there exist constants Cl ;::: 1
and C2 such that
EL(f~n ) <
-
8
2
? f L- (f)
logn
+ Cl JET
In
+ C2 MPdim(F)
,
n
where 8 2 = E [Y - E[YIX]]2 is the noise variance, i(f) = E [(E[YIX] - f(X))2] is
the approximation error of f, and
is a function from the class F that approximately minimizes the sample average of the quadratic loss. Making use of recently
derived bounds [MM97] on the approximation error, inf JET i(f), which are equal,
up to logarithmic factors, to those obtained for networks of units with the standard sigmoidal function u{u) = (1 + e-u)-l , and combining with the considerably
lower pseudo-dimension bounds for piecewise polynomial networks, we obtain much
better error rates than are currently available for sigmoid networks.
in
3
LOWER BOUND
We now compute a lower bound on the VC dimension of neural networks with
continuous activation functions. This result generalizes the lower bound in [KS97],
since it holds for any number of layers.
Almost Linear VC Dimension Bounds for Piecewise Polynomial Networks
195
Theorem 3.1 Suppose f : R -+ R has the following properties:
1. limo-too f(a)
= 1 and limo-t-oo f(a) = 0,
2. f is differentiable at some point
Xo
and
with derivative f'(xo)
=1=
O.
Then for any L ~ 1 and W ~ 10L - 14, there is a feedforward network with the
following properties: The network has L layers and W parameters, the output unit
is a linear unit, all other computation units have activation function f, and the set
sgn(F) of functions computed by the network has
VCdim(sgn(F?
l l
~ ~ J ~ J'
where l u J is the largest integer less than or equal to u.
PROOF
As in [KS97], the proof follows that of Theorem 2.5 in [GJ95], but we
show how the functions described in [GJ95] can be computed by a network, and
keep track of the number of parameters and layers required. We first prove the
lower bound for a network containing linear threshold units and linear units (with
the identity activation function), and then show that all except the output unit
can be replaced by units with activation function f, and the resulting network still
shatters the same set. For further details of the proof, see the full paper [BMM98].
Fix positive integers M, N E N. We now construct a set of M N points, which
may be shattered by a network with O(N) weights and O(M) layers. Let {ad,
i = 1,2, ... ,N denote a set of N parameters, where each ai E [0,1) has an M -bit
binary representation ai = E~l 2-jai,j, ai,j E {O, I}, i.e. the M-bit base two
representation of ai is ai = O.ai,l ai,2 ... ai,M. We will consider inputs in B N X B M,
where BN = {ei : 1 ~ i ~ N}, ei E {O, I}N has i-th bit 1 and all other bits 0, and
BM is defined similarly. We show how to extract the bits of the ai, so that for
input x = (el' ern) the network outputs al,rn. Since there are N M inputs of the
form (el,e rn ), and al,rn can take on all possible 2MN values, the result will follow.
There are three stages to the computation of al,rn: (1) computing ai, (2) extracting
al,k from ai, for every k, and (3) selecting al,rn among the al,ks.
Suppose the network input is x = ((Ul,'" ,UN),(Vt, ... ,VM? = (el,e rn ). Using
one linear unit we can compute E~l Uiai = al. This involves N + 1 parameters
and one computation unit in one layer. In fact, we only need N parameters, but we
need the extra parameter when we show that this linear unit can be replaced by a
unit with activation function f.
Consider the parameter Ck = O.al,k ... al,M, that is, Ck = E~k 2k-1-jal,j for k =
1, ... ,M. Since Ck ~ 1/2 iff al,k = 1, clearly sgn(ck - 1/2) = al,k for all k. Also,
Cl = al and Ck = 2Ck-l - al ,k-l' Thus, consider the recursion
Ck
= 2Ck-l
- al,k-l
= sgn(ck - 1/2)'
with initial conditions CI = al and au = sgn(al - 1/2). Clearly, we can compute
al,l, ... ,al,M-l and C2,' .. ,CM-l in another 2(M - 2) + 1 layers, using 5(M - 2) + 2
parameters in 2(M - 2) + 1 computational units.
al,k
We could compute al,M in the same way, but the following approach gives fewer
layers. Set b = sgn (2C M - 1 - al,M - l then the input vector (VI, ... ,VM)
b = sgn(cM) = sgn(O.al,M) = al,M.
E~~I Vi)'
= eM,
If m
=1= M
then b = O. If m
and thus E~~lvi
= 0,
=M
implying that
P L. Bartlett, V. Maiorov and R. Meir
196
In order to conclude the proof, we need to show how the variables al,m may be
recovered, depending on the inputs (VI, V2, ... ,VM). We then have al,m = b V
V';~I(al,i/\vi). Since for boolean x and y, x/\y = sgn(x+y-3/2), and V';I Xi =
sgn(2:,;1 Xi - 1/2), we see that the computation of al,m involves an additional 5M
parameters in M + 1 computational units, and adds another 2 layers.
In total, there are 2M layers and 10M + N -7 parameters, and the network shatters
a set of size N M. Clearly, we can add parameters and layers without affecting
the function of the network. So for any L, WEN, we can set M = lL/2J and
N = W + 7 - 10M, which is at least lW/2J provided W :2: 10L - 14. In that case,
the VC-dimension is at least l L /2 J l W /2 J .
The network just constructed uses linear threshold units and linear units. However,
it is easy to show (see [KS97], Theorem 5) that each unit except the output unit can
be replaced by a unit with activation function f so that the network still shatters the
set of size M N. For linear units, the input and output weights are scaled so that the
linear function can be approximated to sufficient accuracy by f in the neighborhood
of the point Xo. For linear threshold units, the input weights are scaled so that the
behavior of f at infinity accurately approximates a linear threshold function.
?
References
M. Anthony and P. L. Bartlett. Neural Network Learning: Theoretical
Foundations. Cambridge University Press, 1999 (to appear).
[BEHW89] A. Blumer, A. Ehrenfeucht, D. Haussler, and M. K. Warmuth. Learnability and the Vapnik-Chervonenkis dimension. J. ACM, 36(4):929965, 1989.
[BMM98] P. L. Bartlett, V. Maiorov, and R. Meir. Almost linear VC-dimension
bounds for piecewise polynomial networks.
Neural Computation,
10:2159- 2173, 1998.
P.W. Goldberg and M.R. Jerrum. Bounding the VC Dimension of
[GJ95]
Concept Classes Parameterized by Real Numbers. Machine Learning,
18:131 - 148, 1995.
P. Koiran and E.D. Sontag. Neural Networks with Quadratic VC Di[KS97] .
mension. Journal of Computer and System Science, 54:190- 198, 1997.
[ABar]
[Maa94]
W. Maass. Neural nets with superlinear VC-dimension. Neural Computation, 6(5):877- 884, 1994.
[MM97]
V. Maiorov and R. Meir.
On the Near Optimality of the
Stochastic Approximation of Smooth Functions by Neural Networks.
Submitted for publication, 1997.
A. Sakurai. Tighter bounds on the VC-dimension of three-layer networks. In World Congress on Neural Networks, volume 3, pages 540543, Hillsdale, NJ, 1993. Erlbaum.
[Sak93]
[Sak99]
A. Sakurai. Tight bounds for the VC-dimension of piecewise polynomial networks. In Advances in Neural Information Processing Systems,
volume 11. MIT Press, 1999.
[Vap82]
V. N. Vapnik. Estimation of Dependences Based on Empirical Data.
Springer-Verlag, New York, 1982.
M Vidyasagar. A Theory of Learning and Generalization. Springer
Verlag, New York, 1996.
[Vid96]
| 1515 |@word behw89:2 briefly:1 polynomial:31 bn:1 mention:1 recursively:1 initial:1 selecting:1 chervonenkis:1 recovered:1 activation:18 si:2 realistic:1 partition:8 implying:1 fewer:1 warmuth:1 lr:1 node:1 ron:1 sigmoidal:1 unbounded:1 c2:4 constructed:2 prove:1 behavior:1 multi:2 jm:1 cardinality:1 begin:1 provided:1 moreover:2 bounded:1 notation:1 israel:2 cm:2 minimizes:1 nj:1 pseudo:3 every:4 act:1 ti:4 k2:1 scaled:2 unit:60 appear:1 before:2 positive:2 engineering:2 congress:1 limit:1 solely:1 approximately:1 au:2 k:1 range:1 directed:1 practical:1 eml:1 procedure:1 empirical:1 binaryvalued:1 convenient:1 get:1 superlinear:1 layered:1 bh:3 demonstrated:1 independently:2 haussler:1 construction:1 suppose:5 goldberg:2 us:1 ze:1 approximated:1 electrical:1 region:3 complexity:1 tight:2 chapter:1 talk:1 distinct:2 dichotomy:1 neighborhood:1 limo:2 larger:1 valued:3 say:1 jerrum:2 maiorov:6 final:1 ip:1 differentiable:1 net:7 oflayers:1 maximal:1 combining:2 iff:1 sna:1 derive:1 oo:2 ac:1 depending:1 involves:2 implies:2 australian:1 closely:1 stochastic:1 vc:29 australia:1 sgn:21 hillsdale:1 vcdim:8 fix:2 generalization:1 tighter:1 pl:1 hold:2 considered:2 koiran:2 vary:1 estimation:1 currently:1 largest:4 wl:1 mit:1 clearly:5 yix:3 ck:9 varying:2 publication:1 derived:1 improvement:1 el:4 shattered:3 hidden:1 labelings:1 among:1 denoted:1 logn:1 equal:2 construct:2 represents:1 dumbo:1 piecewise:19 wen:1 national:1 replaced:4 limx:2 yielding:1 edge:6 wl2:1 logarithm:1 haifa:2 theoretical:1 mk:1 boolean:1 tp:1 sakurai:4 assignment:3 subset:3 technion:3 erlbaum:1 too:2 learnability:1 considerably:1 vm:3 containing:1 choose:3 derivative:1 leading:1 wk:1 ad:1 vi:4 break:2 characterizes:1 contribution:1 il:1 accuracy:1 variance:1 identify:1 accurately:1 multiplying:1 unaffected:1 ah:2 submitted:1 definition:2 proof:8 associated:5 di:1 cee:1 recall:2 emerges:1 feed:1 attained:1 follow:1 response:7 improved:1 arranged:2 furthermore:1 just:1 stage:1 ei:2 grows:3 concept:1 true:1 maass:2 ehrenfeucht:1 ll:1 presenting:1 outline:1 recently:1 superior:1 sigmoid:1 behaves:2 volume:2 jl:1 slight:1 approximates:1 cambridge:1 ai:11 rd:1 mathematics:1 similarly:1 base:2 add:2 inf:1 scenario:1 verlag:2 binary:4 arbitrarily:1 vt:1 yi:1 additional:1 determine:1 ii:1 full:1 smooth:1 technical:1 jet:2 regression:1 achieved:1 affecting:1 grow:1 w2:3 extra:1 induced:1 vitaly:1 integer:4 extracting:1 near:1 feedforward:4 split:1 easy:3 wn:2 xj:6 architecture:1 fm:1 idea:2 whether:1 ese:1 bartlett:7 ul:1 wo:2 peter:2 sontag:2 jal:1 york:2 amount:1 ph:3 rw:2 meir:6 sl:2 exist:1 notice:1 sign:4 rmeir:1 track:1 threshold:7 drawn:1 shatters:3 ce:1 graph:1 sum:2 parameterized:1 almost:5 bit:5 bound:34 layer:38 opposition:1 breakpoints:1 quadratic:3 infinity:1 x2:1 argument:1 optimality:1 ern:2 department:3 kd:1 em:2 wi:3 making:1 intuitively:1 xo:3 remains:1 discus:1 describing:2 available:1 generalizes:1 v2:1 remaining:1 giving:1 k1:1 establish:2 question:1 dependence:1 parametrized:1 unknown:1 upper:9 precise:1 rn:6 arbitrary:1 pair:1 required:2 kl:2 vap82:2 established:1 able:1 xm:3 vidyasagar:1 indicator:1 recursion:1 mn:1 imply:1 extract:1 sn:1 jai:1 l2:1 loss:2 acyclic:1 foundation:1 degree:11 affine:2 sufficient:1 jsl:1 land:1 last:1 bias:2 warren:1 dimension:32 xn:1 stand:1 world:1 forward:1 made:1 refinement:1 bm:1 keep:1 conclude:4 xi:3 continuous:1 un:1 cl:5 anthony:1 domain:2 lvi:1 main:3 motivation:1 s2:1 noise:1 bounding:1 allowed:1 cpi:1 canberra:1 wish:1 xl:6 lw:1 theorem:7 specific:1 pac:1 exists:2 vapnik:2 ci:1 anu:1 logarithmic:1 simply:1 springer:2 corresponds:1 acm:1 identity:3 blumer:1 determined:1 except:3 wt:1 lemma:6 called:5 total:2 |
565 | 1,516 | Probabilistic Image Sensor Fusion
Ravi K. Sharma1 , Todd K. Leen 2 and Misha Pavel 1
1 Department
of Electrical and Computer Engineering
2Department of Computer Science and Engineering
Oregon Graduate Institute of Science and Technology
P.O. Box 91000 , Portland , OR 97291-1000
Email: {ravi,pavel} @ece.ogi.edu, tleen@cse .ogi.edu
Abstract
We present a probabilistic method for fusion of images produced
by multiple sensors . The approach is based on an image formation
model in which the sensor images are noisy, locally linear functions
of an underlying, true scene. A Bayesian framework then provides
for maximum likelihood or maximum a posteriori estimates of the
true scene from the sensor images. Maximum likelihood estimates
of the parameters of the image formation model involve (local)
second order image statistics, and thus are related to local principal
component analysis. We demonstrate the efficacy of the method
on images from visible-band and infrared sensors .
1
Introduction
Advances in sensing devices have fueled the deployment of multiple sensors in several
computational vision systems [1, for example]. Using multiple sensors can increase
reliability with respect to single sensor systems. This work was motivated by a
need for an aircraft autonomous landing guidance (ALG) system [2, 3] that uses
visible-band, infrared (IR) and radar-based imaging sensors to provide guidance
to pilots for landing aircraft in low visibility. IR is suitable for night operation,
whereas radar can penetrate fog. The application requires fusion algorithms [4] to
combine the different sensor images .
Images from different sensors have different characteristics arising from the varied
physical imaging processes. Local contrast may be polarity reversed between visibleband and IR images [5 , 6] . A particular sensor image may contain local features
not found in another sensor image , i.e., sensors may report complementary features .
Finally, individual sensors are subj ect to noise. Fig . l(a) and l(b) are visible-band
and IR images respectively, of a runway scene showing polarity reversed (rectangle)
825
Probabilistic Image Sensor Fusion
and complementary (circle) features. These effects pose difficulties for fusion.
An obvious approach to fusion is to average the pixel intensities from different
sensors. Averaging, Fig. l(c), increases the signal to noise ratio, but reduces the
contrast where there are polarity reversed or complementary features [7].
Transform-based fusion methods [8, 5, 9] selectfrom one sensor or another for fusion.
They consist of three steps: (i) decompose the sensor images using a specified
transform e.g. a multiresolution Laplacian pyramid, (ii) fuse at each level of the
pyramid by selecting the highest energy transform coefficient, and (iii) invert the
transform to synthesize the fused image. Since features are selected rather than
averaged, they are rendered at full contrast, but the methods are sensitive to sensor
noise, see Fig. l(d).
To overcome the limitations of averaging or selection methods, and put sensor fusion
on firm theoretical grounds, we explicitly model the production of sensor images
from the true scene, including the effects of sensor noise. From the model, and
sensor images, one can ask What is the most probable true scene? This forms
the basis for fusing the sensor images. Our technique uses the Laplacian pyramid
representation [5], with the step (ii) above replaced by our probabilistic fusion. A
similar probabilistic framework for sensor fusion is discussed in ([10]).
2
The lInage Forlnation Model
The true scene, denoted s, gives rise to a sensor image through a noisy, non-linear
transformation. For ALG, s would be an image of the landing scene under conditions
of uniform lighting, unlimited visibility, and perfect sensors. We model the map from
the true scene to a sensor image by a noisy, locally affine transformation whose
parameters are allowed to vary across the image (actually across the Laplacian
pyramid)
ai(~ t)
= (3i(~ t) s(~ t) + O'i(~ t) + Ei(~ t)
(1)
r
where, s is the true scene, ai is ith sensor image, == (x, y, k) is the hyperpixel
location, with x, y the pixel coordinates and k the level of the pyramid, t is the
time, 0' is the sensor offset, {3 is the sensor gain (which includes the effects of local
polarity reversals and complementarity), and E is the (zero-mean) sensor noise. To
simplify notation, we adopt the matrix form
a = (3s
=
+ 0' + l
(2)
=
where a = [al,a2, . . . ,aqr, f3
[(31,(32, ... , (3qr, Q'
[0'1,0'2, ... ,O'qr, s is a
and we have dropped the reference to location and
scalar and l = [El,E2, ... ,E q
time.
r,
Since the image formation parameters f3, Q', and the sensor noise covariance E~ can
vary from hyperpixel to hyperpixel, the model can express local polarity reversals,
complementary features, spatial variation of sensor gain, and noise.
We do assume, however, that the image formation parameters and sensor noise
distribu tion vary slowly with location 1 . Hence, a particular set of parameters is
considered to hold true over a spatial region of several square hyperpixels. We will
use this assumption implicitly when we estimate these parameters from data.
The model (2) fits the framework of the factor analysis model in statistics [11,
12] . Here the hyperpixel values of the true scene s are the latent variables or
1 Specifically the parameters vary slowly on the spatia-temporal scales over which the
true scene s may exhibit large variations.
R. K. Sharma, T. K. Leen and M. Pavel
826
common factors, f3 contains the factor loadings, and the sensor noise ? values are
the independent factors. Estimation of the true scene is equivalent to estimating
the common factors from the observations a.
3
Bayesian Fusion
Given the sensor intensities a, we will estimate the true scene s by appeal to a
Bayesian framework. We assume that the probability density function of the latent
variables s is a Gaussian with local mean so(~ t) and local variance u;(~ t). An
attractive benefit of this setup is that the prior mean So might be obtained from
knowledge in the form of maps, or clear-weather images of the scene. Thus, such
database information can be folded into the sensor fusion in a natural way.
The density on the sensor images conditioned on the true scene, P(als), is normal
with mean f3 s+a and covariance E? :::: diag[u;l' U;2" .. ,u;J The marginal density
P(a) is normal with mean I'm :::: f3 So + a and covariance
C :::: E?
+ u;f3f3 T
(3)
Finally, the posterior density on s, given the sensor data a, P(sla) is also normal
with mean M- 1 (f3T E;l (a -a)+ so/u;), and covariance M- 1 :::: (f/ E;l f3+ l/u;fl.
Given these densities, there are two obvious candidates for probabilistic fusion :
maximum likelihood (ML) 5 :::: max. P(als), and maximum a posteriori (MAP)
5:::: max. P(sla) .
The MAP fusion estimate is simply the posterior mean
5::::
[f3TE;If3+1/u;r 1 (f3 TE ;l(a_a) + so/un
(4)
(5)
To obtain the ML fusion estimate we take the limit
u; -+
00
in either (4) or (5).
For both ML and MAP, the fused image 5 is a locally linear combination of the sensor
images that can, through the spatio-temporal variations in f3, a, and E?, properly
respond to changes in the sensor characteristics that tax averaging or selection
schemes. For example, if the second sensor has a polarity reversal relative to the
first, then f32 is negative and the two sensor contributions are properly subtracted.
If the first sensor has high noise (large u;J, its contribution to the fused image is
attenuated. Finally, a feature missing from sensor 1 corresponds to f31 :::: O. The
model compensates by accentuating the contribution from sensor 2.
4
Model Parameter Estimates
We need to estimate the local image formation model parameters a(~ t), f3(~ t) and
the local sensor noise covariance?E?(~ t). We estimate the latter from successive,
motion compensated video frames from each sensor. First we estimate the average
value at each hyperpixel (ai(t)), and the average square (a;(t)) by exponential
moving averages . We next estimate the noise variance by the difference
(t) ::::
a; (t) -
U;i
ai 2 (t).
To estimate f3 and a, we assume that f3, a, E?, So and u; are nearly constant
over small spatial regions (5 x 5 blocks) surrounding the hyperpixel for which the
Probabilistic Image Sensor Fusion
827
parameters are desired. Essentially we are invoking a spatial analog of ergodicity,
where ensemble averages are replaced by spatial averages, carried out locally over
regions in which the statistics are approximately constant.
To form a maximum likelihood (ML) estimate of a, we extremize the data loglikelihood C = Z=;;=llog[P(an)] with respect to a to obtain
a ML = I'a -
f3 so
(6)
,
where I'a is the data mean, computed over a 5 x 5 hyperpixellocal region (N
points).
= 25
To obtain a ML estimate of f3, we set the derivatives of C with respect to
to zero and recover
equal
(C - Ea)C
-1
f3 = 0
f3
(7)
where Ea is the data covariance matrix, also computed over a 5 x 5 hyperpixel local
region . The only non-trivial solution to (7) is
f3 ML
= E,!-(X-l)t
U
r
u~
(8)
where U , A are the principal eigenvector and eigenvalue of the weighted data co_
_1.
_1.
variance matrix, Ea == E, 2 EaE ?
and r = ?l.
2,
An alternative to maximum likelihood estimation is the least squares (LS) approach [11] . We obtain the LS estimate aLS by minimizing
(9)
with respect to a . This gives
aLS
The least squares estimate
f3 LS
f3 . The
(10)
is obtained by minimizing
E{3
with respect to
= I'a - f3 so .
= II Ea -
C
W
(11)
solution to this minimization is
f3LS =
At
-Ur
u~
(12)
where U, A are the principal eigenvector and eigenvalue of the noise-corrected covariance matrix (Ea - E f ), and r = ? l. 2
The estimation procedures cannot provide values of the priors u~ and So. Were we
dealing with a single global model, this would pose no problem. But we must impose
a constraint in order to smoothly piece together our local models. We impose that
11.811 1 everywhere, or by (12)
= A. Recall that A is the leading eigenvalue of
~a - ~, and thus captures the scale of variations in a that arise from variations in
s . Thus we would expect A ex u~. Our constraint insures that the proportionality
constant be the same for each local model. Next, note that changing So causes a shift
=
u;
2The least squares and maximum likelihood solutions are identical when the model is
exact Ea == C, i.e. the observed data covariance is exactly of the form dictated by the
model. Under this condition, U = (U TE;lU)-1/2Ee -1/2U and (~- 1) = ~(UTE;lU).
The LS and ML solutions are also identical when the noise covariance is homoscedastic
Ee = I, even if the model is not exact.
(1;
R. K. Sharma, T. K. Leen and M. Pavel
828
in s. To maintain consistency between local regions, we take So
These choices for 11'; and So constrain the parameter estimates to
f3 LS
r V
aLS
Pa .
= 0 everywhere.
and
(13)
In (5) 11'; and So are defined at each hyperpixel. However, to estimate f3 and a,
we used spatial averages to compute the sample mean and covariance. This is
somewhat inconsistent, since the spatial variation of So (e.g. when there are edges
in the scene) is not explicitly captured in the model mean and covariance. These
variations are, instead , attributed to 11';, resulting in overestimation of the latter.
A more complete model would explicitly model the spatial variations of So, though
we expect this will produce only minor changes in the results .
Finally, the sign parameter r is not specified. In order to properly piece together
our local models , we must choose r at each hyperpixel in such a way that f3 changes
direction slowly as we move from hyperpixel to hyperpixel and encounter changes
in the local image statistics. That is, large direction changes due to arbitrary sign
reversals are not allowed . We use a simple heuristic to accomplish this.
5
Relation to peA
The MAP and ML fusion rules are closely related to PCA. To see this, assume that
the noise is homoscedastic EE = 11';1 and use the parameter estimates (13) in the
MAP fusion rule (5), reducing the latter to
1
s= 1+I1'UI1';
T
Va(a-Pa)
1
+ 1+11';;11'~
So
(14)
where Va is the principal eigenvector of the data covariance matrix Ea. The MAP
estimate s is simply a scaled and shifted local PCA projection of the sensor data.
Both the scaling and shift arise because the prior distribution on s tends to bias s
towards So. When the prior is flat 11'; -+ 00, (or equivalently when using the ML
fusion estimate), or when the noise variance vanishes, the fused image is given by a
simple local PCA projection
(15 )
6
Experilllents and Results
We applied our fusion method to visible-band and IR runway images, Fig. 1, containing additive Gaussian noise. Fig. l(e) shows the result of ML fusion with f3
and a estimated using (13) . ML fusion performs better than either averaging or
selection in regions that contain local polarity reversals or complementary features.
ML fusion gives higher weight to IR in regions where the features in the two images are common , thus reducing the effects of noise in the visible-band image. ML
fusion gives higher weight to the appropriate sensor in regions with complementary
features.
Fig. l(f) shows the result of MAP fusion (5) with the priors 11'; and So those dictated
by th e consistency requirements discussed in section 4. Clearly, the MAP image is
less noisy than the ML image. In regions of low sensor image contrast, 11'; is low
(since>. is low), thus the contribution from the sensor images is attenuated compared
to the ML fusion rule. Hence the noise is attenuated. In regions containing features
such as edges, 11'; is high (since>. is high); hence the contribution from the sensor
images is similar to that in ML fusion.
Probabilistic Image Sensor Fusion
829
(a) Visible-band image
(b) IR image
(c) Averaging
(d) Selection
(e) ML
(f) MAP
Figure 1: Fusion of visible-band and IR images containing additive Gaussian noise
In Fig. 2 we demonstrate the use of a database image for fusion. Fig. 2(a) and 2(b)
are simulated noisy sensor images from visible-band and JR, that depict a runway
with an aircraft on it. Fig. 2(c) is an image of the same scene as might be obtained
from a terrain database. Although this image is clean, it does not show the actual
situation on the runway. One can use the database image pixel intensities as the
prior mean So in the MAP fusion rule (5). The prior variance u; in (5) can be
regarded as a m-e asure of confidence in the database image - it's value controls the
relative contribution of the sensors vs. the database image in the fused image. (The
parameters f3 and a, and the sensor noise covariance EIE were estimated exactly
as before.) Fig. 2(d), 2(e) and 2(f) show the MAP-fused image as a function of
increasing 0";. Higher values of 0"; accentuate the contribution of the sensor images,
whereas lower values of 0"; accentuate the contribution of the database.
7
Discussion
We presented a model-based probabilistic framework for fusion of images from multiple sensors and exercised the approach on visible-band and IR images. The approach
provides both a rigorous framework for PCA-like fusion rules, and a principled way
to combine information from a terrain database with sensor images.
We envision several refinements of the approach given here. Writing new image
formation models at each hyperpixel produces an overabundance of models. Early
experiments show that this can be relaxed by using the same model parameters over
regions of several square hyperpixels, rather than recalculating for each hyperpixel.
A further refinement could be provided by adopting a mixture of linear models to
build up the non-linear image formation model. Finally, we have used multiple
frames from a video sequence to obtain ML and MAP fused sequences, and one
should be able to produce superior parameter estimates by suitable use of the video
sequence.
R. K. Sharma, T. K. Leen and M Pavel
830
(a) Visible-band image
(b) IR image
(c) Database image
Figure 2: Fusion of simulated visible-band and IR images using database image
Acknowledgments - This work was supported by NASA Ames Research Center
grant NCC2-S11. TKL was partially supported by NSF grant ECS-9704094.
References
[1] L. A. Klein. Sensor and Data Fusion Concepts and Applications. SPIE, 1993.
[2] J. R. Kerr, D. P. Pond, and S. Inman. Infrared-optical muItisensor for autonomous
landing guidance. Proceedings of SPIE, 2463:38-45, 1995.
[3] B. Roberts and P. Symosek. Image processing for flight crew situation awareness.
Proceedings of SPIE, 2220:246-255, 1994.
[4] M. Pavel and R. K. Sharma. Model-based sensor fusion for aviation. In J. G. Verly,
editor, Enhanced and Synthetic Vision 1997, volume 3088, pages 169-176. SPIE, 1997.
[5] P. J. Burt and R. J. Kolczynski. Enhanced image capture through fusion. In Fourth
Int. Conf. on Computer Vision, pages 173-182. IEEE Compo Soc., 1993.
[6] H. Li and Y. Zhou. Automatic visual/IR image registration. Optical Engineering,
35(2):391-400, 1996.
'
[7] M . Pavel, J. Larimer, and A. Ahumada. Sensor fusion for synthetic vision. In Proceedings of the Society for Information Display, pages 475-478. SPIE, 1992.
[8] P. Burt. A gradient pyramid basis for pattern-selective image fusion. In Proceedings
of the Society for Information Display, pages 467-470. SPIE, 1992.
[9] A. Toet. Hierarchical image fusion. Machine Vision and Applications, 3:1-11, 1990.
[10] J. J. Clark and A. L. Yuille. Data Fusion for Sensory Information Processing Systems.
Kluwer, Boston, 1990.
[11] A. Basilevsky. Statistical Factor Analysis and Related Methods. Wiley, 1994.
[12] M. E. Tipping and C. M. Bishop. Probabilistic principal component analysis. Technical report, NCRG/97/01O, Neural Computing Research Group, Aston University,
UK,1997.
| 1516 |@word aircraft:3 f32:1 loading:1 proportionality:1 covariance:13 pavel:7 invoking:1 contains:1 efficacy:1 selecting:1 envision:1 must:2 additive:2 visible:11 visibility:2 depict:1 v:1 selected:1 device:1 ith:1 compo:1 provides:2 cse:1 location:3 successive:1 ames:1 ect:1 combine:2 actual:1 increasing:1 provided:1 estimating:1 notation:1 underlying:1 what:1 eigenvector:3 transformation:2 temporal:2 exactly:2 scaled:1 uk:1 control:1 grant:2 before:1 ncc2:1 engineering:3 local:19 todd:1 dropped:1 limit:1 tends:1 approximately:1 might:2 deployment:1 graduate:1 averaged:1 acknowledgment:1 block:1 procedure:1 weather:1 projection:2 confidence:1 cannot:1 selection:4 put:1 writing:1 landing:4 equivalent:1 map:14 missing:1 compensated:1 center:1 l:5 penetrate:1 rule:5 regarded:1 autonomous:2 coordinate:1 variation:8 enhanced:2 exact:2 us:2 runway:4 complementarity:1 synthesize:1 pa:2 infrared:3 database:10 observed:1 electrical:1 capture:2 region:12 highest:1 principled:1 vanishes:1 overestimation:1 radar:2 yuille:1 basis:2 surrounding:1 formation:7 firm:1 whose:1 heuristic:1 loglikelihood:1 compensates:1 statistic:4 transform:4 noisy:5 sequence:3 eigenvalue:3 multiresolution:1 tax:1 qr:2 requirement:1 produce:3 perfect:1 pose:2 minor:1 soc:1 crew:1 direction:2 closely:1 pea:1 decompose:1 probable:1 hold:1 considered:1 ground:1 normal:3 vary:4 adopt:1 a2:1 homoscedastic:2 early:1 estimation:3 sensitive:1 exercised:1 weighted:1 minimization:1 clearly:1 sensor:66 gaussian:3 rather:2 zhou:1 properly:3 portland:1 likelihood:6 contrast:4 rigorous:1 posteriori:2 el:1 relation:1 selective:1 i1:1 pixel:3 denoted:1 spatial:8 marginal:1 equal:1 tkl:1 f3:24 identical:2 nearly:1 report:2 simplify:1 individual:1 replaced:2 maintain:1 recalculating:1 mixture:1 misha:1 fog:1 edge:2 circle:1 desired:1 guidance:3 theoretical:1 fusing:1 f3t:1 uniform:1 accomplish:1 synthetic:2 density:5 probabilistic:10 together:2 fused:7 containing:3 choose:1 slowly:3 conf:1 derivative:1 leading:1 li:1 includes:1 coefficient:1 int:1 oregon:1 explicitly:3 piece:2 tion:1 recover:1 contribution:8 square:6 ir:12 variance:5 characteristic:2 ensemble:1 bayesian:3 produced:1 lu:2 lighting:1 eie:1 email:1 energy:1 obvious:2 e2:1 attributed:1 spie:6 gain:2 pilot:1 ask:1 recall:1 knowledge:1 actually:1 ea:7 nasa:1 higher:3 tipping:1 leen:4 box:1 though:1 ergodicity:1 flight:1 night:1 ei:1 effect:4 contain:2 true:13 concept:1 hence:3 attractive:1 ogi:2 complete:1 demonstrate:2 performs:1 motion:1 image:73 common:3 superior:1 accentuating:1 physical:1 volume:1 ncrg:1 discussed:2 analog:1 kluwer:1 ai:4 automatic:1 consistency:2 reliability:1 moving:1 ute:1 spatia:1 posterior:2 dictated:2 s11:1 captured:1 somewhat:1 impose:2 relaxed:1 sharma:4 signal:1 ii:3 multiple:5 full:1 reduces:1 technical:1 laplacian:3 va:2 vision:5 essentially:1 ui1:1 accentuate:2 adopting:1 pyramid:6 invert:1 whereas:2 inconsistent:1 ee:3 iii:1 fit:1 attenuated:3 shift:2 motivated:1 pca:4 cause:1 clear:1 involve:1 locally:4 band:11 extremize:1 nsf:1 shifted:1 sign:2 estimated:2 arising:1 klein:1 express:1 group:1 sla:2 changing:1 clean:1 ravi:2 registration:1 rectangle:1 imaging:2 fuse:1 everywhere:2 fourth:1 fueled:1 respond:1 scaling:1 tleen:1 fl:1 display:2 subj:1 constraint:2 constrain:1 scene:17 flat:1 unlimited:1 optical:2 rendered:1 department:2 combination:1 jr:1 across:2 ur:1 kerr:1 reversal:5 operation:1 hierarchical:1 appropriate:1 subtracted:1 alternative:1 encounter:1 build:1 society:2 move:1 pond:1 exhibit:1 gradient:1 reversed:3 simulated:2 trivial:1 polarity:7 ratio:1 minimizing:2 equivalently:1 setup:1 robert:1 negative:1 rise:1 observation:1 situation:2 frame:2 varied:1 arbitrary:1 intensity:3 burt:2 specified:2 able:1 pattern:1 including:1 max:2 video:3 suitable:2 difficulty:1 natural:1 scheme:1 technology:1 aston:1 carried:1 prior:7 relative:2 expect:2 limitation:1 clark:1 awareness:1 affine:1 editor:1 production:1 supported:2 distribu:1 bias:1 institute:1 benefit:1 f31:1 overcome:1 toet:1 sensory:1 refinement:2 ec:1 implicitly:1 dealing:1 ml:19 global:1 spatio:1 terrain:2 un:1 latent:2 alg:2 ahumada:1 diag:1 noise:21 arise:2 allowed:2 complementary:6 fig:10 wiley:1 exponential:1 candidate:1 bishop:1 showing:1 sensing:1 offset:1 appeal:1 fusion:42 consist:1 te:2 conditioned:1 boston:1 smoothly:1 simply:2 insures:1 visual:1 partially:1 scalar:1 corresponds:1 towards:1 change:5 specifically:1 folded:1 corrected:1 reducing:2 averaging:5 llog:1 aviation:1 principal:5 ece:1 latter:3 ex:1 |
566 | 1,517 | On the optimality of incremental neural network
algorithms
Ron Meir*
Department of Electrical Engineering
Technion, Haifa 32000, Israel
rmeir@dumbo.technion.ac.il
Vitaly Maiorov t
Department of Mathematics
Technion, Haifa 32000, Israel
maiorov@tx.technion.ac.il
Abstract
We study the approximation of functions by two-layer feedforward neural networks, focusing on incremental algorithms which greedily add
units, estimating single unit parameters at each stage . As opposed to
standard algorithms for fixed architectures, the optimization at each stage
is performed over a small number of parameters, mitigating many of the
difficult numerical problems inherent in high-dimensional non-linear optimization. We establish upper bounds on the error incurred by the algorithm, when approximating functions from the Sobolev class, thereby
extending previous results which only provided rates of convergence for
functions in certain convex hulls of functional spaces. By comparing our
results to recently derived lower bounds, we show that the greedy algorithms are nearly optimal. Combined with estimation error results for
greedy algorithms, a strong case can be made for this type of approach.
1 Introduction and background
A major problem in the application of neural networks to real world problems is the excessively long time required for training large networks of a fixed architecture. Moreover,
theoretical results establish the intractability of such training in the worst case [9][4]. Additionally, the problem of determining the architecture and size of the network required to
solve a certain task is left open. Due to these problems, several authors have considered
incremental algorithms for constructing the network by the addition of hidden units, and
estimation of each unit's parameters incrementally. These approaches possess two desirable attributes: first, the optimization is done step-wise, so that only a small number of
parameters need to be optimized at each stage; and second, the structure of the network
-This work was supported in part by the a grant from the Israel Science Foundation
tThe author was partially supported by the center for Absorption in Science, Ministry of Immigrant Absorption, State of Israel.
296
R. Meir and V Maiorov
is established concomitantly with the learning, rather than specifying it in advance. However, until recently these algorithms have been rather heuristic in nature, as no guaranteed
performance bounds had been established. Note that while there has been a recent surge
of interest in these types of algorithms, they in fact date back to work done in the early
seventies (see [3] for a historical survey).
The first theoretical result establishing performance bounds for incremental approximations
in Hilbert space, was given by Jones [8]. This work was later extended by Barron [2], and
applied to neural network approximation of functions characterized by certain conditions
on their Fourier coefficients. The work of Barron has been extended in two main directions. First, Lee et at. [10] have considered approximating general functions using Hilbert
space techniques, while Donahue et al. [7] have provided powerful extensions of Jones'
and Barron's results to general Banach spaces. One of the most impressive results of the
latter work is the demonstration that iterative algorithms can, in many cases, achieve nearly
optimal rates of convergence, when approximating convex hulls.
While this paper is concerned mainly with issues of approximation, we comment that it is
highly relevant to the statistical problem of learning from data in neural networks. First,
Lee et at. [10] give estimation error bounds for algorithms performing incremental optimization with respect to the training error. Under certain regularity conditions, they are
able to achieve rates of convergence comparable to those obtained by the much more computationally demanding algorithm of empirical error minimization. Moreover, it is well
known that upper bounds on the approximation error are needed in order to obtain performance bounds, both for parametric and nonparametric estimation, where the latter is
achieved using the method of complexity regularization. Finally, as pointed out by Donahue et al. [7], lower bounds on the approximation error are crucial in establishing worst
case speed limitations for learning.
The main contribution of this paper is as follows. For functions belonging to the Sobolev
class (see definition below), we establish, under appropriate conditions, near-optimal rates
of convergence for the incremental approach, and obtain explicit bounds on the parameter
values of the network. The latter bounds are often crucial for establishing estimation error
rates. In contrast to the work in [10] and [7], we characterize approximation rates for
functions belonging to standard smoothness classes, such as the Sobolev class. The former
work establishes rates of convergence with respect to the convex hulls of certain subsets
of functions, which do not relate in a any simple way to standard functional classes (such
as Lipschitz, Sobolev, Holder, etc.). As far as we are aware, the results reported here are
the first to report on such bounds for incremental neural network procedures. A detailed
version of this work, complete with the detailed proofs, is available in [13].
2 Problem statement
We make use of the nomenclature and definitions from [7]. Let H be a Banach space of
functions with norm II . II. For concreteness we assume henceforth that the norm is given
by the Lq norm, 1 < q < 00, denoted by II . Ilq. Let linn H consist of all sums of the form
L~=l aigi , gi E H and arbitrary ai, and COn H is the set of such sums with ai E [0,1] and
L~=l ai = 1. The distances, measured in the Lq norm, from a function f are given by
dist(1in n H,f) = inf {l lh - fllq : hE linnH},
dist(conH , f) = inf {l lh - fllq : hE conH}.
The linear span of H is given by linH = Un linn H, while the convex-hull of H is coH =
Unco n H. We follow standard notation and denote closures of sets by a bar, e.g. coH is the
closure of the convex hull of H. In this work we focus on the special case where
H = H1} ~ {g : g(x) = eCJ(a T x
+ b), lei::; 1}, IICJ(?)llq ::; I},
(1)
297
On the Optimality of Incremental Neural Network Algorithms
corresponding to the basic building blocks of multilayer neural networks. The restriction
110-011 ::; 1 is not very demanding as many sigmoidal functions can be expressed as a sum
of functions of bounded norm. It should be obvious that linn 1l1) corresponds to a two-layer
neural network with a linear output unit and o--activation functions in the single hidden
layer, while COn 1l1) is equivalent to a restricted form of such a network, where restrictions
are placed on the hidden-to-output weights. In terms of the definitions introduced above,
the by now well known property of universal function approximation over compacta can
be stated as lin1l = C(M), where C(M) is the class of continuous real valued functions
defined over M , a compact subset of Rd . A necessary and sufficient condition for this
has been established by Leshno et af. [11], and essentially requires that 0-(.) be locally
integrable and non-polynomial. We comment that if 'T} = 00 in (l), and c is unrestricted in
sign, then co1l= = lin1l=. The distinction becomes important only if 'T} < 00 , in which
case co1l1) C lin1l1).
For the purpose of incremental approximation, it turns out to be useful to consider the convex hull co1l, rather than the usual linear span, as powerful algorithms and performance
bounds can be developed in this case. In this context several authors have considered
bounds for the approximation of a function 1 belonging to co1l by sequences of functions
belonging to COn 1l . However, it is not clear in general how well convex hulls of bounded
functions approximate general functions . One contribution of this work is to show how
one may control the rate of growth of the bound 'T} in (1), so that general functions, belonging to certain smoothness classes (e.g. Sobolev), may be well approximated. In fact, we
show that the incremental approximation scheme described below achieves nearly optimal
approximation error for functions in the Sobolev space.
Following Donahue et af. [7], we consider c:-greedy algorithms. Let E = (El ' E2, ... ) be a
positive sequence, and similarly for (ai, a 2, ... ), 0 < an < 1. A sequence of functions
hI, h2 ' ... is E-greedy with respect to 1 if for n = 0, 1, 2, .. . ,
Ilhn+I - Illq< inf {llanhn + (1 -
an)g -
Illq :
9 E 1l1)}
+ En ,
(2)
where we set ho = O. For simplicity we set an = (n - l)/n , although other schemes are
also possible. It should be clear that at each stage n, the function h n belongs to COn 1l1).
Observe also that at each step, the infimum is taken with respect to 9 E 1l1)' the function
h n being fixed. In terms of neural networks, this implies that the optimization over each
hidden unit parameters (a, b, c) is performed independently of the others. We note in passing, that while this greatly facilitates the optimization process in practice, no theoretical
guarantee can be made as to the convexity of the single-node error function (see [1] for
counter-examples). The variables En are slack variables, allowing the extra freedom of
only approximate minimization . In this paper we do not optimize over an, but rather fix a
sequence in advance, forfeiting some generality at the price of a simpler presentation. In
any event, the rates we obtain are unchanged by such a restriction.
In the sequel we consider E-greedy approximations of smooth functions belonging to the
Sobolev class of function s,
W; =
{I : m?x WDk 1112 ::; ?1} ,
OS k S r
where k = (k 1 , . ?? , k d ) , k i 2:: 0 and Ikl = ki + ... k d . Here Vk is the partial derivative
operator of order k. All functions are defined over a compact domain K C Rd.
3
Upper bound for the L2 norm
First, we consider the approximation of functions from WI using the L2 norm. In distinction with other Lq norms, there exists an inner product in this case, defined through
R. Meir and V Maiorov
298
(".) = II?II~. This simp1ification is essential to the proof in this case.
We begin by recalling a result from [12], demonstrating that any function in L2 may be
exactly expressed as a convex integral representation of the form
f(x)
=Q
J
(3)
h(x, O)w(O)dO,
where 0 < Q < 00 depends on f, and w( 0) is a probability density function (pdf) with
respect to the multi-dimensional variable O. Thus, we may write f(x) = QEw{h(x, e)},
where Ew denotes the expectation operator with respect to the pdf w . Moreover, it was
shown in [12], using the Radon and wavelet transforms, that the function h(x, 0) can be
taken to be a ridge function with 0 = (a, b, e) and h(x, 0) = ea(a T x + b).
In the case of neural networks, this type of convex representation was first exploited by
Barron in [2], assuming f belongs to a class of functions characterized by certain moment
conditions on their Fourier transforms. Later, Delyon et al. [6] and Maiorov and Meir
[12] extended Barron's results to the case of wavelets and neural networks, respectively,
obtaining rates of convergence for functions in the Sobolev class.
The basic idea at this point is to generate an approximation, hn(x), based on n draws of
random variables en = {e l , e 2, ... ,en}, e i ,....., w(?), resulting in the random function
hn(x; en)
Q
=-
n
2: h(x, e
n
(4)
i ).
i=l
Throughout the paper we conform to standard notation, and denote random variables by
uppercase letters, as in e, and their realization by lower case letters, as in O. Let w n =
TI~=l Wi represent the product pdf for {e l , ... ,en}. Our first result demonstrates that,
on the average, the above procedure leads to good approximation of functions belonging to
W{.
Theorem 3.1 Let K C Rd be a compact set. Then/or any
there exists a constant e > 0, such that
Ew "
Ilf -
f
hn(x; en)112 :S en- rld +E ,
E W{, n
>
0 and c
>
0
(5)
where Q < en(1/ 2- r l d)+, and (x)+ = max(O,x).
The implication of the upper bound on the expected value, is that there exists a set of
values o* ,n = {Or, ... , O~}, for which the rate (5) can be achieved. Moreover, as long as
the functions h(x, Od in (4) are bounded in the L2 norm, a bound on Q implies a bound on
the size of the function h n itself.
Proof sketch The proof proceeds by expressing f as the sum of two functions, iI
and 12 . The function iI is the best approximation to f from the class of multi-variate
splines of degree r. From [12] we know that there exist parameters on such that
IliI (.) - h n {-, on)112 :S en-rid. Moreover, using the results of [5] it can be shown that
1112112 :S en-rid. Using these two observations, together with the triangle inequality
Ilf - h n l1 2:S IliI - h nl1 2+ 1112112, yields the desired result.
I
Next, we show that given the approximation rates attained in Theorem 3.1, the same rates
may be obtained using an c -greedy algorithm. Moreover, since in [12] we have established
the optimality of the upper bound (up to a logarithmic factor in n), we conclude that greedy
approximations can indeed yield near-optimal perfonnance, while at the same time being
much more attractive computationally. In fact, in this section we use a weaker algorithm,
which does not perform a full minimization at each stage.
On the Optimality of Incremental Neural Network Algorithms
Incremental algorithm: (q = 2) Let an
1. Let 0i be chosen to satisfy
=1-
lin, 6: n
299
=1-
an = lin.
Ilf(x) - Qh(x,Onll~ = EWl {llf(x) - Qh(x,edIID?
2. Assume that 0i ,
f(x) -
?2,...
::~;
,O~-l
have been generated. Select O~ to obey
n-l
2
i=l
2
L h(x,On - 6:nQh(x,O~)
f(x) - Qa n n-l
"h(x, On - 6: n Qh(x, en)
22} .
n-lL..,;
i=l
Define
which measures the error incurred at the n-th stage by this incremental procedure. The
main result of this section then follows.
Theorem 3.2 For any
bounded as
f
E
WI and c > 0, the error of the incremental algorithm above is
for some finite constant c.
Proof sketch The claim will be established upon showing that
(6)
namely, the error incurred by the incremental procedure is identical to that of the nonincremental one described preceding Theorem 3.1. The result will then follow upon using
Holder's inequality and the upper bound (5) for the r.h.s. of (6). The remaining details are
I
straightforward, but tedious, and can be found in the full paper [13].
4
Upper bound for general Lq norms
Having established rates of convergence for incremental approximation of WI in the L2
norm, we move now to general Lq norms. First, note that the proof of Theorem 3.2 relies
heavily on the existence on an inner product. This useful tool is no longer available in the
case of general Banach spaces such as L q . In order to extend the results to the latter norm,
we need to use more advanced ideas from the theory of the geometry of Banach spaces. In
particular, we will make use of recent results from the work of Donahue et al. [7]. Second,
we must keep in mind that the approximation of the Sobolev space WI using the Lq norm
only makes sense if the embedding condition rid> (1/2 - l/q)+ holds, since otherwise
the Lq norm may be infinite (the embedding condition guarantees its finiteness; see [14]
for details).
We first present the main result of this section, followed by a sketch of the proof. The full
details of the rather technical proof can be found in [13]. Note that in this case we need to
use the greedy algorithm (2) rather than the algorithm of Section 3.
R. Meir and V Maiorov
300
Theorem 4.1 Let the embedding condition r / d
0< r < r*, r* =
JEW;
andf.
~ + (~- ~)+
>
(1/2 - 1/ q) + hold for 1 < q <
00,
andassumethatllh(-,O)llq:S IforallO. Thenforany
>0
where
;I- (~-P
!+% _qd
q > 2,
(7)
~
q :s 2,
c = c(r, d , K) and h n(-, on) is obtained via the incremental greedy algorithm (2) with
cn = O.
, =
{
Proof sketch The main idea in the proof of Theorem 4.1 is a two-part approximation
scheme. First, based on [13], we show that any JEW; may be well approximated by
functions in the convex class COn ('111/) for an appropriate value of TJ (see Lemma 5.2 in
[13]), where R1/ is defined in (1). Then, it is argued, making use of results from [7] (in
particular, using Corollary 3.6) , that an incremental greedy algorithm can be used to approximate the closure of the class co( R 1/) by the class COn (111/). The proof is completed by
using the triangle inequality. The proof along the above lines is done for the case q > 2. In
the case q 2, a simple use of the Holder inequality in the form Ilfllq ~ IKI 1 /q- l/21IfI12,
where IKI is the volume of the region K, yields the desired result, which, given the lower
I
bounds in [12], is nearly optimal.
:s
5
Discussion
We have presented a theoretical analysis of an increasingly popular approach to incremental
learning in neural networks . Extending previous results , we have shown that near-optimal
rates of convergence may be obtained for approximating functions in the Sobolev class
These results extend and clarify previous work dealing solely with the approximation
of functions belonging to the closure of convex hulls of certain sets of functions. Moreover,
we have given explicit bounds on the parameters used in the algorithm , and shown that the
restriction to COn 111/ is not too stringent. In the case q ~ 2 the rates obtained are as good (up
to logarithmic factors) to the rates obtained for general spline functions, which are known
to be optimal for approximating Sobolev spaces. The rates obtained in the case q > 2
are sub-optimal as compared to spline functions, but can be shown to be provably better
than any linear approach. In any event, we have shown that the rates obtained are equal,
up to logarithmic factors, to approximation from linn 111/' when the size of TJ is chosen
appropriately, implying that positive input-to-output weights suffice for approximation. An
open problem remaining at this point is to demonstrate whether incremental algorithms for
neural network construction can be shown to be optimal for every value of q. In fact, this
is not even known at this stage for neural network approximation in general.
W; .
References
[1] P. Auer, M. Herbster, and M. Warmuth. Exponentially many local minima for single
neurons. In D.S. Touretzky, M .e. Mozer, and M.E. Hasselmo, editors, Advances in
Neural Information Processing Systems 8, pages 316-322. MIT Press, 1996.
[2] AR. Barron. Universal approximation bound for superpositions of a sigmoidal function. IEEE Trans. In! Th., 39:930-945, 1993.
[3] AR. Barron and R.L. Barron. Statistical learning networks: a unifying view. In
E. Wegman , editor, Computing Science and Statistics: Proceedings 20th Symposium
Interface , pages 192-203, Washington D.e. , 1988. Amer. Statis. Assoc.
On the Optimality of Incremental Neural Network Algorithms
301
[4] A. Blum and R. Rivest. Training a 3-node neural net is np-complete. In D.S. Touretzky, editor, Advances in Neural Information Processing Systems I, pages 494-50l.
Morgan Kaufmann, 1989.
[5] c. de Boor and G. Fix. Spline approximation by quasi-interpolation. J. Approx.
Theory , 7:19-45,1973 .
[6] B. Delyon, A. Juditsky, and A. Benveniste. Accuracy nalysis for wavelet approximations. IEEE Transaction on Neural Networks, 6:332-348, 1995.
[7] M .J. Donahue, L. Gurvits, C. Darken, and E. Sontag. Rates of convex approximation
in non-hilbert spaces. Constructive Approx. , 13: 187-220, 1997.
[8] L. Jones. A simple lemma on greedy approximation in Hilbert space and convergence rate for projection pursuit regression and neural network training. Ann. Statis. ,
20:608-613, 1992.
[9] S. Judd. Neural Network Design and the Complexity of Learning. MIT Press, Boston,
USA,1990.
[10] W.S . Lee, P.S. Bartlett, and R.c. Williamson. Efficient Agnostic learning of neural
networks with bounded fan-in . IEEE Trans. In! Theory, 42(6):2118-2132, 1996.
[11] M. Leshno, V. Lin, A. Pinkus, and S. Schocken. Multilayer Feedforward Networks
with a Nonpolynomial Activation Function Can Approximate any Function. Neural
Networks, 6:861-867,1993.
[12] V.E. Maiorov and R. Meir. On the near optimality of the stochastic approximation of
smooth functions by neural networks. Technical Report CC-223, Technion, Department of Electrical Engineering, November ]997. Submitted to Advances in Computational Mathematics.
[13] R. Meir and V. Maiorov.
On the optimality of neural network approximaSubmitted for publication, October 1998.
tion using incremental algorithms.
ftp://dumbo.technion.ac.il/pub/PAPERSlincrementa] .pdf.
[14] H. Triebel. Theory of Function Spaces. Birkhauser, Basel, 1983.
| 1517 |@word version:1 polynomial:1 norm:15 tedious:1 open:2 closure:4 iki:2 thereby:1 moment:1 pub:1 comparing:1 od:1 activation:2 must:1 numerical:1 statis:2 ilii:2 juditsky:1 implying:1 greedy:11 warmuth:1 node:2 ron:1 forfeiting:1 sigmoidal:2 simpler:1 along:1 symposium:1 boor:1 indeed:1 expected:1 surge:1 dist:2 multi:2 leshno:2 becomes:1 provided:2 estimating:1 moreover:7 notation:2 bounded:5 begin:1 suffice:1 rivest:1 israel:4 agnostic:1 developed:1 guarantee:2 every:1 ti:1 growth:1 exactly:1 demonstrates:1 assoc:1 control:1 unit:6 grant:1 positive:2 engineering:2 local:1 establishing:3 solely:1 interpolation:1 specifying:1 co:1 practice:1 block:1 procedure:4 universal:2 empirical:1 projection:1 operator:2 context:1 ilf:3 restriction:4 equivalent:1 optimize:1 center:1 straightforward:1 independently:1 convex:12 survey:1 simplicity:1 embedding:3 qh:3 construction:1 heavily:1 ilq:1 approximated:2 electrical:2 worst:2 region:1 counter:1 mozer:1 convexity:1 complexity:2 upon:2 triangle:2 tx:1 wdk:1 heuristic:1 solve:1 valued:1 schocken:1 otherwise:1 statistic:1 gi:1 itself:1 maiorov:8 sequence:4 net:1 product:3 relevant:1 realization:1 date:1 achieve:2 ecj:1 convergence:9 regularity:1 extending:2 r1:1 incremental:22 ftp:1 ac:3 measured:1 strong:1 implies:2 qd:1 direction:1 attribute:1 hull:8 stochastic:1 stringent:1 argued:1 tthe:1 fix:2 absorption:2 extension:1 clarify:1 hold:2 considered:3 claim:1 major:1 achieves:1 early:1 purpose:1 estimation:5 superposition:1 hasselmo:1 establishes:1 tool:1 minimization:3 mit:2 rather:6 nqh:1 publication:1 corollary:1 derived:1 focus:1 vk:1 mainly:1 greatly:1 contrast:1 greedily:1 sense:1 el:1 hidden:4 quasi:1 mitigating:1 provably:1 issue:1 denoted:1 special:1 equal:1 aware:1 gurvits:1 having:1 washington:1 identical:1 seventy:1 jones:3 nearly:4 dumbo:2 report:2 others:1 spline:4 inherent:1 np:1 geometry:1 recalling:1 freedom:1 llq:2 interest:1 highly:1 uppercase:1 llf:1 tj:2 implication:1 integral:1 partial:1 necessary:1 lh:2 perfonnance:1 concomitantly:1 haifa:2 desired:2 theoretical:4 ar:2 subset:2 technion:6 too:1 characterize:1 reported:1 combined:1 density:1 herbster:1 sequel:1 lee:3 together:1 opposed:1 hn:3 henceforth:1 derivative:1 linh:1 de:1 coefficient:1 satisfy:1 depends:1 performed:2 later:2 h1:1 view:1 tion:1 contribution:2 il:3 holder:3 accuracy:1 kaufmann:1 yield:3 cc:1 submitted:1 touretzky:2 definition:3 obvious:1 e2:1 proof:12 con:7 rld:1 popular:1 hilbert:4 ea:1 back:1 auer:1 focusing:1 attained:1 follow:2 amer:1 done:3 generality:1 stage:7 until:1 sketch:4 o:1 ikl:1 incrementally:1 infimum:1 lei:1 usa:1 building:1 excessively:1 former:1 regularization:1 attractive:1 ll:1 pdf:4 complete:2 ridge:1 demonstrate:1 l1:6 interface:1 wise:1 recently:2 functional:2 exponentially:1 pinkus:1 banach:4 extend:2 he:2 volume:1 expressing:1 nonpolynomial:1 ai:4 smoothness:2 rd:3 approx:2 mathematics:2 similarly:1 pointed:1 had:1 impressive:1 longer:1 etc:1 add:1 recent:2 inf:3 belongs:2 certain:8 inequality:4 exploited:1 integrable:1 morgan:1 ministry:1 unrestricted:1 minimum:1 preceding:1 ewl:1 ii:7 full:3 desirable:1 smooth:2 technical:2 characterized:2 af:2 long:2 lin:3 basic:2 regression:1 multilayer:2 essentially:1 expectation:1 represent:1 achieved:2 background:1 addition:1 finiteness:1 crucial:2 appropriately:1 extra:1 posse:1 comment:2 facilitates:1 vitaly:1 near:4 feedforward:2 concerned:1 variate:1 architecture:3 inner:2 idea:3 cn:1 triebel:1 whether:1 bartlett:1 sontag:1 nomenclature:1 passing:1 useful:2 detailed:2 clear:2 transforms:2 nonparametric:1 locally:1 generate:1 meir:7 exist:1 sign:1 rmeir:1 conform:1 write:1 demonstrating:1 blum:1 linn:4 nalysis:1 concreteness:1 sum:4 letter:2 powerful:2 throughout:1 sobolev:11 draw:1 radon:1 comparable:1 layer:3 bound:24 hi:1 guaranteed:1 ki:1 followed:1 fan:1 fourier:2 speed:1 optimality:7 span:2 performing:1 jew:2 coh:2 department:3 belonging:8 increasingly:1 wi:5 making:1 restricted:1 taken:2 computationally:2 turn:1 slack:1 needed:1 know:1 mind:1 available:2 pursuit:1 observe:1 obey:1 barron:8 appropriate:2 ho:1 existence:1 denotes:1 remaining:2 completed:1 unifying:1 establish:3 approximating:5 unchanged:1 move:1 parametric:1 usual:1 distance:1 assuming:1 demonstration:1 difficult:1 october:1 statement:1 relate:1 stated:1 design:1 basel:1 perform:1 allowing:1 upper:7 observation:1 neuron:1 darken:1 finite:1 november:1 wegman:1 extended:3 arbitrary:1 introduced:1 immigrant:1 namely:1 required:2 optimized:1 distinction:2 established:6 qa:1 trans:2 able:1 bar:1 proceeds:1 below:2 max:1 event:2 demanding:2 advanced:1 scheme:3 l2:5 determining:1 limitation:1 foundation:1 h2:1 incurred:3 degree:1 sufficient:1 nonincremental:1 editor:3 intractability:1 benveniste:1 supported:2 placed:1 weaker:1 judd:1 world:1 author:3 made:2 historical:1 far:1 transaction:1 compacta:1 approximate:4 compact:3 keep:1 dealing:1 rid:3 conclude:1 un:1 iterative:1 continuous:1 additionally:1 nature:1 obtaining:1 williamson:1 constructing:1 domain:1 main:5 en:12 sub:1 explicit:2 lq:7 wavelet:3 donahue:5 theorem:7 showing:1 consist:1 exists:3 essential:1 delyon:2 boston:1 illq:2 logarithmic:3 expressed:2 partially:1 corresponds:1 relies:1 presentation:1 ann:1 lipschitz:1 price:1 infinite:1 birkhauser:1 lemma:2 ew:2 select:1 latter:4 constructive:1 andf:1 |
567 | 1,518 | Contrast adaptation in simple cells by changing
the transmitter release probability
Peter Adorjan
Klaus Obennayer
Dept. of Computer Science, FR2-1, Technical University Berlin
Franklinstrasse 28/2910587 Berlin, Germany
{adp, oby} @cs.tu-berlin.de http://www.ni.cs.tu-berlin.de
Abstract
The contrast response function (CRF) of many neurons in the primary visual cortex saturates and shifts towards higher contrast values following
prolonged presentation of high contrast visual stimuli. Using a recurrent
neural network of excitatory spiking neurons with adapting synapses we
show that both effects could be explained by a fast and a slow component in the synaptic adaptation. (i) Fast synaptic depression leads to saturation of the CRF and phase advance in the cortical response to high
contrast stimuli. (ii) Slow adaptation of the synaptic transmitter release
probability is derived such that the mutual information between the input
and the output of a cortical neuron is maximal. This component-given
by infomax learning rule-explains contrast adaptation of the averaged
membrane potential (DC component) as well as the surprising experimental result, that the stimulus modulated component (Fl component)
of a cortical cell's membrane potential adapts only weakly. Based on our
results, we propose a new experiment to estimate the strength of the effective excitatory feedback to a cortical neuron, and we also suggest a
relatively simple experimental test to justify our hypothesized synaptic
mechanism for contrast adaptation.
1
Introduction
Cells in the primary visual cortex have to encode a wide range of contrast levels, and they
still need to be sensitive to small changes in the input intensities. Because the signaling
capacity is limited, this paradox can be resolved only by a dynamic adaptation to changes
in the input intensity distribution: the contrast response function (CRF) of many neurons
in the primary visual cortex shifts towards higher contrast values following prolonged presentation of high contrast visual stimuli (Ahmed et al. 1997, Carandini & Ferster 1997).
On the one hand, recent experiments, suggest that synaptic plasticity has a major role
77
Contrast Adaptation and Infomax
in contrast adaptation. Because local application of GABA does not mediate adaptation
(Vidyasagar 1990) and the membrane conductance does not increase significantly during
adaptation (Ahmed et al. 1997, Carandini & Ferster 1997), lateral inhibition is unlikely to
account for contrast adaptation. In contrast, blocking glutamate (excitatory) autoreceptors
decreases the degree of adaptation (McLean & Palmer 1996). Furthermore, the adaptation
is stimulus specific (e.g. Carandini et al. 1998), it is strongest if the adapting and testing
stimuli are the same. On the other hand, plasticity of synaptic weights (e.g. Chance et
al. 1998) cannot explain the weak adaptation of the stimulus driven modulations in the
membrane potential (FI component) (Carandini & Ferster 1997) and the retardation of the
response phase after high contrast adaptation (Saul 1995). These experimental findings
motivated us to explore how presynaptic factors, such as a long term plasticity mediated
by changes in the transmitter release probability (Finlayson & Cynader 1995) affect contrast adaptation.
2
The single cell and the cortical circuit model
The cortical cells are modeled as leaky integrators with a firing threshold of -55 mY. The
interspike membrane potential dynamics is described by
8Vi(t)
Cm~ = -91eak
(Vi(t) -
Eresd -
~
L9ij(t) (Vi(t) -
Esyn) .
(I)
J
The postsynaptic conductance 9ij (t) is the integral over the previous presynaptic events
and is described by the alpha-function
(2)
t;
where is the arrival time of spike number s from neuron j. Including short term synaptic
depressIOn, the effective conductance is weighted by the portion of the synaptic resource
Pij (t) . Rij (t) that targets the postsynaptic side. The model parameters are C m = 0.5 nF,
91e ak = 31 nS, E rest = -65 mY, Esyn = -5 mY, 9'::':~x = 7.8 nS, and Tpeak = 1 ms,
and the absolute refractory period is 2 ms, and after a spike, the membrane potential is reset 1 mV below the resting potential. Following Tsodyks & Markram (1997) a synapse
between neurons j and i is characterized by the relative portion of the available synaptic
transmitter or resource Rij. After a presynaptic event, Rij decreases by Pij Rij, and recovers exponentially, where Pij is the transmitter release probability. The time evolution
of Rij between two presynaptic spikes is then
Rij(t) = 1 - (1 - (Rij{t) - pij(t)Rij{t))) exp
A
?
(-(t -i))
,
(3)
Tree
where ? is the last spike time, and the recovery time constant Tree = 200 ms. Assuming
Poisson distributed presynaptic firing, the steady state of the expected resource is
(4)
The stationary mean excitatory postsynaptic current (EPSC) Ii] (fj , Pij) is proportional to
the presynaptic firing frequency fj and the activated transmitter Pij Ri] (fj , Pij )
Ii] (fj, Pij) ex f Pij Ri] (fj, Pij) .
(5)
The mean current saturates for high input rates /j and it also depends on the transmitter
release probability Pij: with a high release probability the function is steeper at low presynaptic frequencies but saturates earlier than for a low release probability.
P Adorjem and K. Obermayer
78
_
g
..
?
80 r = = = = - '- -- - - ,
--- p=O.55
,'.
60 - p=O.24
,A
.?
,,/"
e 40
~.
OD
J!" 20
#
-641.. - - - --.====:::::::;-]
:'
--- p = 0.55
- -64.2 ,
'\ .?
>
p= 02
.4
S -64.4 : \ ;\
.... --
0 0
~ -64.6'
0
t-"
.........
"
0
: \ :\.
~
,-
(5
~
0
...
..
-64.8
?
OF'='-4-~o~~-----~
10?
(a)
10 '
Firing rate [Hz 1
-{)5 0
102
(b)
50
100
150
200
Time [ms[
Figure 1: Short term synaptic dynamics at high and low transmitter release probability,
(a) The estimated transfer function O(f, p) for the cortical cells (Eq. 7) (solid and dashed
lines) in comparison with data obtained by the integrate and fire model (Eq. 1, circles and
asterisks). (b) EPSP trains for a series of presynaptic spikes at intervals of 31 ms (32 Hz).
p=O.55 (0.24) corresponds to adaptation to 1% (50% ) contrast (see Section 4).
In order to study contrast adaptation. 30 leaky-integrator neurons are connected fully via
excitatory fast adapting synapses. Each "cortical" leaky integrator neuron receives its
"geniculate" input through 30 synapses. The presynaptic geniculate spike-trains are independent Poisson-processes. Modeling visual stimulation with a drifting grating, their rates
are modulated sinusoidally with a temporal frequency of2 Hz. The background activity for
each individual "geniculate" source is drawn from a Gaussian distribution with a mean of
20 Hz and a standard deviation of 5 Hz. In the model the mean geniculate firing rate (Fig.
2b) and the amplitude of modulation (Fig. 2a) increases with stimulus log contrast according to the experimental data (Kaplan et al. 1987). In the following simulations CRFs are
determined according to the protocol of Carandini & Ferster (1997). The CRFs are calculated using an initial adaptation period of 5 s and a subsequent series of interleaved test and
re-adaptation stimuli (1 s each).
3
The learning rule
We propose that contrast adaptation in a visual cortical cell is a result of its goal to maximize the amount of information the cell's output conveys about the geniculate input l . Following (Bell & Sejnowski 1995) we derive a learning rule for the transmitter release probability p to maximize the mutual information between a cortical cell's input and output. Let
O(f, p) be the average output firing rate, f the presynaptic firing rate, and p the synaptic
transmitter release probability. Maximizing the mutual information is then equivalent to
maximizing the entropy of a neuron's output if we assume only additive noise:
H [O(f,p)]
-E[ In Prob(O(J, p))]
Prob(f)]
[
-E In I(}O(f ,p)/(}f l
E [In 1(}O~;, p) I] - E[ In Prob(f)]
(6)
(In the following all equations apply locally to a synapse between neurons j and i.)
In order to derive an analytic expression for the relation between 0 and f we use the fact
that the EPSP amplitude converges to its steady state relatively fast compared to the modulation of the geniculate input to the visual cortex, and that the average firing rates of the
I A different approach of maximizing mutual information between input and output of a single
spiking neuron has been developed by Stemmler & Koch (1999). For non-spiking neurons this strategy has been demonstrated experimentally by. e.g. Laughlin (1994).
Contrast Adaptation and Infomax
79
presynaptic neurons are approximately similar. Thus we approximate the activation function by
Ire
O(f,p)
ex
S(f)pRoo (f ,p),
(7)
where S(f) =
accounts for the frequency dependent summation of EPSCs. The
parameters a = 1.8 and
= 15 Hz are determined by fitting O(f, p) to the firing rate
of our integrate and fire single cell model (see Fig. 1a) . The objective function is then
maximized by a stochastic gradient ascent learning rule for the release probability p
e
op _ oH [O(f, p)] _ ~ 1
Tadapt
ot -
Op
- Op n
IoO(f, p) I
of
.
(8)
Evaluating the derivatives we obtain a non-Hebbian learning rule for the transmitter release
probability p,
op
Tadapt
= }-
ot
-
2 Tre e fR
1
Tree(fa - 1)
- 1)
+ p- + a + TreeP(fa
(9)
=
where a
I~e' and the adaptation time constant Tadapt
7 s (Ohzawa et al. 1985).
This is similar in spirit to the anti-Hebbian learning mechanism for the synaptic strength
proposed by Barlow & Foldiak (1989) to explain adaptation phenomena. Here, the first
term is proportional to the presynaptic firing rate f and to the available synaptic resource
R, suggesting a presynaptic mechanism for the learning. Because the amplitude of the
EPSP is proportional to the available synaptic resource, we could interpret R as an output
related quantity and -2Tree f R as an anti-Hebbian learning rule for the "strength of the
synapse", i.e. the probability p of the transmitter release. The second term ensures that pis
always larger than O. In the current model setup for the operating range of the presynaptic
geniculate cells p also stays al ways less than 1. The third term modulates the adaptation
slightly and increases the release probability p most if the input firing rate is close to 20 Hz,
i.e. the stimulus contrast is low.
Image contrast is related to the standard deviation of the luminance levels normalized by
the mean . Because ganglion cells adapt to the mean luminance, contrast adaptation in the
primary visual cortex requires only the estimation of the standard deviation. In a free viewing scenario with an eye saccade frequency of 2-3 Hz, the standard deviation can be estimated based on 10-20 image samples. Thus the adaptation rate can be fast (Tadapt = 7 s),
and it should also be fast in order to maintain good a representation whenever visual contrast changes, e.g. by changing light conditions. Higher order moments (than the standard
deviation) of the statistics of the visual world express image structure and are represented
by the receptive fields' profiles. The statistics of the visual environment are relatively
static, thus the receptive field profiles should be determined and constrained by another
less plastic synaptic parameter. such as the maximal synaptic conductance 9max.
4
Results
Figure 2 shows the average geniculate input, the membrane potential, the firing rate and the
response phase of the modeled cortical cells as a function of stimulus contrast. The CRFs
were calculated for two adapting contrasts 1% (dashed line) and 50% (solid line). The
cortical CRF saturates for high contrast stimuli (Fig. 2e). This is due to the saturation ofthe
postsynaptic current (cf. Fig. I a) and thus induced by the short term synaptic depression. In
accordance with the experimental data (e.g. Carandini et al. 1997) the delay of the cortical
response (Fig. 2f) decreases towards high contrast stimuli. This is a consequence of fast
synaptic depression (c.f. Chance et al. 1998). High modulation in the input firing rate
leads to a fast transient rise in the EPSC followed by a rapid depression.
80
P. Ador}an and K. Obermayer
LGN
>E
<L)
--a
fJj
~
~
....
~
~(f
5
~
0.0
....0
<L)
~
0
100
U' 40
102
.----~---,
<L)
102
~20
~-58
10 1
Contrast [%]
102
~..o--_
,P.-El~-~
101
102
-0
0- 62 L--_~
100
101
~
~
ro
_ _- - '
Contrast [%]
,
I
-40
o
I
~ -60
0..
(d)
p--e'
~-20
'0
U
,,- -
o
~
<c1o
u
o 0 l...-_~~_~
,,
,,
(e) 10
.......
>-54
~30
.S
oa
~
(c)
.s
100
~
101
,,
10
.S
......
t:l..
10 1
fJj
(b)
--~ 20
,,
C
0.0
.S
(a)
<L)
fJj
.......
:-20
u::
,......,
30
(,)
....... 10
U'40
102
-80 l...-_~~_---'
100
10 1
102
(0
Contrast [%]
Figure 2: The DC (a) and the FI (b) component of the geniculate input, and the response of
the cortical units in the model with strong recurrent lateral connections and slow adaptation
of the release probability on both the geniculocortical and lateral synapses. The Fl (c)
and the DC (d) component of the subthreshold membrane potential of a single cortical
unit, the Fl component of the firing rate (e). and the response phase (0 are plotted as a
function of stimulus contrast after adaptation to 1% (solid lines) and to 50% (dashed lines)
contrast stimuli. The CRF for the membrane potential (c, d) is calculated by integrating
Eg. I without spikes and without reset after spikes. The cortical circuitry involves strong
recurrent lateral connections.
The model predicts a shift of 3-5 mV in the DC component of the subthreshold membrane
potential (Fig. 2d)- a smaller amount than measured by Carandini & Ferster (1997). Nevertheless, in accordance with the data the shift caused by the adaptation is larger than the
change in the DC component of the membrane potential from I % contrast to 100% contrast. The largest shift in the DC membrane potential during adaptation occurs for small
contrast stimuli because an alteration in the transmitter release probability has the largest
effect on the postsynaptic current if the presynaptic firing rate is close to the geniculate
background activity of 20 Hz. The maximal change in the Fl component (Fig. 2c) is
around 5mV and it is half of the increase in the FI component of the membrane potential from 1% contrast to 100% contrast. The CRF for the cortical firing rate (Fig. 2e) shifts
to the right and the slope decreases after adaptation to high contrast. The model predicts
that the probability p for the transmitter release decreases by approximately a factor of two.
The Fl component of the cortical firing rate decreases after adaptation because after tonic
decrease in the input modulated membrane potential, the over-threshold area of its FI component decreases. The adaptation in the Fl firing rate is fed back via the recurrent excitatory connections resulting in the observable adaptation in the FI membrane potential.
Without lateral feedback (Fig. 3) the Fl component of the membrane potential is basically
independent of the contrast adaptation. At high release probability a steep rise of the EPSC
to a high amplitude peak is followed by rapid depression if the input is increasing. At low
release probability the current increases slower to a lower amplitude, but the depression is
Contrast Adaptation and lnfomax
81
,......,
u
,......,
u 25
30
-aE 20
tU
tU
-.
<Jl
.?
20
.........
~
- 10
.-1Z
~
~
btl
btl
t:
t:
?c
0
102 (c) 10?
,......,
102
10 1
Contrast [%]
102
0
i-20
~
U
10 1
0
>-54
E
.........
o0..
5
0
(a) 10?
~
1Z
.-.5-58
....
15
10
'"0
.......
tU
~
......
-40
r..-...---..-__
-.l ~ -60
-o--a-o ___ .(). __
0-62
10?
(b)
Figure 3
~
10 1
Contrast [%]
102
-80
10?
(d)
10 1
Contrast [%]
102
-60 L--~~_~...-l
10?
10 1
102
(b)
Contrast [%]
Figure 4
Figure 3: The membrane potential (a, b), the phase (d) of the Fl component of the firing
rate, and the Fl component (c) averaged for the modeled cortical cells after adaptation to
1% (dashed lines) and 50% (solid lines) contrast. The weight of cortical connections is
set to zero. The CRF for the membrane potential (a, b) is calculated by integrating Eq. 1
without spikes and without reset after spikes.
Figure 4: Hysteresis curve revealed by following the ramp method protocol (Carandini &
Ferster 1997). After adaption to I % contrast, test stimuli of 2 s duration were applied with
a contrast successively increasing from 1% to 100% (asterisks). and then decreasing back
to 1% (circles).
less pronounced too. As a consequence, the power at the first harmonic (Fl component)
of the subthreshold membrane potential does not change if the release probability is modulated. It is modulated to a large extent by the recurrent excitatory feedback . The adaptation
of the Fl component of the firing rate could therefore be used to measure the effective
strength of the recurrent excitatory input to a simple ceIl in the primary visual cortex.
Additional simulations (data not shown) revealed that changing the transmitter release
probability of the geniculocortical synapses is responsible for the adaptation in our model
network. Fixing the value of p for the geniculocortical synapses abolishes contrast adaptation, while fixing the release probability p for the lateral synapses has no effect. Simulations show that increasing the release probability of the recurrent excitatory synapses leads
to oscillatory activity (e.g. Senn et al. 1996) without altering the mean activity of simple
cells. These results suggest an efficient functional segregation of feedforward and recurrent excitatory connections. Plasticity of the geniculocortical connections may playa key
role in contrast adaptation, while-without affecting the CRF-plasticity of the recurrent
excitatory synapses could could playa key role in dynamic feature binding and segregation
in the visual cortex (e.g. Engel et al. 1997).
Figure 4 shows the averaged CRF of the cortical model neurons revealed by the ramp
method (see figure caption) for strong recurrent feedback and adapting feedforward and
recurrent synapses. We find hysteresis curves for the Fl component of the firing rate simi-
82
P. Adorjan and K. Obermayer
lar to the results reported by Carandini & Ferster (1997), and for the response phase.
In summary, by assuming two different dynamics for a single synapse we explain the saturation of the CRFs, the contrast adaptation, and the increase in the delay of the cortical
response to low contrast stimuli. For the visual cortex of higher mammals, adaptation of
release probability p as a substrate for contrast adaptation is so far only a hypothesis. This
hypothesis, however, is in agreement with the currently available data, and could additionally be justified experimentally by intracellular measurements of EPSPs evoked by stimulating the geniculocortical axons. The model predicts that after adaptation to a low contrast
stimulus the amplitude of the EPSPs decreases steeply from a high value, while it shows
only small changes after adaptation to a high contrast stimulus (cf. Fig. 1b).
Acknowledgments The authors are grateful to Christian Piepenbrock for fruitful discussions. Funded by the German Science Foundation (Ob 102/2-1, GK 120-2).
References
Ahmed, B., Allison, J. D., Douglas, R. 1. & Martin, K A. C. (1997), 'Intracellular study of the
contrast-dependence of neuronal activity in cat visual cortex.' , Cerebral Cortex 7,559-570.
Barlow, H. B. & F6ldiak, P. (1989), Adaptation and decorrelation in the cortex, in R. Durbin, C. Miall
&
Mitchison, eds, 'The computing neuron', Workingham: Addison-Wesley, pp. 54-72.
c.
Bell, A. 1. & Sejnowski, T. J. (1995), 'An information-maximization approach to blind sepertation
and blind deconvolution', Neur. Comput. 7(6),1129-1159.
Carandini, M. & Ferster, D. (1997), 'A tonic hyperpolarization underlying contrast adaptation in cat
visual cortex', Science 276, 949-952.
Carandini, M., Heeger, D. J. & Movshon, 1. A. (1997), 'Linearity and normalization in simple cells
of the macaque primary visual cortex' , J. Neurosci. 17,8621-8644.
Carandini, M., Movshon, J. A. & Ferster, D. (1998), 'Pattern adaptation and cross-orientation interactions in the primary visual cortex' , Neuropharmacology 37, 501-51 I.
Chance, F. S., Nelson, S. B. & Abbott, L. F. (1998), 'Synaptic depression and the temporal response
characteristics of V I cells', 1. Neurosci. 18,4785-4799.
Engel, A. K, Roelfsema, P. R., Fries, P.. Brecht, M. & Singer. W (1997), 'Role of the temporal
domain for response selection and perceptual binding', Cerebral Cortex 7,571-582.
Finlayson, P. G. & Cynader, M. S. (I 995), 'Synaptic depression in visual cortex tissue slices: am in
vitro model for cortical neuron adaptation', Exp. Brain Res. 106, 145-155.
Kaplan, E., Purpura, K & Shapley, R. M. (1987), 'Contrast affects the transmission of visual information through the mammalian lateral geniculate nucleus', J. PhysioL. 391, 267-288.
Laughlin, S. B. (1994), 'Matching coding, circuits, cells, and molecules to signals: general principles
of retinal design in the fly's eye', Prog. Ret. Eye Res. 13, 165-196.
McLean, 1. & Palmer, L. A. (1996), 'Contrast adaptation and excitatory amino acid receptors in cat
striate cortex', Vis. Neurosci. 13, 1069-1087.
Ohzawa, I., ScJar, G . & Freeman , R. D . (1985), 'Contrast gain control in the cat's visual system', 1.
Neurophysiol. 54, 651-667.
Saul, A. B. (1995), 'Adaptation in single units in visual cortex: response timing is retarted by adapting', Vis. Neurosci. 12, 191-205.
Senn, W, Wyler, K, Streit, J., Larkum, M., Luscher, H.-R., H. Mey, L. M. a. D. S ., Vogt, K &
Wannier, T. (1996), 'Dynamics of a random neural network with synaptic depression', Neural
Networks 9, 575-588.
Stemmler, M. & Koch, C. (1999), Information maximization in single neurons, in 'Advances in Neural Information Processing Systems NIPS II'. same volume.
Tsodyks , M. V. & Markram, H. (1997), 'The neural code between neocortical pyramidal neurons
depends on neurotransmitter release probability', Proc. NatL. Acad. Sci. 94, 719-723.
Vidyasagar, T. R. (1990), 'Pattern adaptation in cat visual cortex is a co-operative phenomenon',
Neurosci. 36, 175-179.
| 1518 |@word vogt:1 simulation:3 mammal:1 solid:4 moment:1 initial:1 series:2 current:6 od:1 surprising:1 activation:1 physiol:1 additive:1 subsequent:1 interspike:1 plasticity:5 analytic:1 christian:1 piepenbrock:1 stationary:1 half:1 short:3 ire:1 fitting:1 shapley:1 wannier:1 expected:1 rapid:2 integrator:3 brain:1 freeman:1 decreasing:1 prolonged:2 increasing:3 underlying:1 linearity:1 circuit:2 cm:1 developed:1 ret:1 finding:1 temporal:3 nf:1 ro:1 control:1 unit:3 local:1 accordance:2 ceil:1 timing:1 consequence:2 acad:1 receptor:1 ak:1 firing:21 modulation:4 approximately:2 evoked:1 co:1 limited:1 palmer:2 range:2 averaged:3 acknowledgment:1 responsible:1 testing:1 signaling:1 area:1 bell:2 adapting:6 significantly:1 matching:1 integrating:2 suggest:3 cannot:1 close:2 selection:1 www:1 equivalent:1 fruitful:1 demonstrated:1 crfs:4 maximizing:3 duration:1 recovery:1 rule:6 oh:1 target:1 caption:1 substrate:1 hypothesis:2 agreement:1 fr2:1 mammalian:1 predicts:3 blocking:1 role:4 fly:1 rij:8 tsodyks:2 epsc:3 ensures:1 connected:1 decrease:9 environment:1 dynamic:6 weakly:1 grateful:1 cynader:2 neurophysiol:1 resolved:1 represented:1 cat:5 neurotransmitter:1 stemmler:2 train:2 fast:8 effective:3 sejnowski:2 oby:1 klaus:1 larger:2 ramp:2 statistic:2 propose:2 interaction:1 maximal:3 epsp:3 adaptation:54 reset:3 tu:5 fr:1 adapts:1 pronounced:1 transmission:1 converges:1 derive:2 recurrent:11 fixing:2 measured:1 ij:1 op:4 eq:3 strong:3 epsps:2 grating:1 c:2 involves:1 stochastic:1 viewing:1 transient:1 explains:1 summation:1 koch:2 around:1 exp:2 circuitry:1 major:1 estimation:1 proc:1 geniculate:11 currently:1 sensitive:1 largest:2 engel:2 weighted:1 gaussian:1 always:1 encode:1 release:25 derived:1 transmitter:15 steeply:1 contrast:62 am:1 dependent:1 el:1 unlikely:1 relation:1 lgn:1 germany:1 adorjan:2 orientation:1 constrained:1 mutual:4 field:2 stimulus:20 individual:1 phase:6 fire:2 maintain:1 conductance:4 allison:1 light:1 activated:1 natl:1 integral:1 tree:4 circle:2 re:3 plotted:1 earlier:1 modeling:1 sinusoidally:1 altering:1 maximization:2 deviation:5 delay:2 too:1 reported:1 my:3 peak:1 stay:1 infomax:3 successively:1 mclean:2 derivative:1 account:2 potential:19 suggesting:1 de:2 retinal:1 alteration:1 coding:1 hysteresis:2 caused:1 mv:3 vi:5 depends:2 blind:2 steeper:1 portion:2 slope:1 ni:1 acid:1 characteristic:1 maximized:1 subthreshold:3 ofthe:1 weak:1 plastic:1 basically:1 tissue:1 explain:3 synapsis:10 strongest:1 oscillatory:1 whenever:1 synaptic:21 ed:1 frequency:5 pp:1 conveys:1 recovers:1 static:1 gain:1 carandini:12 amplitude:6 back:2 wesley:1 higher:4 response:13 synapse:4 furthermore:1 hand:2 receives:1 tre:1 lar:1 effect:3 hypothesized:1 ohzawa:2 normalized:1 barlow:2 evolution:1 eg:1 during:2 steady:2 m:5 crf:9 neocortical:1 fj:5 image:3 harmonic:1 fi:5 stimulation:1 spiking:3 hyperpolarization:1 vitro:1 functional:1 refractory:1 exponentially:1 volume:1 jl:1 cerebral:2 adp:1 resting:1 interpret:1 neuropharmacology:1 measurement:1 of2:1 funded:1 cortex:19 operating:1 inhibition:1 playa:2 recent:1 foldiak:1 driven:1 scenario:1 additional:1 maximize:2 period:2 dashed:4 ii:4 signal:1 hebbian:3 technical:1 characterized:1 ahmed:3 adapt:1 long:1 cross:1 fjj:3 ae:1 poisson:2 normalization:1 cell:18 justified:1 background:2 affecting:1 operative:1 interval:1 pyramidal:1 source:1 ot:2 rest:1 ascent:1 hz:9 induced:1 spirit:1 revealed:3 feedforward:2 affect:2 simi:1 brecht:1 shift:6 motivated:1 expression:1 o0:1 movshon:2 peter:1 depression:10 amount:2 locally:1 http:1 senn:2 estimated:2 express:1 key:2 threshold:2 nevertheless:1 drawn:1 changing:3 douglas:1 abbott:1 btl:2 luminance:2 prob:3 prog:1 roelfsema:1 ob:1 interleaved:1 fl:12 followed:2 durbin:1 activity:5 strength:4 ri:2 relatively:3 martin:1 according:2 neur:1 gaba:1 membrane:19 smaller:1 slightly:1 postsynaptic:5 explained:1 resource:5 equation:1 segregation:2 german:1 mechanism:3 singer:1 addison:1 fed:1 available:4 apply:1 fry:1 slower:1 drifting:1 cf:2 retardation:1 objective:1 quantity:1 spike:10 occurs:1 strategy:1 primary:7 fa:2 receptive:2 dependence:1 striate:1 obermayer:3 gradient:1 berlin:4 capacity:1 lateral:7 oa:1 sci:1 nelson:1 presynaptic:15 extent:1 geniculocortical:5 finlayson:2 assuming:2 code:1 modeled:3 esyn:2 setup:1 steep:1 gk:1 obennayer:1 kaplan:2 rise:2 f6ldiak:1 design:1 neuron:19 anti:2 saturates:4 tonic:2 paradox:1 dc:6 intensity:2 connection:6 macaque:1 nip:1 below:1 pattern:2 saturation:3 including:1 max:1 vidyasagar:2 event:2 power:1 decorrelation:1 glutamate:1 eye:3 mediated:1 relative:1 fully:1 proportional:3 asterisk:2 foundation:1 integrate:2 nucleus:1 degree:1 pij:11 lnfomax:1 principle:1 pi:1 excitatory:12 summary:1 last:1 free:1 side:1 laughlin:2 wide:1 saul:2 markram:2 absolute:1 leaky:3 distributed:1 slice:1 feedback:4 calculated:4 cortical:23 evaluating:1 world:1 curve:2 author:1 far:1 miall:1 eak:1 alpha:1 approximate:1 observable:1 mitchison:1 purpura:1 additionally:1 transfer:1 molecule:1 ioo:1 protocol:2 domain:1 intracellular:2 neurosci:5 noise:1 mediate:1 arrival:1 profile:2 amino:1 neuronal:1 fig:11 slow:3 axon:1 n:2 heeger:1 comput:1 perceptual:1 third:1 c1o:1 specific:1 deconvolution:1 modulates:1 epscs:1 entropy:1 explore:1 ganglion:1 visual:24 saccade:1 binding:2 corresponds:1 chance:3 adaption:1 stimulating:1 goal:1 presentation:2 towards:3 ferster:9 change:8 experimentally:2 determined:3 justify:1 experimental:5 modulated:5 tpeak:1 dept:1 phenomenon:2 ex:2 |
568 | 1,519 | Controlling the Complexity of HMM Systems by
Regularization
Christoph Neukirchen, Gerhard Rigoll
Department of Computer Science
Gerhard-Mercator-University Duisburg
47057 Duisburg, Germany
email: {chn.rigoll}@fb9-ti.uni-duisburg.de
Abstract
This paper introduces a method for regularization ofHMM systems that
avoids parameter overfitting caused by insufficient training data. Regularization is done by augmenting the EM training method by a penalty
term that favors simple and smooth HMM systems. The penalty term
is constructed as a mixture model of negative exponential distributions
that is assumed to generate the state dependent emission probabilities of
the HMMs. This new method is the successful transfer of a well known
regularization approach in neural networks to the HMM domain and can
be interpreted as a generalization of traditional state-tying for HMM systems. The effect of regularization is demonstrated for continuous speech
recognition tasks by improving overfitted triphone models and by speaker
adaptation with limited training data.
1
Introduction
One general problem when constructing statistical pattern recognition systems is to ensure
the capability to generalize well, i.e. the system must be able to classify data that is not
contained in the training data set. Hence the classifier should learn the true underlying data
distribution instead of overfitting to the few data examples seen during system training.
One way to cope with the problem of overfitting is to balance the system's complexity and
flexibility against the limited amount of data that is available for training.
In the neural network community it is well known that the amount of information used in
system training that is required for a good generalization performance should be larger than
,the number of adjustable weights (Baum, 1989). A common method to train a large size
neural network sufficiently well is to reduce the number of adjustable parameters either
by removing those weights that seem to be less important (in (Ie Cun, 1990) the sensitivity of individual network weights is estimated by the second order gradient) or by sharing
C. Neukirchen and G. Rigoll
738
the weights among many network connections (in (Lang, 1990) the connections that share
identical weight values are determined in advance by using prior knowledge about invariances in the problem to be solved). A second approach to avoid overfitting in neural networks is to make use of regularization methods. Regularization adds an extra term to the
training objective function that penalizes network complexity. The simplest regularization
method is weight decay (Plaut, 1986) that assigns high penalties to large weights. A more
complex regularization term is used in soft weight-sharing (Nowlan, 1992) by favoring
neural network weights that fall into a finite set of small weight-clusters. The traditional
neural weight sharing technique can be interpreted as a special case of soft weight-sharing
regularization when the cluster variances tend towards zero.
In continuous speech recognition the Hidden Markov Model (HMM) method is common.
When using detailed context-dependent triphone HMMs, the number ofHMM-states and
parameters to estimate in the state-dependent probability density functions (pdfs) is increasingly large and overfitting becomes a serious problem. The most common approach
,to balance the complexity of triphone HMM systems against the training data set is to reduce the number of parameters by tying, i.e. parameter sharing (Young, 1992). A popular
sharing method is state-tying with selecting the HMM-states to be tied in advance, either
by data-driven state-clustering based on a pdf-dependent distance metric (Young, 1993),
or by constructing binary decision trees that incorporate higher phonetic knowledge (Bahl,
1991). In these methods, the number of state-clusters and the decision tree sizes, respectively, must be chosen adequately to match the training data size. However, a possible
drawback of both methods is that two different states may be selected to be tied (and their
pdfs are forced to be identical) although there is enough training data to estimate the different pdfs of both states sufficiently well. In the following, a method to reduce the complexity
of general HMM systems based on a regularization term is presented. Due to its close relationship to the soft weight-sharing method for neural networks this novel approach can be
interpreted as soft state-tying.
2
Maximum likelihood training in HMM systems
Traditionally, the method most commonly used to determine the set of adjustable parameters 8 in a HMM system is maximum likelihood (ML) estimation via the expectation
maximization (EM) algorithm. If the training observation vector sequence is denoted as
X = (x(l), ... ,x(T)) and the corresponding HMM is denoted as W the ML estimator is
given by:
{)ML = argmax {logpe(XIW)}
(1)
()
In the following, the total number of different HMM states is given by K. The emission pdf
,of the k-th state is denoted as bk (x); for continuous HMMs bk (x) is a mixture of Gaussian
pdfs most commonly; in the case of discrete HMMs the observation vector x is mapped by
a vector quantizer (VQ) on the discrete VQ-Iabel m(x) and the emission pdfis replaced by
the discrete output probability bk (m). By the forward-backward algorithm the probabilistic
state counts rdt) can be determined for each training observation and the log-likelihood
over the training data can be decomposed into the auxiliary function Q(8) optimized in
the EM steps (state transition probabilities are neglected here):
T
Q(8) =
K
L L rk(t) ?logbk(x(t))
(2)
t=l k=l
Sometimes, the observation vector x is split up into several independent streams. If the total
number of streams is given by Z, the features in the z-th stream comprise the subvector x(z)
and in the case of application ofa VQ the corresponding VQ label is denoted as m(z) (x(z?).
739
Controlling the Complexity ofHMM Systems by Regularization
The observation subvectors in different streams are assumed to be statistically independent
thus the states' pdfs can be written as:
Z
bk(x) =
II b~z)(x(z?)
(3)
z=l
3
A complexity measure for HMM systems
When using regularization methods to train the HMM system, the traditional objective
training function Q(0) is augmented by a complexity penalization term 0 and the new
optimization problem becomes:
(reg = argmax {Q(0)
(}
+ v? 0(0)}
(4)
Here, the regulizer term 0 should be small if the HMM system has high complexity and
parameter overfitting becomes a problem; 0 should be large if the HMM-states' pdfs are
shaped smoothly and system generalization works well. The constant v 2: 0 is a control
parameter that adjusts the tradeoff between the pure ML solution and the smoothness of
penalization. In Eqn. (4) the term Q(0) becomes larger the more data is used for training
(which makes the ML estimation become more reliable) and the influence of the term v? 0
gets less important, relatively.
The basic idea when constructing an expression for the regulizer 0 that favors smooth
HMM systems is, that in the case of simple and smooth systems the state-dependent emission pdfs bk (.) should fall into several groups of similar pdfs. This is in contrast to the
traditional state-tying that forces identical pdfs in each group. In the following, these clusters of similar emission pdfs are described by a probabilistic mixture model. Each pdf is
assumed to be generated by a mixture of I different mixture components Pi (. ). In this case
the probability (-density) of generating the emission pdf bk (.) is given by:
I
p(bkO) =
L
Ci . Pi(bk(?))
(5)
i=l
with the mixture weights Ci that are constrained to 0 ::::; Ci ::::; 1 and 1 = 2::=1 ci. The i-th
mixture component Pi (.) is used to model the i-th cluster of HMM-emission pdfs. Each
cluster is represented by a prototype pdf that is denoted as fJi (.) for the i-th cluster; the
distance (using a suitable metric) between a HMM emission pdf bk 0 and the i-th prototype
pdfis denoted as Di(bk (.)). If these distances are small for all HMM emission probabilities
there are several small clusters of emission probabilities and the regulizer term 0 should be
large. Now, it is assumed that the distances follow a negative exponential distribution (with
a deviation parameter Ai), yielding an expression for the mixture components:
p; (b.O) -
(g
A;,,) . exp ( -
~ A;"
. D;" (bh)) )
(6)
In Eqn. (6) the more general case of Z independent streams is given. Hence, the HMM
emission pdfs and the cluster prototype pdfs are split up into Z different pdfs b~) (.) and
fJ;Z) (.), respectively and the stream dependent distances D i,z and parameters Ai,z are used.
Now, for the regulizer term 0 the log-likelihood of the mixture model in Eqn. (5) over all
emission pdfs in the HMM system can be used:
K
0(0) =
L logp(b
k=l
k (?))
(7)
740
4
C. Neukirchen and G. Rigoll
Regularization example: discrete HMMs
As an example for parameter estimation in the regularization framework, a discrete HMM
system with different VQs for each of the Z streams is considered here: Each VQ subdivides the feature space into J z different partitions (i.e. the z-th codebook size is J z) and
the VQ-partition labels are denoted m)z) . If the observation subvector x (z) is in the j-th
VQ-partition the VQ output is m(z) (x(z)) = m)Z).
Since the discrete kind HMM output probabilities b~\m(z)) are used here, the regulizer's
, prototypes are the discrete probabilities (3~z) (m (z) ). As a distance metric between the HMM
emission probabilities and the prototype probabilities used in Eqn. (6) the asymmetric
Kullback-Leibler divergence is applied:
(8)
4.1
Estimation of HMM parameters using regularization
The parameter set e of the HMM system to be estimated mainly consists of the discrete
HMM emission probabilities (transition probabilities are not subject of regularization here).
To get an iterative parameter estimation in the EM style, Eqn. (4) must be maximized; e.g.
by setting the derivative of Eqn. (4) with respect to the HMM -parameter b~) (m )z) ) to zero
and application of Lagrange multipliers with regard to the constraint 1 = EJ~ 1 biz)(m ;z)) .
This leads to a quite complex solution that can be only solved numerically.
The optimization problem can be simplified if the mixture in Eqn. (5) is replaced by the
maximum approximation; i.e. only the maximum component in the sum is considered. The
corresponding index of the maximum component is denoted i * :
(9)
In this simplified case the HMM parameter estimation is given by:
(10)
This is a weighted sum of the well known ML solution and the regulizer's prototype probability (3i~ (.) that is selected by the maximum search in Eqn. (9). The larger the value ofthe
constant II, the stronger is the force that pushes the estimate of the HMM emission probability biz) (m ;z)) towards the prototype probability (3i~ (.). The situation when II tends towards
infinity corresponds to the case of traditional state-tying, because all different states that
fall into the same cluster i* make use of (3i~ (.) as emission probability in the z-th stream.
4.2
Estimation of regulizer parameters
The parameter set ~ of the regulizer consists of the mixture weights Ci, the deviation parameters Ai,z , and of the discrete prototype probabilities (3~z) (m ;z) ) in the case of regulizing
741
Controlling the Complexity ofHMM Systems by Regularization
discrete HMMs. These parameters can be set in advance by making use of prior knowledge; e.g. the prototype probabilities can be obtained from a simple HMM system that
uses a small number of states. Alternatively, the regulizer's parameters can be estimated in
a similar way as in (Nowlan, 1992) by maximizing Eqn. (7). Since there is no direct solution to this optimization problem, maximization must be performed in an EM-like iterative
procedure that uses the HMM emission pdfs bk (.) as training data for the mixture model
and by increasing the following auxiliary function in each step:
R(~)
K
I
k=1
i=1
K
I
k=1
i=1
L L P(ilbk(?)) ?logp(i, bk(?))
=
L L P(ilb k(?)) . log (Ci . Pi(bk(?)))
(11)
with the posterior probability used as weighting factor given by:
P(ilb k (.))
ICi . Pi(bk('))
2:: 1=1 Cl . Pl(bk(?))
(12)
Again, maximization of Eqn. (11) can be performed by setting the derivative of R(~)
with respect to the regulizer's parameters to zero under consideration of the constraints
ci and 1 =
f3~Z)(m~Z)) by application of Lagrange multipliers. For the es1=
timation of the regulizer parameters this yields:
2:::=1
o
=
2:::::1
K
~ .L
Ci =
P(ilb k (-))
(13)
2:::=1 P(ilbk(?))
(14)
k=1
~.
_
~,z ~(z)(
i
2:::=1 Di,z(b~) (.)) . P(ilbk(-))
exp (
(z)) _
mj
-
~
~exp
2:::=1 P(ilbk(')) 'IOgb~)(m)Z)))
K.
2::k=1 P(zlbk('))
P(llb k (.)) 'IOgb~)(m}Z)))
(2:::=1
1=1
(15)
K
2::k=1 P(llb k (?))
The estimate Ci can be interpreted as the a:-erage probability that a HMM emission probability falls into the i-th mixture cluster; Ai,z is the inverse ofthe weighted average distance between the emission probabilities and the prototype probability f3;z) ( .). The estimate
~;z)(m)zl) is the average probability over all emission probabilities for the VQ-label m~zl
weighted in the log-domain.
If the Euclidean distance between the discrete probabilities is used instead of Eqn. (8) to
measure the differences between the HMM emission probabilities and the prototypes
Di 'Z (b~)(m(zl)) =
Jz
L (f3jz\myl) - b~z)(m;Zl))
2
(16)
j=1
the estimate of the prototype probabilities is given by the average of the HMM probabilities
weighted in the original space:
(17)
742
5
C. Neukirchen and G. Rigo/l
Experimental results
To investigate the performance of the regularization methods described above a HMM
speech recognition system for the speaker-independent resource management (RM) continuous speech task is built up. For training 3990 sentences from 109 different speakers are
used. Recognition results are given as word error rates averaged over the official DARPA
RM test sets feb'89, oct'89, feb'91 and sep'92, consisting of 1200 sentences from 40 different speakers, totally. Recognition is done via a beam search guided Viterbi decoder using
the DARPA RM word pair grammar (perplexity: 60).
'As acoustic features every 10 ms 12 MFCC coefficients and the relative signal power are
extracted from the speech signal along with the dynamic ~- and ~~-features, comprising
39 features per frame. The HMM system makes use of standard 3-state discrete probability phonetic models. Four different neural networks, trained by the MMI method, that
is described in in (Rigoll, 1997) and extended in (Neukirchen, 1998), are used as VQ to
quantize the features into Z = 4 different streams of discrete labels. The codebook size in
each stream is set to 200.
A simple system with models for 47 monophones and for the most prominent 33 function
words (totally 394 states) yields a word error rate of 8.6%. A system that makes use of
the more detailed (but untied) word internal triphone models (totally 6921 states) yields
12.2% word error. Hence HMM overfitting because of insufficient training data is a severe
problem in this case. Traditional methods to overcome the effects of overfitting like interpolating between triphones and monophones (Bahl, 1983), data driven state-clustering and
decision tree clustering yield error rates of 6.5%, 8.3% and 6.4%, respectively. It must be
noted that in contrast to the usual training procedure in (Rigoll, 1996) no further smoothing
methods are applied to the HMM emission probabilities here.
In a first series of experiments the untied triphone system is regulized by a quite simple
mixture of I = 394 density components, i.e. the number of clusters in the penalty term is
identical to the number of states in the monophone system. In this case the prototype probabilities are initialized by the emission probabilities of the monophone system; the mixture
weights and the deviation parameters in the regulizer are set to be uniform, initially. In
order to test the inluence of the tradeoff parameter v it is set to 50, 10 and 2, respectively.
The corresponding word error rates are 8.4%, 6.9% and 6.3%, respectively. In the case
of large vs regularization degrades to a tying of trip hone states to monophone states and
,the error rate tends towards the monophone system performance. For smaller vs there is a
good tradeoff between data fitting and HMM smoothness yielding improved system performance. The initial prototype probability settings provided by the monophone system do not
seem to be changed much by regulizer parameter estimation, since the system performance
only changes slightly when the regulizer's parameter reestimation is not incorporated.
In preliminary experiments the regularization method is also used for speaker adaptation.
A speaker-independent system trained on the Wall Street Journal (WSJ) database yields an
error rate of32.4% on the Nov. 93 S33>0 test set with 10 different non-native speakers. The
speaker-independent HMM emission probabilities are used to initialize the prototype probabilities of the regulizer. Then, speaker-dependent systems are built up for each speaker
using only 40 fast enrollment sentences for training along with regularization (v is set to
10). Now, the error rate drops to 25.7% what is better than the speaker adaptation method
described in (Rottland, 1998) that yields 27.3% by a linear feature space transformation. In
combination both methods achieve 23.0% word error.
6
Summary and Discussion
A method to avoid parameter overfitting in HMM systems by application of a regularization term that favor smooth and simple models has been presented here. The complexity
Controlling the Complexity of HMM Systems by Regularization
743
measure applied to the HMMs is based on a finite mixture of negative exponential distributions, that generates the state-dependent emission probabilities. This kind of regularization
term can be interpreted as a soft state-tying, since it forces the HMM emission probabilities
to form a finite set of clusters. The effect of regularization has been demonstrated on the
RM task by improving overfitted trip hone models. On a WSJ non-native speaker adaption
task with limited training data, regularization outperforms feature space transformations.
Eqn. (4) may be also interpreted from a perspective of Bayesian inference: the term v .
n plays the role of setting a prior distribution on the HMM parameters to be estimated.
Hence, the use of a mixture model for n is equivalent to using a special kind of prior in the
framework of MAP estimation for HMMs (Gauvain, 1994).
References
L.R. Bahl, F. Jelinek, L.R. Mercer, 'A Maximum Likelihood Approach to Continuous
Speech Recognition', IEEE Trans. Pattern Analysis and Machine Intelligence, Vol. 5, No.2
Mar. 1983, pp. 179-190.
L.R. Bahl, P.v. de Souza, P.S. Gopalakrishnan,D. Nahamoo, M.A. Picheny, (1991) Context
dependent modeling of phones in continuous speech using decision trees. Proc. DARPA
speech and natural language processing workshop, 264-270.
E.B. Baum, D. Haussler, (1989) What size net gives valid generalization? Neural Computation, 1:151-160.
Y. Ie Cun, J. Denker, S. Solla, R.E. Howard, L.D. Jackel, (1990) Optimal brain damage.
Advances in Neural Information Processing Systems 2, San Mateo, CA, Morgan Kauffinan.
J.L. Gauvain, C.R. Lee, (1994) Maximum a posteriori estimation for multivariate Gaussian
mixture observations of markov chains. IEEE Transaction Speech and Audio Proc., Vol. 2,
2:291-298.
KJ. Lang, A.H. Waibel, G.E. Hinton, (1990) A time-delay neural network architecture for
isolated word recognition. Neural Networks, 3:23~3.
Ch. Neukirchen, D. Willett, S. Eickeler, S. Muller, (1998) Exploiting acoustic feature
correlations by joint neural vector quantizer design in a discrete HMM system. Proc.
ICASSP'98,5-8.
S.J. Nowlan, G.E. Hinton, (1992) Simplifying neural networks by soft weight-sharing.
Neural Computation, 4:473~93.
D.C. Plaut, S.J. Nowlan, G.E. Hinton, (1986) Experiments on learning by backpropagation. technical report CMU-CS-86-126, Carnegie-Mellon University, Pittsburgh,
PA.
G. Rigoll, Ch. Neukirchen, J. Rottland, (1996) A new hybrid system based on MMI-neural
networks for the RM speech recognition task. Proc. ICASSP'96, 865-868.
G. Rigoll, Ch. Neukirchen, (1997) A new approach to hybrid HMMIANN speech recognition using mutual information neural networks. Advances in Neural Information Processing Systems 9, Cambridge, MA, MIT Press, 772-778.
J. Rottland, Ch. Neukirchen, G. Rigoll, (1998) Speaker adaptation for hybrid MMIconnectionist speech recognition systems. Pmc. ICASSP '98, 465~68.
'S.J. Young, (1992) The general use of tying in phoneme-based HMM speech recognizers.
Proc. ICASSP '92, 569- 572.
SJ. Young, P.C. Woodland (1993) The use of state tying in continuous speech recognition.
Proc. Eurospeech '93, 2203-2206.
| 1519 |@word stronger:1 simplifying:1 initial:1 series:1 selecting:1 outperforms:1 nowlan:4 gauvain:2 lang:2 must:5 written:1 partition:3 drop:1 v:2 intelligence:1 selected:2 quantizer:2 plaut:2 codebook:2 along:2 constructed:1 direct:1 become:1 consists:2 fitting:1 iogb:2 brain:1 decomposed:1 subvectors:1 increasing:1 becomes:4 totally:3 provided:1 underlying:1 s33:1 what:2 tying:10 kind:3 interpreted:6 transformation:2 every:1 ti:1 ofa:1 classifier:1 rm:5 control:1 zl:4 tends:2 mercator:1 mateo:1 christoph:1 hmms:8 limited:3 statistically:1 averaged:1 backpropagation:1 procedure:2 word:9 get:2 close:1 bh:1 context:2 influence:1 rottland:3 equivalent:1 map:1 demonstrated:2 baum:2 maximizing:1 assigns:1 pure:1 estimator:1 adjusts:1 haussler:1 traditionally:1 controlling:4 play:1 gerhard:2 us:2 pa:1 recognition:12 asymmetric:1 native:2 database:1 role:1 solved:2 solla:1 overfitted:2 complexity:12 neglected:1 dynamic:1 trained:2 subdivides:1 sep:1 darpa:3 joint:1 icassp:4 represented:1 train:2 forced:1 fast:1 quite:2 larger:3 grammar:1 favor:3 sequence:1 net:1 adaptation:4 flexibility:1 achieve:1 exploiting:1 cluster:13 generating:1 wsj:2 augmenting:1 auxiliary:2 c:1 guided:1 drawback:1 generalization:4 wall:1 preliminary:1 pl:1 sufficiently:2 considered:2 exp:3 viterbi:1 estimation:10 proc:6 label:4 jackel:1 weighted:4 mit:1 gaussian:2 avoid:2 ej:1 es1:1 emission:26 pdfs:16 likelihood:5 mainly:1 contrast:2 posteriori:1 inference:1 dependent:9 initially:1 hidden:1 favoring:1 comprising:1 germany:1 among:1 denoted:8 constrained:1 special:2 smoothing:1 initialize:1 mutual:1 comprise:1 f3:2 shaped:1 identical:4 report:1 ilb:3 serious:1 few:1 divergence:1 individual:1 hmmiann:1 replaced:2 argmax:2 consisting:1 investigate:1 severe:1 introduces:1 mixture:18 yielding:2 chain:1 tree:4 euclidean:1 penalizes:1 initialized:1 isolated:1 monophone:5 neukirchen:9 classify:1 soft:6 modeling:1 enrollment:1 rigo:1 logp:2 maximization:3 deviation:3 uniform:1 delay:1 successful:1 eurospeech:1 density:3 sensitivity:1 ie:2 probabilistic:2 lee:1 again:1 management:1 derivative:2 style:1 de:2 coefficient:1 caused:1 stream:10 reg:1 performed:2 capability:1 variance:1 phoneme:1 maximized:1 yield:6 ofthe:2 generalize:1 bayesian:1 mmi:2 mfcc:1 sharing:8 email:1 rdt:1 against:2 pp:1 di:3 popular:1 knowledge:3 higher:1 follow:1 improved:1 done:2 mar:1 correlation:1 eqn:12 bahl:4 biz:2 effect:3 true:1 multiplier:2 adequately:1 regularization:27 hence:4 leibler:1 during:1 speaker:13 noted:1 m:1 prominent:1 pdf:6 fj:1 consideration:1 novel:1 common:3 bko:1 numerically:1 willett:1 mellon:1 cambridge:1 ai:4 smoothness:2 language:1 recognizers:1 add:1 feb:2 triphones:1 posterior:1 multivariate:1 perspective:1 driven:2 perplexity:1 phone:1 phonetic:2 binary:1 muller:1 seen:1 morgan:1 triphone:5 determine:1 signal:2 ii:3 smooth:4 technical:1 match:1 basic:1 metric:3 expectation:1 cmu:1 sometimes:1 llb:2 beam:1 extra:1 subject:1 tend:1 seem:2 split:2 enough:1 architecture:1 reduce:3 idea:1 prototype:15 tradeoff:3 vqs:1 rigoll:9 expression:2 penalty:4 monophones:2 speech:14 woodland:1 detailed:2 amount:2 simplest:1 generate:1 estimated:4 per:1 discrete:14 carnegie:1 vol:2 group:2 four:1 duisburg:3 backward:1 sum:2 inverse:1 decision:4 nahamoo:1 constraint:2 infinity:1 untied:2 generates:1 relatively:1 department:1 waibel:1 combination:1 smaller:1 slightly:1 em:5 increasingly:1 cun:2 making:1 resource:1 vq:10 count:1 fji:1 available:1 denker:1 original:1 clustering:3 ensure:1 objective:2 xiw:1 degrades:1 damage:1 usual:1 traditional:6 gradient:1 distance:8 mapped:1 hmm:48 decoder:1 street:1 gopalakrishnan:1 index:1 relationship:1 insufficient:2 balance:2 pmc:1 negative:3 design:1 adjustable:3 observation:7 markov:2 hone:2 howard:1 finite:3 situation:1 extended:1 incorporated:1 hinton:3 frame:1 community:1 souza:1 bk:14 pair:1 required:1 subvector:2 trip:2 connection:2 optimized:1 sentence:3 acoustic:2 trans:1 able:1 pattern:2 built:2 reliable:1 power:1 suitable:1 natural:1 force:3 hybrid:3 kj:1 prior:4 relative:1 ofhmm:4 penalization:2 mercer:1 share:1 pi:5 changed:1 summary:1 fall:4 jelinek:1 regard:1 overcome:1 transition:2 avoids:1 valid:1 forward:1 commonly:2 ici:1 simplified:2 san:1 cope:1 transaction:1 picheny:1 nov:1 sj:1 uni:1 kullback:1 ml:6 overfitting:9 reestimation:1 pittsburgh:1 assumed:4 alternatively:1 continuous:7 iterative:2 search:2 timation:1 learn:1 transfer:1 mj:1 jz:1 ca:1 improving:2 quantize:1 complex:2 cl:1 constructing:3 domain:2 official:1 interpolating:1 augmented:1 exponential:3 tied:2 weighting:1 young:4 removing:1 rk:1 decay:1 workshop:1 ci:9 push:1 chn:1 smoothly:1 lagrange:2 contained:1 ch:4 corresponds:1 adaption:1 extracted:1 ma:1 oct:1 towards:4 change:1 determined:2 total:2 invariance:1 experimental:1 internal:1 incorporate:1 audio:1 |
569 | 152 | 436
SIMULATION AND MEASUREMENT OF
THE ELECTRIC FIELDS GENERATED
BY WEAKLY ELECTRIC FISH
Brian Rasnow 1, Christopher Assad2, Mark E. Nelson3 and James M. Bow~
Divisions of Physics1,Elecbical Engineerini, and Biolo~
Caltech, Pasadena, 91125
ABSTRACT
The weakly electric fish, Gnathonemus peters;;, explores its environment by generating pulsed elecbic fields and detecting small pertwbations in the fields resulting from
nearby objects. Accordingly, the fISh detects and discriminates objects on the basis of a
sequence of elecbic "images" whose temporal and spatial properties depend on the timing of the fish's electric organ discharge and its body position relative to objects in its environmenl We are interested in investigating how these fish utilize timing and body-position during exploration to aid in object discrimination. We have developed a fmite-element simulation of the fish's self-generated electric fields so as to reconstruct the electrosensory consequences of body position and electric organ discharge timing in the fish.
This paper describes this finite-element simulation system and presents preliminary electric field measurements which are being used to tune the simulation.
INTRODUCTION
The active positioning of sensory structures (i.e. eyes, ears, whiskers, nostrils, etc.)
is characteristic of the information seeking behavior of all exploratory animals. Yet, in
most existing computational models and in many standard experimental paradigms, the
active aspects of sensory processing are either eliminated or controlled (e.g. by stimulating fIXed groups of receptors or by stabilizing images). However, it is clear that the active positioning of receptor surfaces directly affects the content and quality of the sensory
infonnation received by the nervous system. Thus. controlling the position of sensors
during sensory exploration constitutes an important feature of an animals strategy for
making sensory discriminations. Quantitative study of this process could very well shed
light on the algorithms and internal representations used by the nervous system in discriminating peripheral objects.
Studies of the active use of sensory surfaces generally can be expected to pose a
number of experimental challenges. This is because, in many animals, the sensory surfaces involved are themselves structurally complicated, making it difficult to reconstruct p0sition sequences or the consequences of any repositioning. For example, while the sen-
Simulation and Measurement of the Weakly Electric Fish
sory systems of rats have been the subjects of a great deal of behavioral (Welker, 1964)
and neurophysiological study (Gibson & Welker, 1983), it is extremely difficult to even
monitor the movements of the perioral surfaces (lips, snout, whiskers) used by these animals in their exploration of the world let alone reconstruct the sensory consequences. For
these reasons we have sought an experimental animal with a sensory system in which
these sensory-motor interactions can be more readily quantified.
The experimental animal which we have selected for studying the control of sensory
surface position during exploration is a member of a family of African freshwater fish
(Monniridae) that use self-generated electric fields to detect and discriminate objects in
their environment (Bullock & Heiligenberg, 1986). The electrosensory system in these
fish relies on an "electric organ" in their tails which produces a weak pulsed electric field
in the surrounding environment (significant within 1-2 body lengths) that is then detected
with an array of electrosensors that are extremely sensitive to voltage drops across the
skin. These "electroreceptors" allow the fISh to respond to the perturbations in the electric field resulting from objects in the environment which differ in conductivity from the
surrounding water (Fig. 1).
. . conducting
?
object
IIID electric organ
?
electroreceptors
electric
field lines
Figure 1. The peripheral electrosensory system of Gnathonemus petersii
consists of an "electric organ" current source at the base of the tail and several thousand "electroreceptor" cells distributed non uniformly over the
fish's body. A conducting object near the fish causes a local increase in the
current through the skin.
These fISh are nocturnal, and rely more on their electric sense than on any other sensory
system in perfonning a wide range of behaviors (eg. detecting and localizing objects such
as food). It is also known that these fish execute exploratory movements, changing their
body position actively as they attempt an electrosensory discrimination (Toerring &
Belbenoit, 1979). Our objective is to understand how these movements change the distribution of the electric field on the animals skin, and to determine what, if any, relationship
this has to the discrimination process.
There are several clear advantages of this system for our studies. First, the electrore-
437
438
Rasnow, Assad, Nelson and Bower
ceptors are in a fixed position with respect to each other on the surface of the animal.
Therefore, by knowing the overall body position of the animal it is possible to know the
exact spatial relationship of electroreceptors with respect to objects in the environment.
Second, the physical equations governing the self-generated electric fIeld in the fish's environment are well understood. As a consequence, it is relatively straightforward to reconstruct perturbations in the electric field resulting from objects of different shape and
conductance. Third, the electric potential can be readily measured, providing a direct
measure of the electric field at a distance from the fish which can be used to reconstruct
the potential difference across the animals skin. And finally, in the particular species of
fish we have chosen to work with, Gnathonemus petersii, individual animals execute a
brief (100 J.1Sec) electric organ discharge (BOD) at intervals of 30 msec to a few seconds.
Modification of the firing pattern is 1cnown to be correlated with changes in the electrical
environment (Lissmann, 1958). Thus, when the electric organ discharges, it is probable
that the animal is interested in "taking a look" at its surroundings. In few other sensory
systems is there as direct an indication of the attentional state of the subject.
Having stated the advantages of this system for the study we have undertaken, it is
also the case that considerable effort will still be necessary to answer the questions we
have posed. For example, as described in this paper, in order to use electric field measurements made at a distance to infer the voltages across the surface of the animal's skin,
it is necessary to develop a computer model of the fish and its environment. This will
allow us to predict the field on the animal's skin surface given different body poSitions
relative to objects in the environment. This paper describes our first steps in constructing
this simulation system.
Experimental Approach and Methods
Simulations of Fish Electric Fields
The electric potential, cll(x), generated by the EOD of a weakly electric fish in a fish
tank is a solution ofPoisson's equation:
Ve(pVell) = f
where p(x)and f(x) are the impedance magnitude and source density at each point x inside and surrounding the fish. Our goal is to solve this equation for ell given the current
source density, f, generated by the electric organ and the impedances, p, corresponding to
the properties of the fish and external objects (rocks, worms, etc.). Given p and f. this
equation can be solved for the potential ell using a variety of iterative approximation
schemes. Iterative methods, in general, first discretize the spatial domain of the problem
into a set of "node" points, and convert Poisson's equation into a set of algebraic equations with the nodal potentials as the unknown parameters. The node values, in this case,
each represent an independent degree of freedom of the system and, as a consequence,
there are as many equations as there are nodes. This very large system of equations can
Simulation and Measurement of the Weakly Electric Fish
then be solved using a variety of standard techniques, including relaxation methods, conjugate gradient minimization, domain decomposition and multi-grid methods.
To simulate the electric fields generated by a fish, we currently use a 2-dimensional
fmite element domain discretization (Hughes, 1987) and conjugate gradient solver. We
chose the finite element method because it allows us to simulate the electric fields at
much higher resolution in the area of interest close to the animal's body where the electric field is largest and where errors due to the discretization would be most severe. The
fmite element method is based on minimizing a global function that corresponds to the
potential energy of the electric field. To compute this energy, the domain is decomposed
into a large number of elements, each with uniform impedance (see Fig. 2). The global
energy is expressed as a sum of the contributions from each element, where the potential
within each element is assumed to be a linear interpolation of the potentials at the nodes
or vertices of each element The conjugate gradient solver determines the values of the
node potentials which minimize the global energy function.
1\ IVrv'V
1\ 1\ rv:V J J 1\/1\ .J1\/
1\11\/1\/1\11\/1\/1\ 1\11\/[\ 1\1 IV
r--..
V
v
7'
v
[7
If\ If\ '\ '\ V\ V If\ J\ 1'\
'w
l/\ V 11'\ '\ 1/ :1'\ '\ '\ '\ V '\ 1'\" '\ 11\1/\ \ '\ V
Figure 2. The inner region of a fmite element grid constructed for simulating in 2-dimensions the electric field generated by an electric fish.
Measurement of Fish Electric Fields
Another aspect of our experimental approach involves the direct measurement of
the potential generated by a fish's EOD in a fish tank using arrays of small electrodes and
differential amplifiers. The electrodes and electronics have a high impedance which minimizes their influence on the electric fields they are designed to measure. The electrodes
are made by pulling a 1mm glass capillary tube across a heated tungsten filament, resulting in a fine tapered tip through which a 1~ silver wire is run. The end of this wire is
melted in a flame leaving a 200J,un ball below the glass insulation. Several electrodes are
then mounted as an array on a microdrive attached to a modified X-Yplotter under computer control and giving better than 1mm positioning accuracy. Generated potentials are
amplified by a factor of 10 - 100, and digitized at a rate of 100kHz per channel with a 12
bit AID converter using a Masscomp 5700 computer. An array processor searches this
439
440
Rasnow, Assad, Nelson and Bower
continuous stream of data for EOD wavefonns. which are extracted and saved along with
the position of the electrode array.
Results
Calibration of the Simulator
In order to have confidence in the overall system, it was fD'St necessary to calibrate
both the recording and the simulation procedures. To do this we set up relatively simple
geometrical arrangements of sources and conductors in a fish tank for which the potential
could be found analytically. The calibration source was an electronic "fake fish" circuit
that generated signals resembling the discharge of Gnathonemus.
Point current source
A point source in a 2-dimensional box is perhaps the simplest configuration with
which to initially test our electric field reconstruction system. The analytic solution for
the potential from a point current source centered in a grounded. conducting 2-dimensional box is:
. (.n7t). (n7tx). h (.n7ty )
00
4>(x. y) =
L
n =1
sm("2 sm
L
sm
\L
ri1t
n L cosh(T)
Our fmite element simulation. based on a regular 80 x 80 node grid differs from the
above expression by less than 1%. except in the elements adjacent to the source. where
the potential change across these elements is large and is not as accurately reconstructed
by a linear interpolation (Fig. 3). Smaller elements surrounding the source would improve the accuracy. however. one should note the analytic solution is infmite at the location of the "point" source whereas the measured and simulated sources (and real fish)
have finite current densities.
To measure the real equivalent of a point source in a 2-dimensional box. we used a
linear current source (a wire) which ran the full depth of a real 3-dimensional tank.
Measurements made in the midplane of the tank agree with the simulation and analytic
solution to better than 5% (Fig. 3.). Uncertainty in the positions of the ClUTent source and
recording sites relative to the position of the conducting walls probably accounts for
much of this difference.
Simulation and Measurement of the Weakly Electric Fish
1----~~--~--~----~-------
o - measured
x - simulated
00
2
4
6
8
10
12
14
16
dislaDce from source
Figure 3. Electric potential of a point current source centered in a grounded
2-dimensional box.
Measurements of Fish Fields and 2-Dimensional Simulations
Calibration of our fmite element model of an electric fish requires direct measurements of the electric potential close to a discharging fish. Fig. 4 shows a recording of a
single EOD sampled with 5 colinear electrodes near a restrained fish. The wavefonn is
bipolar, with the fIrst phase positive if recorded near the animals head and negative if recorded near the tail (relative to a remote reference). We used the peak amplitude of the
larger second phase of the wavefonn to quantify the electric potential recorded at each
location. Note that the potential reverses sign at a point approximately midway along the
tail. This location corresponds to the location of the null potential shown in Fig. 5.
1500
1000
$'
5500
I
0
-soo
-1000
-1ro
200
~sec
Figure 4. EOD waveform sampled simultaneously from 5 electrodes.
441
442
Rasnow, Assad, Nelson and Bower
Measurements of EODs from a restrained fish exhibited an extraordinarily small variance in amplitude and waveform over long periods of time. In fact, the peak-peak amplitude of the EOD varied by less than 0.4% in a sample of 40 EOD's randomly chosen during a 30 minute period. Thus we are able to directly compare waveforms sampled sequentially without renonnalizing for fluctuations in EOD amplitude.
Figure 5 shows equipotential lines reconstructed from a set of 360 measurements
made in the midplane of a restrained Gnathonemus. Although the observed potential resembles that from a purely dipolar source (Fig. 6), careful inspection reveals an asymmetry between the head and tail of the fISh. This asymmetry can be reproduced in our simulations by adjusting the electrical properties of the fish. Qualitatively, the measured
fields can be reproduced by assigning a low impedance to the internal body cavity and a
high impedance to the skin. However, in order to match the location of the null potential,
the skin impedance must vary over the length of the body. We are currently quantifying
these parameters, as described in the next section.
!!!!m 1'!I!fl!IPf!~m
!~ ...... II..
?? 1.. . ...... 1 ???
Figure 5. Measured potentials (at peak of second phase of EOD) recorded
from a restrained Gnathonemus petersii in the midplane of the fish.
Equipotential lines are 20 mV apart. Inset shows relative location of fish
and sampling points in the fISh tank.
Figure 6. Equipotential lines from a 2-dimensional finite element simulation of a dipole using the grid of Fig. 2. The resistivity of the fish was set
equal to that of the sWToundings in this simulation.
Simulation and Measurement of the Weakly Electric Fish
Future Directions
There is still a substantial amount of work that remains to be done before we
achieve our goal of being able to fully reconstruct the pattern of electroreceptor activation for any arbitrary body position in any particular environment. First. it is clear that
we require more information about the electrical structure of the fISh itself. We need an
accurate representation of the internal impedance distribution p(x) of the body and skin
as well as of the source density f(x) of the electric organ. To some extent this can be addressed as an inverse problem, namely given the measured potential cl>(x), what choice of
p(x) and f(x) best reproduces the data. Unfortunately, in the absence of further constraints, there are many equally valid solution, thus we will need to directly measure the
skin and body impedance of the fish. Second, we need to extend our finite-element simulations of the fish to 3-dimensions which, although conceptually straight forward, requires substantial technical developments to be able to (a) specify and visualize the
space-filling set of 3-dimensional finite-elements (eg. tetrahedrons) for arbitrary configurations, (b) compute the solution to the much larger set of equations (typically a factor of
100-1(00) in a reasonable time, and (c) visualize and analyze the resulting solutions for
the 3-dimensional electrical fields. As a possible solution to (b), we are developing and
testing a parallel processor implementation of the simulator.
References
Bullock, T. H. & Heiligenberg, W. (Eds.) (1986). "Electroreception", Wiley & Sons,
New York.
Gibson, J. M. & Welker. W. I. (1983). Quantitative Studies of Stimulus Coding in FirstOrder Vibrissa Afferents of Rats. 1. Receptive Field Properties and Threshold
Distributions. Somatosensory Res. 1:51-67.
Hughes, T. J. (1987). The Finite Element Method: Linear Static and Dynamic Finite
Element Analysis. Prentice-Hall, New Jersey.
Lissmann. H.W. (1958). On the function and evolution of electric organs in fish. J. Exp.
Bioi. 35:156-191.
Toening, M. J. and Belbenoit. P. (1979). Motor Programmes and Electroreception in
Monnyrid Fish. Behav. Ecol. Sociobiol. 4:369-379.
Welker, W. I. (1964). Analysis of Sniffing of the Albino Rat Behaviour 22:223-244.
443
| 152 |@word simulation:18 electrosensory:4 decomposition:1 electroreceptors:3 fmite:6 electronics:1 configuration:2 existing:1 current:8 discretization:2 activation:1 yet:1 assigning:1 must:1 readily:2 j1:1 midway:1 shape:1 analytic:3 motor:2 drop:1 designed:1 discrimination:4 alone:1 selected:1 nervous:2 accordingly:1 perioral:1 inspection:1 detecting:2 node:6 location:6 nodal:1 constructed:1 direct:4 differential:1 along:2 consists:1 behavioral:1 inside:1 expected:1 behavior:2 themselves:1 multi:1 simulator:2 detects:1 decomposed:1 food:1 solver:2 circuit:1 null:2 what:2 minimizes:1 developed:1 eod:9 temporal:1 quantitative:2 firstorder:1 shed:1 bipolar:1 ro:1 control:2 conductivity:1 discharging:1 positive:1 before:1 understood:1 timing:3 local:1 consequence:5 receptor:2 firing:1 interpolation:2 approximately:1 fluctuation:1 chose:1 resembles:1 quantified:1 tungsten:1 range:1 filament:1 testing:1 wavefonns:1 hughes:2 differs:1 insulation:1 procedure:1 area:1 gibson:2 confidence:1 regular:1 close:2 prentice:1 influence:1 heated:1 equivalent:1 resembling:1 straightforward:1 resolution:1 stabilizing:1 clutent:1 dipole:1 array:5 exploratory:2 discharge:5 controlling:1 exact:1 element:20 observed:1 electrical:4 solved:2 thousand:1 region:1 remote:1 movement:3 ran:1 substantial:2 discriminates:1 environment:10 dynamic:1 weakly:7 depend:1 colinear:1 purely:1 division:1 basis:1 jersey:1 surrounding:4 detected:1 extraordinarily:1 whose:1 posed:1 solve:1 larger:2 reconstruct:6 itself:1 reproduced:2 sequence:2 advantage:2 indication:1 rock:1 sen:1 reconstruction:1 interaction:1 bow:1 achieve:1 amplified:1 bod:1 electrode:7 asymmetry:2 produce:1 generating:1 silver:1 object:14 develop:1 pose:1 measured:6 received:1 electroreception:2 involves:1 somatosensory:1 revers:1 quantify:1 differ:1 direction:1 waveform:3 saved:1 exploration:4 centered:2 require:1 behaviour:1 wall:1 preliminary:1 brian:1 probable:1 mm:2 ecol:1 hall:1 exp:1 great:1 predict:1 visualize:2 sought:1 vary:1 currently:2 infonnation:1 sensitive:1 largest:1 organ:10 minimization:1 sensor:1 modified:1 voltage:2 sense:1 detect:1 glass:2 typically:1 initially:1 pasadena:1 interested:2 tank:6 overall:2 iiid:1 development:1 animal:16 spatial:3 ell:2 field:28 equal:1 having:1 eliminated:1 sampling:1 look:1 constitutes:1 filling:1 future:1 stimulus:1 few:2 surroundings:1 randomly:1 simultaneously:1 ve:1 individual:1 phase:3 attempt:1 freedom:1 conductance:1 amplifier:1 interest:1 fd:1 severe:1 light:1 electroreceptor:2 accurate:1 necessary:3 iv:1 re:1 localizing:1 calibrate:1 vertex:1 uniform:1 answer:1 repositioning:1 st:1 density:4 explores:1 freshwater:1 discriminating:1 cll:1 peak:4 tip:1 ear:1 tube:1 recorded:4 sniffing:1 external:1 actively:1 account:1 potential:23 sec:2 coding:1 mv:1 afferent:1 stream:1 analyze:1 complicated:1 parallel:1 contribution:1 minimize:1 capillary:1 accuracy:2 variance:1 characteristic:1 conducting:4 conceptually:1 weak:1 accurately:1 straight:1 processor:2 african:1 resistivity:1 ed:1 tetrahedron:1 energy:4 involved:1 james:1 nostril:1 static:1 sampled:3 adjusting:1 amplitude:4 higher:1 specify:1 execute:2 box:4 done:1 snout:1 governing:1 christopher:1 quality:1 perhaps:1 ipf:1 pulling:1 evolution:1 analytically:1 restrained:4 infmite:1 deal:1 eg:2 adjacent:1 during:4 self:3 rat:3 equipotential:3 midplane:3 heiligenberg:2 geometrical:1 image:2 physical:1 attached:1 khz:1 extend:1 tail:5 measurement:14 significant:1 grid:4 calibration:3 surface:8 etc:2 base:1 pulsed:2 apart:1 caltech:1 determine:1 paradigm:1 period:2 signal:1 ii:1 rv:1 full:1 infer:1 positioning:3 match:1 technical:1 long:1 equally:1 controlled:1 dipolar:1 poisson:1 represent:1 grounded:2 cell:1 whereas:1 fine:1 interval:1 addressed:1 source:19 leaving:1 exhibited:1 probably:1 subject:2 recording:3 member:1 near:4 variety:2 affect:1 converter:1 inner:1 knowing:1 expression:1 effort:1 peter:1 algebraic:1 york:1 cause:1 behav:1 generally:1 fake:1 clear:3 nocturnal:1 tune:1 amount:1 cosh:1 simplest:1 fish:51 sign:1 per:1 group:1 threshold:1 monitor:1 changing:1 tapered:1 undertaken:1 utilize:1 welker:4 relaxation:1 convert:1 sum:1 run:1 inverse:1 uncertainty:1 respond:1 family:1 reasonable:1 electronic:1 bit:1 fl:1 constraint:1 nearby:1 aspect:2 simulate:2 extremely:2 relatively:2 developing:1 peripheral:2 ball:1 conjugate:3 describes:2 across:5 smaller:1 son:1 wavefonn:2 bullock:2 making:2 modification:1 equation:9 agree:1 remains:1 know:1 end:1 studying:1 simulating:1 giving:1 seeking:1 skin:10 objective:1 question:1 arrangement:1 strategy:1 receptive:1 gradient:3 distance:2 attentional:1 simulated:2 sory:1 nelson:3 extent:1 reason:1 water:1 length:2 relationship:2 providing:1 minimizing:1 difficult:2 unfortunately:1 stated:1 negative:1 implementation:1 unknown:1 discretize:1 wire:3 sm:3 finite:8 head:2 digitized:1 perturbation:2 varied:1 vibrissa:1 arbitrary:2 namely:1 able:3 below:1 pattern:2 challenge:1 including:1 soo:1 rely:1 scheme:1 improve:1 brief:1 eye:1 relative:5 fully:1 whisker:2 mounted:1 degree:1 allow:2 understand:1 wide:1 taking:1 distributed:1 dimension:2 depth:1 world:1 valid:1 sensory:13 forward:1 made:4 qualitatively:1 programme:1 reconstructed:2 cavity:1 global:3 active:4 investigating:1 sequentially:1 reveals:1 reproduces:1 assumed:1 un:1 iterative:2 search:1 continuous:1 lip:1 impedance:9 channel:1 cl:1 electric:46 constructing:1 domain:4 body:14 fig:8 site:1 aid:2 wiley:1 structurally:1 position:13 msec:1 bower:3 third:1 minute:1 inset:1 rasnow:4 magnitude:1 neurophysiological:1 expressed:1 albino:1 corresponds:2 determines:1 relies:1 extracted:1 stimulating:1 bioi:1 goal:2 quantifying:1 careful:1 absence:1 content:1 change:3 considerable:1 except:1 uniformly:1 conductor:1 worm:1 specie:1 discriminate:1 experimental:6 assad:3 internal:3 mark:1 eods:1 perfonning:1 correlated:1 |
570 | 1,520 | Exploiting generative models
discriminative classifiers
Tommi S. Jaakkola*
MIT Artificial Intelligence Laboratorio
545 Technology Square
Cambridge, MA 02139
?
In
David Haussler
Department of Computer Science
University of California
Santa Cruz, CA 95064
Abstract
Generative probability models such as hidden ~larkov models provide a principled way of treating missing information and dealing
with variable length sequences. On the other hand , discriminative
methods such as support vector machines enable us to construct
flexible decision boundaries and often result in classification performance superior to that of the model based approaches. An ideal
classifier should combine these two complementary approaches. In
this paper, we develop a natural way of achieving this combination by deriving kernel functions for use in discriminative methods
such as support vector machines from generative probability models. We provide a theoretical justification for this combination as
well as demonstrate a substantial improvement in the classification
performance in the context of D~A and protein sequence analysis.
1
Introduction
Speech, vision , text and biosequence data can be difficult to deal with in the context
of simple statistical classification problems. Because the examples to be classified
are often sequences or arrays of variable size that may have been distorted in particular ways, it is common to estimate a generative model for such data, and then
use Bayes rule to obtain a classifier from this model. However. many discriminative methods, which directly estimate a posterior probability for a class label (as
in Gaussian process classifiers [5]) or a discriminant function for the class label
(as in support vector machines [6]) have in other areas proven to be superior to
* Corresponding author.
T. S. Jaakkola and D. Haussler
488
generative models for classification problems. The problem is that there has been
no systematic way to extract features or metric relations between examples for use
with discriminative methods in the context of difficult data types such as those
listed above. Here we propose a general method for extracting these discriminatory
features using a generative model. V{hile the features we propose are generally
applicable, they are most naturally suited to kernel methods.
2
Kernel methods
Here we provide a brief introduction to kernel methods; see, e.g., [6] [5] for more
details. Suppose now that we have a training set of examples Xl and corresponding
binary labels 51 (?1) . In kernel methods. as we define them. the label for a new
example X is obtained from a weighted sum of the training labels. The weighting of
each training label 52 consists of two parts: 1) the overall importance of the example
Xl as summarized with a coefficient '\1 and 2) a measure of pairwise "similarity"
between between XI and X, expressed in terms of a kernel function K(X2' X). The
predicted label S for the new example X is derived from the following rule:
s ~ sign ( ~ S, '\,K(X,. X) )
(1)
We note that this class of kernel methods also includes probabilistic classifiers, in
\vhich case the above rule refers to the label with the maximum probability. The
free parameters in the classification rule are the coefficients '\1 and to some degree
also the kernel function K . To pin down a particular kernel method. two things
need to be clarified. First , we must define a classification loss . or equivalently, the
optimization problem to solve to determine appropriate values for the coefficients
'\1' Slight variations in the optimization problem can take us from support vector
machines to generalized linear models. The second and the more important issue is
the choice of the kernel function - the main topic of this paper. \Ve begin with a
brief illustration of generalized linear models as kernel methods.
2.1
Generalized linear models
For concreteness we consider here only logistic regression models. while emphasizing
that the ideas are applicable to a larger class of models l . In logistic regression
models , the probability of the label 5 given the example X and a parameter vector
e is given by2
P(5IX. e) = (7 (5e T X)
(2)
where (7(z) = (1 + e- z) - l is the logistic function. To control the complexity of
the model when the number of training examples is small we can assign a prior
distribution p(e) over the parameters. \Ve assume here that the prior is a zero
mean Gaussian with a possibly full covariance matrix L:. The maximum a posteriori
(l\IAP) estimate for the parameters e given a training set of examples is found by
1 Specifically. it applies to all generalized linear models whose transfer functions are
log-concave.
2Here we assume that the constant + 1 is appended to every feature vector X so that
an adjustable bias term is included in the inner product T X.
e
Exploiting Generative Models in Discriminative Classifiers
489
maximizing the following penalized log-likelihood:
I: log P(S, IX
1,
B)
+ log P(B)
where the constant c does not depend on B. It is straightforward to show, simply
by taking the gradient with respect to the parameters , that the solution to this
(concave) maximization problem can be written as 3
(4)
Xote that the coefficients A, appear as weights on the training examples as in the
definition of the kernel methods . Indeed. inserting the above solution back into the
conditional probability model gives
(5)
By identifying !..:(X/. X) = X;'f.X and noting that the label with the maximum
probability is the aile that has the same sign as the sum in the argument. this gives
the decision rule (1).
Through the above derivation , we have written the primal parameters B in terms
of the dual coefficients A,.J. Consequently. the penalized log-likelihood function can
be also written entirely in terms of A, : the resulting likelihood function specifies
how the coefficients are to be optimized. This optimization problem has a unique
solution and can be put into a generic form. Also , the form of the kernel function
that establishes the connection between the logistic regression model and a kernel
classifier is rather specific , i.e .. has the inner product form K(X,. X) = X;'f.X.
However. as long as the examples here can be replaced with feature vectors derived
from the examples. this form of the kernel function is the most general. \Ve discuss
this further in the next section.
3
The kernel function
For a general kernel fUIlction to be valid. roughly speaking it only needs to be positive semi-definite (see e.g. [7]). According to the t-Iercer 's theorem. any such valid
kernel function admits a representation as a simple inner product bet\\'een suitably
defined feature vectors. i.e .. !":(X,.Xj) = 0\,0.'\) . where the feature vectors come
from some fixed mapping X -> ? .'\. For example. in the previous section the kernel
function had the form X;'f.Xj ' which is a simple inner product for the transformed
feature vector ? .'\ = 'f. 1- X.
Specifying it simple inner product in the feature space defines a Euclidean metric space. Consequently. the Euclidean distances between the feature vectors are
obtained directly from the kernel fUllction: with the shorthand notation K ,} =
3This corresponds to a Legendre transformation of the loss functions log a( z) .
.}This is possible for all those e that could arise as solutions to the maximum penalized
likelihood problem: in other words. for all relevant e.
490
T. S. Jaakkola and D. Haussler
K(Xi , Xj) we get II<Px, - <PxJ W = K ti - 2Ktj + K jj . In addition to defining the
metric structure in the feature space, the kernel defines a pseudo metric in the original example space through D(Xi,Xj) = II<px. - <pxJII. Thus the kernel embodies
prior assumptions about the metric relations between the original examples. No
systematic procedure has been proposed for finding kernel functions, let alone finding ones that naturally handle variable length examples etc. This is the topic of the
next section.
4
Kernels from generative probability models: the Fisher
kernel
The key idea here is to derive the kernel function from a generative probability
model. We arrive at the same kernel function from two different perspectives, that
of enhancing the discriminative power of the model and from an attempt to find
a natural comparison between examples induced by the generative model. Both of
these ideas are developed in more detail in the longer version of this paper[4].
We have seen in the previous section that defining the kernel function automatically
implies assumptions about metric relations between the examples. We argue that
these metric relations should be defined directly from a generative probability model
P(XIO). To capture the generative process in a metric between examples we use
the gradient space of the generative model. The gradient of the log-likelihood with
respect to a parameter describes how that parameter contributes to the process of
generating a particular example 5 . This gradient space also naturally preserves all
the structural assumptions that the model encodes about the generation process.
To develop this idea more generally, consider a parametric class of models P(XIO) ,
o E e. This class of probability models defines a Riemannian manifold Ale with
a local metric given by the Fisher information matrix 6 I, where I = Ex{UxU{},
Us = \1 () log P(XIB), and the expectation is over P(XIO) (see e.g. [1]). The gradient
of the log-likelihood , Us , is called the Fisher score, and plays a fundamental role in
our development. The local metric on lvle defines a distance between the current
model P(XIO) and a nearby model P(XIO+J). This distance is given by D(O, 0+15) =
~JT 16, which also approximates the KL-divergence between the two models for a
sufficiently small 6.
The Fisher score Us = \l(} log P(XIB) maps an example X into a feature vector
that is a point in the gradient space of the manifold Ale. We call this the Fisher
score mapping. This gradient Us can be used to define the direction of steepest
ascent in log P(X 10) for the example X along the manifold, i.e. , the gradient in the
direction 6 that maximizes log P( X 10) while traversing the minimum distance in
the manifold as defined by D(O , 0 + 6). This latter gradient is known as the natural
gradient (see e.g. [1]) and is obtained from the ordinary gradient via <Ps = I - I Ux.
We will call the mapping X ~ <Px the natural mapping of examples into feature
vectors 7 . The natural kernel of this mapping is the inner product between these
5For the exponential family of distributions, under the natural parameterization () ,
these gradients, less a normalization constant that depends on () , form sufficient statistics
for the example.
6For simplicity we have suppressed the dependence of I and Ux on the parameter
setting (), or equivalently, on the position in the manifold .
7 Again, we have suppressed dependence on the parameter setting () here.
Exploiting Generative Models in Discriminative Classifiers
491
feature vectors relative to the local Riemannian metric:
(6)
We call this the Fisher kernel owing to the fundamental role played by the Fisher
scores in its definition. The role of the information matrix is less significant; indeed,
in the context of logistic regression models, the matrix appearing in the middle of
the feature vectors relates to the covariance matrix of a Gaussian prior, as show
above. Thus, asymptotically, the information matrix is immaterial, and the simpler
kernel KU(X i , Xj) ex
Ux) provides a suitable substitute for the Fisher kernel.
u.Z,
We emphasize that the Fisher kernel defined above provides only the basic comparison between the examples, defining what is meant by an "inner product" between
the examples when the examples are objects of various t.ypes (e .g. variable length
sequences). The way such a kernel funct.ion is used in a discriminative classifier
is not specified here. Using the Fisher kernel directly in a kernel classifier, for example, amounts to finding a linear separating hyper-plane in the natural gradient.
(or Fisher score) feature space. The examples may not. be linearly separable in this
feature space even though the natural metric st.ructure is given by t.he Fisher kernel.
It may be advantageous to search in the space of quadratic (or higher order) decision boundaries, which is equivalent to transforming the Fisher kernel according to
R(X t , Xj) = (1 + K(X t ? x)))m and using the resulting kernel k in the classifier.
\Ve are now ready to state a few properties of the Fisher kernel function. So long as
the probability model P(XIB) is suitably regular then the Fisher kernel derived from
it is a) a valid kernel function and b) invariant to any invertible (and differentiable)
transformation of the model parameters. The rather informally stated theorem
below motivates the use of this kernel function in a classification setting.
Theorem 1 Given any suitably regular probability model P(XIB) with parameters
B and assuming that the classification label is included as a latent variable, the
Fisher kernel K(X 1 , X)) = V~, I-I Ux] derived from this model and employed in
a kernel classifier is. asymptotically. never inferior to the MAP decision rule from
this model.
The proofs and other related theorems are presented in the longer version of this
paper [4].
To summarize, we have defined a generic procedure for obtaining kernel functions
from generative probability models. Consequently the benefits of generative models are immediately available to the discriminative classifier employing this kernel
function . We now turn the experimental demonstration of the effectiveness of such
a combined classifier.
5
Experimental results
Here we consider two relevant examples from biosequence analysis and compare
the performance of the combined classifier to the best generative models used in
these problems. vVe start with a DNA splice site classification problem, where the
objective is to recognize true splice sites, i.e. , the boundaries between expressed
regions (exons) in a gene and the intermediate regions (introns) . The dat.a set used
in our experiments consisted of 9350 DNA fragments from C. elegans. Each of the
T S. Jaakkola and D. Haussler
492
2029 true examples is a sequence X over the DNA alphabet {A, G, T, C} of length
25; the 7321 false examples are similar sequences that occur near but not at 5'
splice sites. All recognition rates we report on this data set are averages from 7-fold
cross-validation.
To use the combined classifier in this setting requires us to choose a generative
model for the purpose of deriving the kernel function. In order to test how much
the performance of the combined classifier depends on the quality of the underlying
generative model, we chose the poorest model possible. This is the model where
the DKA residue in each position in the fragment is chosen independently of others,
i.e., P(XIB) =
P(XzIBz) and , furthermore , the parameters Bz are set such that
P( Xzl OJ) = 1/4 for all i and all Xl E {A. G, T, C} . This model assigns the same
probability to all examples X. We can still derive the Fisher kernel from such a
model and use it in a discriminative classifier. In this case we used a logistic regression model as in (5) with a quadratic Fisher kernel K(X/. X j ) = (1 + K(Xz, Xj))2.
Figure 1 shows the recognition performance of this kernel method, using the poor
generative model, in comparison to the recognition performance of a naive Bayes
model or a hierarchical mixture model. The comparison is summarized in ROC
style curves plotting false positive errors (the errors of accepting false examples)
as a function of false negative errors (the errors of missing true examples) when
we vary the classification bias for the labels. The curves show that even with such
a poor underlying generative model, the combined classifier is consistently better
than either of the better generative models alone.
n;!l
In the second and more serious application of the combined classifier. we consider
the well-known problem of recognizing remote homologies (evolutionary/structural
similarities) between protein sequences 8 that have low residue identity. Considerable
recent work has been done in refining hidden l\Iarkov models for this purpose as
reviewed in [2], and such models current achieve the best performance. We use
these state-of-the-art HMMs as comparison cases and also as sources for deriving the
kernel function. Here we used logistic regression with the simple kernel K u (X1 ' X J)'
as the number of parameters in the Hj\IMs was several thousand.
The experiment was set up as follows. We picked a particular superfamily (glycosyltransferases) from the TIl'vI-barrel fold in the SCOP protein structure classification
[3], and left out one of the four major families in this superfamily for testing while
training the HMJlvI as well as the combined classifier on sequences corresponding
to the remaining three families . The false training examples for the discriminative
method came from those sequences in the same fold but not in the same superfamily. The test sequences consisted of the left-out family (true examples) and proteins
outside the TIM barrel fold (false examples). The number of training examples varied around 100 depending on the left-out family. As the sequences among the four
glycosyltransferase families are extremely different, this is a challenging discrimination problem. Figure lc shows the recognition performance curves for the HMM
and the corresponding kernel method, averaged over the four-way cross validation.
The combined classifier yields a substantial improvement in performance over the
HJl..IM alone.
8These are variable length sequences thus rendering many discriminative methods
inapplicable.
493
Exploiting Generative Models in Discriminative Classifiers
022
022
02
02
002'
002
0'.
!!O16
;01.
.016
~O ,.
1012
'to 12
i
~ 0'
~
a)
~
0015
00'
0'
~008
0 08
006
006
004
004
0020
002
0 04
006
008
FaN t'leQllttve tata
0'
b)
0020
000'
002
004
006
False I'Wgalllle rate
008
00
0'
0"
c)
06
la1H~_rala
Figure 1: a) & b) Comparison of classification performance between a kernel classifiers from t he uniform model (solid line) and a mixt ure model (dashed line) . In
a) t he mixt ure model is a naive Bayes model and in b) it has t hree components in
each class . c) Comparison of homology recognition performance between a hidden
Mar kov model (das hed line) and t he corresponding kernel classifier (solid line).
6
Discussion
The model based kernel fun ction derived in this pa per provides a generic mechanism
for incorporating generative models into discriminative classifiers. For d iscrimination, the resulting combined classifier is guaranteed to be superior t o the generative
model alone wit h little addi t ional computational cost . Vie not e that t he power of
t he new classifier arises to a large ext.ent from the use of Fisher scores as features
in place of original exa mples. It is possible to use t hese features with any classifier.
e.g. a feed-forward neur al net, but kernel methods are most naturally suited for
incorp orating them .
F inally we note that while we have used classification t o guide the development of
the kernel fun ction , t he results are directly applicable t o regression . clustering. or
even interpolation problems, all of which can easily exploit metric relations among
the examples defined by the Fisher kernel.
References
[1] S.-I. Amari . Natural gradient works efficient ly in learning.
10:251- 276, 1998.
Neural Computation,
[2] R. Durbin , S. Eddy, A. K rogh, a nd G . :\Iitchison . Biological Sequence Analysis: Probabilistic Models of Proteins and Nucleic Aczds. Cambridge C niversity Press, 1998.
[3] T. Hubba rd , A. Murzin , S. Brenner , a nd C. C hothia . seo?: a structural classification
of proteins database. NA R , 25(1) :236- 9, Jan . 199 7.
[4] T. S. Jaakkola a nd D . Haussler .
Exploiting generative models in discriminat ive classifiers.
1998.
Revised and exten ded version . \Vill be available from
http : //r,l'.lY . ai . mit . edu/ rv tommi.
[5] D . J. C. MacKay. Introduction to gaussia n processes.
http : //wol.ra . phy . cam . ac.uk/mackay/.
1997.
Availa ble from
[6] V. Vapnik. The nature of statistical learning theory. Springer-Verlag. 1995.
[7] G. Wahba . Spline models f or observational data. CB:\IS-NSF Regional Conference
Series in Applied t>.lathematics , 1990.
| 1520 |@word middle:1 version:3 advantageous:1 nd:3 suitably:3 covariance:2 solid:2 phy:1 series:1 score:6 fragment:2 current:2 must:1 written:3 cruz:1 treating:1 discrimination:1 alone:4 generative:26 intelligence:1 parameterization:1 plane:1 steepest:1 accepting:1 provides:3 clarified:1 ional:1 simpler:1 along:1 consists:1 shorthand:1 kov:1 combine:1 pairwise:1 ra:1 indeed:2 roughly:1 xz:1 automatically:1 little:1 begin:1 notation:1 underlying:2 maximizes:1 barrel:2 what:1 developed:1 finding:3 transformation:2 pseudo:1 vhich:1 every:1 ti:1 concave:2 fun:2 classifier:30 uk:1 control:1 ly:2 appear:1 positive:2 vie:1 local:3 ext:1 ure:2 interpolation:1 chose:1 specifying:1 challenging:1 hmms:1 discriminatory:1 averaged:1 unique:1 testing:1 definite:1 procedure:2 jan:1 area:1 word:1 refers:1 regular:2 protein:6 get:1 put:1 context:4 equivalent:1 map:2 missing:2 maximizing:1 murzin:1 straightforward:1 independently:1 wit:1 simplicity:1 identifying:1 immediately:1 assigns:1 rule:6 haussler:5 array:1 deriving:3 handle:1 variation:1 justification:1 suppose:1 play:1 mples:1 pa:1 recognition:5 database:1 role:3 capture:1 thousand:1 region:2 remote:1 principled:1 substantial:2 transforming:1 complexity:1 xio:5 cam:1 hese:1 immaterial:1 depend:1 hjl:1 funct:1 inapplicable:1 exon:1 easily:1 various:1 derivation:1 alphabet:1 ction:2 artificial:1 dka:1 hyper:1 outside:1 whose:1 larger:1 solve:1 ive:1 amari:1 statistic:1 addi:1 sequence:13 differentiable:1 net:1 propose:2 product:7 inserting:1 relevant:2 achieve:1 ent:1 exploiting:5 p:1 generating:1 object:1 tim:1 derive:2 develop:2 depending:1 ac:1 predicted:1 come:1 implies:1 tommi:2 direction:2 owing:1 enable:1 wol:1 observational:1 assign:1 discriminat:1 biological:1 im:1 sufficiently:1 around:1 cb:1 mapping:5 major:1 vary:1 purpose:2 applicable:3 label:12 by2:1 pxj:1 seo:1 establishes:1 weighted:1 mit:2 gaussian:3 rather:2 hj:1 bet:1 jaakkola:5 derived:5 refining:1 improvement:2 consistently:1 likelihood:6 posteriori:1 hidden:3 relation:5 transformed:1 overall:1 classification:14 flexible:1 issue:1 dual:1 among:2 development:2 art:1 mackay:2 construct:1 never:1 report:1 others:1 spline:1 serious:1 few:1 preserve:1 ve:4 divergence:1 recognize:1 replaced:1 attempt:1 hed:1 mixture:1 primal:1 biosequence:2 traversing:1 euclidean:2 theoretical:1 maximization:1 ordinary:1 cost:1 uniform:1 recognizing:1 combined:9 st:1 fundamental:2 gaussia:1 systematic:2 probabilistic:2 invertible:1 na:1 again:1 choose:1 possibly:1 exten:1 style:1 til:1 scop:1 summarized:2 includes:1 coefficient:6 depends:2 vi:1 picked:1 start:1 bayes:3 appended:1 square:1 yield:1 classified:1 definition:2 naturally:4 proof:1 riemannian:2 eddy:1 back:1 feed:1 higher:1 done:1 though:1 mar:1 furthermore:1 hand:1 defines:4 logistic:7 quality:1 consisted:2 true:4 homology:2 deal:1 inferior:1 generalized:4 demonstrate:1 mixt:2 superior:3 common:1 he:7 approximates:1 slight:1 ims:1 significant:1 cambridge:2 ai:1 rd:1 had:1 xib:5 similarity:2 longer:2 etc:1 posterior:1 recent:1 perspective:1 verlag:1 binary:1 came:1 seen:1 minimum:1 employed:1 determine:1 dashed:1 semi:1 relates:1 ale:2 ii:2 rv:1 full:1 cross:2 long:2 regression:7 basic:1 vision:1 expectation:1 enhancing:1 bz:1 metric:13 kernel:60 normalization:1 ion:1 addition:1 residue:2 source:1 regional:1 ascent:1 induced:1 thing:1 elegans:1 effectiveness:1 call:3 extracting:1 structural:3 near:1 noting:1 ideal:1 intermediate:1 rendering:1 xj:7 een:1 wahba:1 inner:7 idea:4 hile:1 fullction:1 speech:1 speaking:1 jj:1 generally:2 santa:1 listed:1 informally:1 amount:1 dna:3 http:2 specifies:1 nsf:1 sign:2 per:1 key:1 four:3 achieving:1 asymptotically:2 concreteness:1 sum:2 distorted:1 arrive:1 aile:1 family:6 place:1 decision:4 ble:1 poorest:1 entirely:1 guaranteed:1 played:1 fold:4 quadratic:2 fan:1 exa:1 durbin:1 occur:1 x2:1 encodes:1 nearby:1 argument:1 extremely:1 separable:1 px:3 department:1 according:2 neur:1 combination:2 poor:2 legendre:1 describes:1 suppressed:2 ypes:1 invariant:1 pin:1 discus:1 turn:1 mechanism:1 iap:1 available:2 hierarchical:1 appropriate:1 generic:3 appearing:1 original:3 substitute:1 hree:1 remaining:1 clustering:1 embodies:1 exploit:1 dat:1 objective:1 parametric:1 dependence:2 evolutionary:1 gradient:14 vill:1 distance:4 separating:1 hmm:1 topic:2 manifold:5 argue:1 tata:1 discriminant:1 assuming:1 length:5 illustration:1 demonstration:1 equivalently:2 difficult:2 stated:1 negative:1 motivates:1 availa:1 adjustable:1 nucleic:1 revised:1 defining:3 varied:1 david:1 kl:1 specified:1 optimized:1 connection:1 california:1 below:1 summarize:1 oj:1 power:2 suitable:1 natural:9 technology:1 brief:2 ready:1 extract:1 naive:2 text:1 prior:4 relative:1 loss:2 generation:1 proven:1 validation:2 degree:1 sufficient:1 plotting:1 penalized:3 free:1 bias:2 guide:1 taking:1 superfamily:3 benefit:1 boundary:3 curve:3 valid:3 author:1 forward:1 employing:1 emphasize:1 gene:1 dealing:1 discriminative:15 xi:3 search:1 latent:1 reviewed:1 ku:1 transfer:1 nature:1 ca:1 obtaining:1 contributes:1 da:1 main:1 linearly:1 arise:1 ded:1 complementary:1 x1:1 vve:1 site:3 roc:1 lc:1 position:2 exponential:1 xl:3 weighting:1 ix:2 splice:3 down:1 emphasizing:1 theorem:4 specific:1 jt:1 intron:1 admits:1 incorporating:1 false:7 vapnik:1 importance:1 suited:2 simply:1 expressed:2 ux:4 applies:1 springer:1 corresponds:1 ma:1 conditional:1 identity:1 consequently:3 fisher:20 considerable:1 brenner:1 included:2 specifically:1 called:1 experimental:2 support:4 latter:1 arises:1 meant:1 ex:2 |
571 | 1,521 | SMEM Algorithm for Mixture Models
N aonori U eda Ryohei Nakano
{ueda, nakano }@cslab.kecl.ntt.co.jp
NTT Communication Science Laboratories
Hikaridai, Seika-cho, Soraku-gun, Kyoto 619-0237 Japan
Zoubin Ghahramani Geoffrey E. Hinton
zoubin@gatsby.uc1.ac.uk g.hinton@ucl.ac.uk
Gatsby Computational Neuroscience Unit, University College London
17 Queen Square, London WC1N 3AR, UK
Abstract
We present a split and merge EM (SMEM) algorithm to overcome the local
maximum problem in parameter estimation of finite mixture models. In the
case of mixture models, non-global maxima often involve having too many
components of a mixture model in one part of the space and too few in another, widely separated part of the space. To escape from such configurations
we repeatedly perform simultaneous split and merge operations using a new
criterion for efficiently selecting the split and merge candidates. We apply
the proposed algorithm to the training of Gaussian mixtures and mixtures of
factor analyzers using synthetic and real data and show the effectiveness of
using the split and merge operations to improve the likelihood of both the
training data and of held-out test data.
1
INTRODUCTION
Mixture density models, in particular normal mixtures, have been extensively used
in the field of statistical pattern recognition [1]. Recently, more sophisticated mixture density models such as mixtures of latent variable models (e.g., probabilistic
PCA or factor analysis) have been proposed to approximate the underlying data
manifold [2]-[4]. The parameter of these mixture models can be estimated using the
EM algorithm [5] based on the maximum likelihood framework [3] [4]. A common
and serious problem associated with these EM algorithm is the local maxima problem. Although this problem has been pointed out by many researchers, the best
way to solve it in practice is still an open question.
Two of the authors have proposed the deterministic annealing EM (DAEM) algorithm [6], where a modified posterior probability parameterized by temperature is
derived to avoid local maxima. However, in the case of mixture density models,
local maxima arise when there are too many components of a mixture models in
one part of the space and too few in another. It is not possible to move a component from the overpopulated region to the underpopulated region without passing
N Ueda, R. Nakano, Z. Ghahramani and G. E. Hinton
600
through positions that give lower likelihood. We therefore introduce a discrete move
that simultaneously merges two components in an overpopulated region and splits
a component in an underpopulated region.
The idea of split and merge operations has been successfully applied to clustering
or vector quantization (e.g., [7]). To our knowledge, this is the first time that
simultaneous split and merge operations have been applied to improve mixture
density estimation. New criteria presented in this paper can efficiently select the
split and merge candidates. Although the proposed method, unlike the DAEM
algorithm, is limited to mixture models, we have experimentally comfirmed that our
split and merge EM algorithm obtains better solutions than the DAEM algorithm.
2
Split and Merge EM (SMEM) Algorithm
The probability density function (pdf) of a mixture of M density models is given by
M
p(x; 8) =
L amP(xlwm;Om),
where
am 2: 0 and
(1)
m=l
The p(xlwm ; Om) is a d-dimensional density model corresponding to the component
The EM algorithm, as is well known, iteratively estimates the parameters 8 =
{(am, Om), m = 1, ... , M} using two steps. The E-step computes the expectation
of the complete data log-likelihood.
Wm .
Q(818(t?) =
L L P(wmlx; 8(t?) logamP(xlwm;Om),
X
(2)
m
where P(wmlx; 8(t?) is the posterior probability which can be computed by
P(
Wm
I'
8(t?) =
X,
M
a~p(xlwm;
o~;h
(t)
(t)
Lm'=l am,p(xlwm,;Om')
.
(3)
Next, the M-step maximizes this Q function with respect to 8 to estimate the new
parameter values 8(t+1).
Looking at (2) carefully, one can see that the Q function can be represented in
the form of a direct sum; i.e., Q(818(t?) = L~=l qm(818(t?), where qm(818(t?) =
LXEx P(wmlx ; 8(t?) logamP(xlw m; Om) and depends only on am and Om . Let 8*
denote the parameter values estimated by the usual EM algorithm. Then after the
EM algorithm has converged, the Q function can be rewritten as
Q* = q7
+ q; + qk +
L
q:n.
(4)
m,m,/i,j,k
We then try to increase the first three terms of the right-hand side of (4) by merging
two components Wi and Wj to produce a component Wi', and splitting the component
Wk into two components Wj' and Wk" To reestimate the parameters of these new
components, we have to initialize the parameters corresponding to them using 8*.
The initial parameter values for the merged component Wi' can be set as a linear
combination of the original ones before merge:
and
Oi'
P(wJ
?l x,? 8*)
= O*~
wX P(w ? lx?8*)+O*~
J wX
Lx P(wil x ; 8*) + Lx P(wjlx; 8*)
t
t,
(5)
SMEM Algorithm for Mixture Models
601
On the other hand, as for two components Wj' and
Wk',
we set
(6)
where t is some small random perturbation vector or matrix (i.e., Iltll?IIOk 11)1.
The parameter reestimation for m = i', j' and k' can be done by using EM steps,
but note that the posterior probability (3) should be replaced with (7) so that this
reestimation does not affect the other components.
o~;h
La~p(xlwm;
(I
am,p x m ,
(t)
W
.O(t))
m'
x
L
m'=i,j,k
P(wm'lx; 8*),
m = i',j', k'.
m'=i',j',k'
(7)
Clearly Lm'=i',j',k' P(wm, Ix; 8(t)) = Lm=i,j,k P{wmlx; 8*) always holds during
the reestimation process. For convenience, we call this EM procedure the partial
EM procedure. After this partial EM procedure, the usual EM steps, called the full
EM procedure, are performed as a post processing. After these procedures, if Q
is improved, then we accept the new estimate and repeat the above after setting
the new paramters to 8*. Otherwise reject and go back to 8* and try another
candidate. We summarize these procedures as follows:
[SMEM Algorithm]
1. Perform the usual EM updates. Let 8* and Q* denote the estimated parameters
and corresponding Q function value, respectively.
2. Sort the split and merge candidates by computing split and merge criteria (described in the next section) based on 8*. Let {i, j, k}c denote the cth candidate.
3. For c = 1, ... , C max , perform the following: After initial parameter settings
based on 8 *, perform the partial EM procedure for {i, j, k }c and then perform
the full EM procedure. Let 8** be the obtained parameters and Q** be the
corresponding Q function value. If Q** > Q*, then set Q* f - Q**, 8* f - 8**
and go to Step 2.
4. Halt with 8* as the final parameters.
Note that when a certain split and merge candidate which improves the Q function
value is found at Step 3, the other successive candidates are ignored. There is
therefore no guarantee that the split and the merge candidates that are chosen will
give the largest possible improvement in Q. This is not a major problem, however,
because the split and merge operations are performed repeatedly. Strictly speaking,
C max = M(M -1)(M - 2)/2, but experimentally we have confirmed that C max '" 5
may be enough.
3
Split and Merge Criteria
Each of the split and merge candidates can be evaluated by its Q function value
after Step 3 of the SMEM algorithm mentioned in Sec.2. However, since there
are so many candidates, some reasonable criteria for ordering the split and merge
candidates should be utilized to accelerate the SMEM algorithm.
In general, when there are many data points each of which has almost equal posterior
probabilities for any two components, it can be thought that these two components
1 In the case of mixture Gaussians, covariance matrices E i , and E k , should be positive
definite. In this case, we can initialize them as E i , = E k , = det(Ek)l/d Id indtead of (6).
N. Ueda. R. Nakano. Z. Ghahramani and G. E. Hinton
602
might be merged.
criterion:
To numerically evaluate this, we define the following merge
Jmerge(i,j; 8*) = P i (8*fp j (8*),
(8)
where Pi(8*) = (P(wilxl; 8*), ... , P(wilxN; 8*))T E nN is the N-dimensional
vector consisting of posterior probabilities for the component Wi. Clearly, two components Wi and Wj with larger Jmerge(i,j; 8*) should be merged.
As a split criterion (Jsplit), we define the local Kullback divergence as:
J split (k ; 8- *) =
J(
Pk x; 8*)
- Iog Pk(X;
(I 8*)
e) d x,
P x Wk; k
(9)
which is the distance between two distributions: the local data density Pk(X) around
the component Wk and the density of the component Wk specified by the current
parameter estimate ILk and ~k' The local data density is defined as:
(10)
This is a modified empirical distribution weighted by the posterior probability so
that the data around the component Wk are focused. Note that when the weights
are equal, i.e., P(wklx; 8*) = 11M, (10) is the usual empirical distribution, i.e.,
Pk(X; 8*) = (liN) E~=l 6(x - xn). Since it can be thought that the component
with the largest Jspl it (k; 8*) has the worst estimate of the local density, we should
try to split it. Using Jmerge and Jsp/it, we sort the split and merge candidates as
follows. First, merge candidates are sorted based on Jmerge. Then, for each sorted
merge ' candidate {i,j}e, split candidates excluding {i,j}e are sorted as {k}e. By
combining these results and renumbering them, we obtain {i, j , k }e.
4
4.1
Experiments
Gaussian mixtures
First, we show the results of two-dimensional synthetic data in Fig. 1 to visually
demonstrate the usefulness of the split and merge operations. Initial mean vectors
and covariance matrices were set to near mean of all data and unit matrix, respectively. The usual EM algorithm converged to the local maximum solution shown in
Fig. l(b), whereas the SMEM algorithm converged to the superior solution shown
in Fig. led) very close to the true one. The split of the 1st Gaussian shown in
Fig. l(c) seems to be redundant, but as shown in Fig. led) they are successfully
merged and the original two Gaussians were improved. This indicates that the split
and merge operations not only appropriately assign the number of Gaussians in a
local data space, but can also improve the Gaussian parameters themselves.
Next, we tested the proposed algorithm using 20-dimensional real data (facial images) where the local maxima make the optimization difficult. The data size was 103
for training and 103 for test. We ran three algorithms (EM, DAEM, and SMEM)
for ten different initializations using the K-means algorithm. We set M = 5 and
used a diagonal covariance for each Gaussian. As shown in Table 1, the worst solution found by the SMEM algorithm was better than the best solutions found by
the other algorithms on both training and test data.
603
SMEM Algorithmfor Mixture Models
:'(:0: .
.~.~
. ' " .; ':.. .
\.';,,:
," ...... .
. :~~1 co
-
...';q\'
~.t./
?"{J??
.... O?:6
'..!' " ~:. "
2 .:
"
? i,:"'.....
....:...:
. ,,' :. '
(a) True Gaussians and
generated data
(b) Result by EM (t=72)
.
,' ..
(c) Example of split
and merge (t=141)
?
I
.'
? :: . : :,?? t
, ' ??
(d) Final result by SMEM
(t=212)
Figure 1: Gaussian mixture estimation results.
Table 1: Log-likelihood I data point
-145
Training
data
Test
data
initiall value
EM
DAEM
SMEM
min
-159.1
1.n
-157.3
-163.2
-148.2
0.24
-147.7
-148.6
-147.9
0.04
-147.8
-147.9
-145.1
0.08
-145.0
-145.2
mean
std
max
min
-168.2
2.80
-165.5
-174.2
-159.8
1.00
-158.0
-160.8
-159.7
0.37
-159.6
-159.8
-155.9
0.09
-155.9
-156.0
mean
std
max
'~-150
~
'0
8-'55
?
~
g.-,,,
...J
Table 2: No. of iterations
mean
sId
max
min
EM
DAEM
SMEM
47
16
65
37
147
39
189
103
155
44
219
109
,
EM
-1&5
I
ste~
:
,
,
,
I
with split and'merge
I
:
:
I-------'--~ ,-.----~--~~
,
~
~
..
.. ~ m ~
No. of iterations
~
____
~
~
=
Figure 2: Trajectories of loglikelihood. Upper (lower)
corresponds to training (test) data.
Figure 2 shows log-likelihood value trajectories accepted at Step 3 of the SMEM
algorithm during the estimation process 2. Comparing the convergence points at
Step 3 marked by the '0' symbol in Fig. 2, one can see that the successive split
and merge operations improved the log-likelihood for both the training and test
data, as we expected. Table 2 compares the number of iterations executed by the
three algorithms. Note that in the SMEM algorithm, the EM-steps corresponding
to rejected split and merge operations are not counted. The average rank of the
accepted split and merge candidates was 1.8 (STD=O.9) , which indicates that the
proposed split and merge criteria work very well . Therefore, the SMEM algorithm
was about 155 x 1.8/47 c::: 6 times slower than the original EM algorithm.
4.2
Mixtures of factor analyzers
A mixture of factor analyzers (MFA) can be thought of as a reduced dimension
mixture of Gaussians [4]. That is, it can extract locally linear low-dimensional
manifold underlying given high-dimensional data. A single FA model assumes that
an observed D-dimensional variable x are generated as a linear transformation of
some lower K-dimensionallatent variable z rv N(O, I) plus additive Gaussian noise
v rv N(O, w). w is diagonal. That is, the generative model can be written as
2Dotted lines in Fig. 2 denote the starting points of Step 2. Note that it is due to the
initialization at Step 3 that the log-likelihood decreases just after the split and merge.
N. Ueda, R. Nakano, Z. Ghahrarnani and G. E. Hinton
604
??~~~~--~~?-X~l"?
?~?~~~~--~'"O':',,-x-:.,.
?~~~~--~~?-x~;
(a) Initial values
(b) Result by EM
(c) Result by SMEM
Rgure 3: Extraction of 1D manifold by using a mixture of factor analyzers.
x = Az + v + J-L. Here J-L is a mean vector. Then from simple calculation, we
can see that x
N(J-L, AAT + '11). Therefore, in the case of a M mixture of FAs,
x
L~=l omN(J-Lm, AmA~ + 'lim). See [4] for the details. Then, in this case,
the Q function is also decomposable into M components and therefore the SMEM
algorithm is straightforwardly applicable to the parameter estimation of the MFA
models.
t'V
t'V
Figure 3 shows the results of extracting a one-dimensional manifold from threedimensional data (nOisy shrinking spiral) using the EM and the SMEM algorithms 3.
Although the EM algorithm converged to a poor local maxima, the SMEM algorithm successfully extracted data manifold. Table 3 compares average log-likelihood
per data point over ten different initializations. The log-likelihood values were drastically improved on both training and test data by the SMEM algorithm.
The MFA model is applicable to pattern recognition tasks [2][3] since once an MFA
model is fitted to each class, we can compute the posterior probabilities for each data
point. We tried a digit recognition task (10 digits (classes))4 using the MFA model.
The computed log-likelihood averaged over ten classes and recognition accuracy for
test data are given in Table 4. Clearly, the SMEM algorithm consistently improved
the EM algorithm on both log-likelihood and recognition accuracy. Note that the
recognition accuracy by the 3-nearest neighbor (3NN) classifier was 88.3%. It is
interesting that the MFA approach by both the EM and SMEM algorithms could
outperform the nearest neighbor approach when K = 3 and M = 5. This suggests
that the intrinsic dimensionality of the data would be three or so.
'
3In this case, each factor loading matrix Am becomes a three dimensional column
vector corresponding to each thick line in Fig. 3. More correctly, the center position and
the direction of each thick line are f..Lm and Am, respectively. And the length of each thick
line is 2 IIAmll.
4The data were created using the degenerate Glucksman's feature (16 dimensional data)
by NTT labs.[8]. The data size was 200/class for training and 200/class for test.
605
SMEM Algorithm/or Mixture Models
Table 4: Digit recognition results
Table 3: Log-likelihood I data point
EM
Training -7.68 (0.151)
Test
-7.75 (0.171)
-7.26 (0.017)
EM
SMEM
EM
SMEM
K=3
M=5
M=10
-3.18
-3.09
-3.15
-3.05
89.0
87.5
91.3
88.7
K=8
M=5
M=10
-3.14
-3.04
-3.11
-3.01
85.3
82.5
87.3
85.1
-7.33 (0.032)
O:STD
5
Log-likelihood / data point Recognition rate ("!o)
SMEM
Conclusion
We have shown how simultaneous split and merge operations can be used to move
components of a mixture model from regions of the space in which there are too
many components to regions in which there are too few. Such moves cannot be
accomplished by methods that continuously move components through intermediate
locations because the likelihood is lower at these locations. A simultaneous split and
merge can be viewed as a way of tunneling through low-likelihood barriers, thereby
eliminating many non-global optima. In this respect, it has some similarities with
simulated annealing but the moves that are considered are long-range and are very
specific to the particular problems that arise when fitting a mixture model. Note
that the SMEM algorithm is applicable to a wide variety of mixture models, as long
as the decomposition (4) holds. To make the split and merge method efficient we
have introduced criteria for deciding which splits and merges to consider and have
shown that these criteria work well for low-dimensional synthetic datasets and for
higher-dimensional real datasets. Our SMEM algorithm conSistently outperforms
standard EM and therefore it would be very useful in practice.
References
[1] MacLachlan, G. and Basford K., "Mixture models: Inference and application
to clustering," Marcel Dekker, 1988.
[2] Hinton G. E., Dayan P., and Revow M., "Modeling the minifolds of images of
handwritten digits," IEEE Trans. PAMI, vol.8, no.1, pp. 65-74, 1997.
[3] Tipping M. E. and Bishop C. M., "Mixtures of probabilistic principal component analysers," Tech. Rep. NCRG-97-3, Aston Univ. Birmingham, UK, 1997.
[4] Ghahramani Z. and Hinton G. E., "The EM algorithm for mixtures of factor
analyzers," Tech. Report CRG-TR-96-1, Univ. of Toronto, 1997.
[5] Dempster A. P., Laird N. M. and Rubin D. B., "Maximum likelihood from
incomplete data via the EM algorithm," Journal of Royal Statistical Society
B, vol. 39, pp. 1-38, 1977.
[6] Ueda N. and Nakano R., "Deterministic annealing EM algorithm," Neural Networks, voLl1, no.2, pp.271-282, 1998.
[7] Ueda N. and Nakano R., "A new competitive learning approach based on an
equidistortion principle for designing optimal vector quantizers," Neural Networks, vol.7, no.8, pp.1211-1227, 1994.
[8] Ishii K., "Design of a recognition dictionary using artificially distorted characters," Systems and computers in Japan, vol.21, no.9, pp. 669-677, 1989.
| 1521 |@word eliminating:1 seems:1 loading:1 dekker:1 open:1 tried:1 covariance:3 decomposition:1 thereby:1 tr:1 xlw:1 initial:4 configuration:1 selecting:1 amp:1 outperforms:1 current:1 comparing:1 written:1 additive:1 wx:2 update:1 generative:1 location:2 lx:4 successive:2 toronto:1 direct:1 ryohei:1 fitting:1 introduce:1 expected:1 themselves:1 seika:1 becomes:1 underlying:2 maximizes:1 transformation:1 guarantee:1 qm:2 classifier:1 uk:4 unit:2 before:1 positive:1 aat:1 local:12 id:1 merge:34 pami:1 might:1 plus:1 initialization:3 suggests:1 co:2 limited:1 range:1 averaged:1 practice:2 definite:1 digit:4 procedure:8 empirical:2 reject:1 thought:3 zoubin:2 convenience:1 close:1 cannot:1 deterministic:2 center:1 go:2 starting:1 focused:1 decomposable:1 splitting:1 designing:1 recognition:9 utilized:1 std:4 q7:1 observed:1 worst:2 region:6 wj:5 ordering:1 decrease:1 ran:1 mentioned:1 dempster:1 wil:1 accelerate:1 represented:1 separated:1 univ:2 london:2 analyser:1 widely:1 solve:1 larger:1 loglikelihood:1 otherwise:1 noisy:1 laird:1 final:2 ucl:1 dimensionallatent:1 ste:1 combining:1 degenerate:1 ama:1 az:1 convergence:1 optimum:1 jsp:1 produce:1 ac:2 nearest:2 marcel:1 direction:1 thick:3 merged:4 assign:1 crg:1 strictly:1 hold:2 around:2 considered:1 normal:1 visually:1 deciding:1 lm:5 major:1 dictionary:1 estimation:5 birmingham:1 applicable:3 largest:2 successfully:3 weighted:1 clearly:3 gaussian:7 always:1 modified:2 avoid:1 derived:1 improvement:1 consistently:2 rank:1 likelihood:17 indicates:2 tech:2 ishii:1 am:7 inference:1 dayan:1 nn:2 accept:1 initialize:2 field:1 equal:2 once:1 having:1 extraction:1 report:1 escape:1 few:3 serious:1 simultaneously:1 divergence:1 replaced:1 consisting:1 mixture:33 held:1 wc1n:1 partial:3 facial:1 incomplete:1 fitted:1 column:1 modeling:1 ar:1 queen:1 usefulness:1 too:6 straightforwardly:1 synthetic:3 cho:1 st:1 density:11 probabilistic:2 continuously:1 daem:6 ek:1 japan:2 sec:1 wk:7 depends:1 performed:2 try:3 lab:1 wm:4 sort:2 competitive:1 om:7 square:1 oi:1 accuracy:3 qk:1 efficiently:2 sid:1 handwritten:1 trajectory:2 confirmed:1 researcher:1 converged:4 simultaneous:4 reestimate:1 eda:1 pp:5 associated:1 basford:1 knowledge:1 lim:1 improves:1 dimensionality:1 sophisticated:1 carefully:1 back:1 higher:1 tipping:1 improved:5 done:1 evaluated:1 rejected:1 just:1 hand:2 true:2 laboratory:1 iteratively:1 during:2 criterion:10 pdf:1 complete:1 demonstrate:1 temperature:1 image:2 recently:1 common:1 superior:1 jp:1 ncrg:1 numerically:1 pointed:1 analyzer:5 similarity:1 posterior:7 certain:1 rep:1 accomplished:1 redundant:1 rv:2 full:2 kyoto:1 ntt:3 calculation:1 renumbering:1 lin:1 long:2 post:1 halt:1 iog:1 expectation:1 iteration:3 whereas:1 annealing:3 appropriately:1 unlike:1 effectiveness:1 call:1 extracting:1 near:1 intermediate:1 split:37 enough:1 spiral:1 variety:1 affect:1 idea:1 det:1 pca:1 soraku:1 passing:1 speaking:1 repeatedly:2 algorithmfor:1 ignored:1 useful:1 involve:1 extensively:1 ten:3 locally:1 reduced:1 outperform:1 dotted:1 neuroscience:1 estimated:3 per:1 correctly:1 paramters:1 discrete:1 vol:4 sum:1 parameterized:1 distorted:1 almost:1 reasonable:1 ueda:6 tunneling:1 ilk:1 min:3 cslab:1 combination:1 poor:1 em:38 character:1 wi:5 cth:1 operation:10 rewritten:1 gaussians:5 apply:1 slower:1 original:3 assumes:1 clustering:2 nakano:7 ghahramani:4 hikaridai:1 threedimensional:1 society:1 move:6 question:1 fa:2 usual:5 diagonal:2 distance:1 simulated:1 gun:1 manifold:5 length:1 difficult:1 executed:1 quantizers:1 design:1 perform:5 upper:1 datasets:2 finite:1 hinton:7 communication:1 looking:1 excluding:1 perturbation:1 introduced:1 specified:1 merges:2 trans:1 pattern:2 fp:1 summarize:1 max:6 royal:1 mfa:6 improve:3 smem:30 aston:1 created:1 extract:1 interesting:1 geoffrey:1 rubin:1 kecl:1 principle:1 pi:1 repeat:1 drastically:1 side:1 neighbor:2 wide:1 barrier:1 overcome:1 dimension:1 xn:1 computes:1 author:1 counted:1 approximate:1 obtains:1 kullback:1 global:2 reestimation:3 latent:1 table:8 artificially:1 pk:4 noise:1 arise:2 fig:8 gatsby:2 shrinking:1 position:2 candidate:16 ix:1 specific:1 bishop:1 symbol:1 intrinsic:1 quantization:1 merging:1 led:2 corresponds:1 extracted:1 sorted:3 marked:1 viewed:1 revow:1 experimentally:2 principal:1 overpopulated:2 called:1 accepted:2 la:1 select:1 college:1 underpopulated:2 evaluate:1 tested:1 lxex:1 |
572 | 1,522 | An entropic estimator for structure discovery
Matthew Brand
Mitsubishi Electric Research Laboratories, 201 Broadway, Cambridge MA 02139
brand@merl.com
Abstract
We introduce a novel framework for simultaneous structure and parameter learning in
hidden-variable conditional probability models, based on an en tropic prior and a solution
for its maximum a posteriori (MAP) estimator. The MAP estimate minimizes uncertainty
in all respects: cross-entropy between model and data; entropy of the model ; entropy
of the data's descriptive statistics. Iterative estimation extinguishes weakly supported
parameters, compressing and sparsifying the model. Trimming operators accelerate this
process by removing excess parameters and, unlike most pruning schemes, guarantee
an increase in posterior probability. Entropic estimation takes a overcomplete random
model and simplifies it, inducing the structure of relations between hidden and observed
variables. Applied to hidden Markov models (HMMs), it finds a concise finite-state
machine representing the hidden structure of a signal. We entropically model music,
handwriting, and video time-series, and show that the resulting models are highly concise,
structured, predictive, and interpretable: Surviving states tend to be highly correlated
with meaningful partitions of the data, while surviving transitions provide a low-perplexity
model of the signal dynamics.
1 . An entropic prior
In entropic estimation we seek to maximize the information content of parameters. For
conditional probabilities , parameters values near chance add virtually no information
to the model, and are therefore wasted degrees of freedom. In contrast, parameters
near the extrema {O, I} are informative because they impose strong constr?aints on the
class of signals accepted by the model. In Bayesian terms, our prior should assert that
parameters that do not reduce uncertainty are improbable. We can capture this intuition in
a surprisingly simple form: For a model of N conditional probabilities 9 = {(h , . . . , ()N }
we write
(1)
whence we can see that the prior measures a model's freedom from ambiguity (H(9) is an
entropy measure). Applying Pe (.) to a multinomial yields the posterior
p (LlI)
e
U
W
<X
P(wI9)Pe(9)
P(w)
<X
[II
.
z
()W;]
1
Pe(9)
P(w)
<X
II
.
(){;/;+w;
z
(2)
z
where Wi is evidence for event type i. With extensive evidence this distribution converges
to "fair"(ML) odds for w, but with scant evidence it skews to stronger odds.
M Brand
724
1.1
MAP estimator
To obtain MAP estimates we set the derivative of log-posterior to zero, using Lagrange
multipliers to ensure L:i (}i = 1,
W'
(3)
1+ (); +IOg(}i+"\
We obtain (}i by working backward from the Lambert W function, a multi-valued inverse
function satisfying W(x)eW(x) =x. Taking logarithms and setting y = logx,
0= -W(x) -logW(x) + logx
- W(e Y ) -log W(e Y ) + y
-1
IjW(e Y ) +logljW(e Y )+logz+y-Iogz
-z
zjW(e Y ) + 10gzjW(e Y ) + y -logz
(4)
Setting (}i = zjW(e Y ) , y = 1 +"\+logz, and z = -Wi, eqn. 4 simplifies to eqn. 3, implying
(5)
Equations 3 and 5 together yield a quickly converging fix-point equation for ..\ and therefore
for the entropic MAP estimate. Solutions lie in the W -1 branch of Lambert's function. See
[Brand, 1997] for methods we developed to calculate the little-known W function.
1.2
Interpretation
The negated log-posterior is equivalent to a sum of entropies:
H(O)
+ D(wIIO) + H(w)
(6)
Maximizing Pe(Olw) minimizes entropy in all respects: the parameter entropy H(O); the
cross-entropy D (w "0) between the parameters 0 and the data's descriptive statistics w;
and the entropy of those statistics H (w), which are calculated relative to the structure
of the model. Equivalently, the MAP estimator minimizes the expected coding length,
making it a maximally efficient compressor of messages consisting of the model and the
data coded relative to the model. Since compression involves separating essential from
accidental structure, this can be understood as a form of noise removal. Noise inflates the
apparent entropy of a sampled process; this systematically biases maximum likelihood
(ML) estimates toward weaker odds, more so in smaller samples. Consequently, the
entropic prior is a countervailing bias toward stronger odds.
725
An Entropic Estimator for Structure Discovery
1.3
Model trimming
Because the prior rewards sparse models, it is possible to remove weakly supported
parameters from the model while improving its posterior probability, such that
P e (O\(}iIX) > P e (OIX). This stands in contrast to most pruning schemes, which typically
try to minimize damage to the posterior. Expanding via Bayes rule and taking logarithms
we obtain
(7)
where hi((}i) is the entropy due to (}i. For small (}i, we can approximate via differentials:
() . {)H(O)
~
{)(}i
>
() . {)logP(XIO)
t
{)(} i
(8)
By mixing the left- and right-hand sides of equations 7 and 8, we can easily identify
trimmable parameters-those that contribute more to the entropy than the log-likelihood.
E.g., for multinomials we set hi ((}i) = -(}i log (}i against r.h.s. eqn. 8 and simplify to obtain
<
exp [ -
{)logP(XIO)]
{)(}i
(9)
Parameters can be trimmed at any time during training; at convergence trimming can
bump the model out of a local probability maximum, allowing further training in a lowerdimensional and possibly smoother parameter subspace.
2 Entropic HMM training and trimming
In entropic estimation of HMM transition probabilities, we follow the conventional E-step,
calculating the probability mass for each transition to be used as evidence w:
T-l
Ij,i
L aj(t) Pilj Pi(Xt+1) fh(t + 1)
(10)
where P ilJ is the current estimate of the transition probability from state j to state i;
Pi(Xt+d is the output probability of observation Xt+1 given state i, and Q, {3 are obtained
from forward-backward analysis and follow the notation of Rabiner [1989]. For the Mstep, we calculate new estimates {Pi lj h = 0 by applying the MAP estimator in ?1.1 to
each w = {,j ,i k That is, w is a vector of the evidence for each kind of transition out of
a single state; from this evidence the MAP estimator calculates probabilities O. (In BaumWelch re-estimation, the maximum-likelihood estimator simply sets Pilj = Ij ,i/ 2:i Ij,d
In iterative estimation, e.g., expectation-maximization (EM), the entropic estimator drives
weakly supported parameters toward zero, skeletonizing the model and concentrating
evidence on surviving parameters until their estimates converge to near the ML estimate.
Trimming appears to accelerate this process by allowing slowly dying parameters to
leapfrog to extinction. It also averts numerical underflow errors.
For HMM transition parameters, the trimming criterion of egn. 9 becomes
(11 )
where Ij (t) is the probability of state j at time t. The multinomial output distributions of a
discrete-output HMM can be en tropically re-estimated and trimmed in the same manner.
M. Brand
726
Entropic versus ML HMM models of Bach chorales
90
7S .-.-~-.-----,...,
go
r:\\~
~~
o
S
20
5
IS
25
3S
.t5.....,5----+.-.,-~
, states at initialization
Figure 1: Left: Sparsification, classification, and prediction superiority of entropically
estimated HMMs modeling Bach chorales. Lines indicate mean performance over 10
trials; error bars are 2 standard deviations . Right: High-probability states and subgraphs of
interest from an entropically estimated 35-state chorale HMM. Tones output by each state
are listed in order of probability. Extraneous arcs have been removed for clarity.
3
Structure learning experiments
To explore the practical utility of this framework, we will use entropically estimated HMMs
as a window into the hidden structure of some human-generated time-series.
Bach Chorales: We obtained a dataset of melodic lines from 100 of I.S. Bach's 371
surviving chorales from the UCI repository [Merz and Murphy, 1998], and transposed all
into the key of C. We compared entropically and conventionally estimated HMMs in
prediction and classification tasks, training both from identical random initial conditions
and trying a variety of different initial state-counts. We trained with 90 chorales and
testing with the remaining 10. In ten trials, all chorales were rotated into the test
set. Figure 1 illustrates that despite substantial loss of parameters to sparsification, the
entropically estimated HMMs were, on average, better predictors of notes. (Each test
sequence was truncated to a random length and the HMMs were used to predict the first
missing note.) They also were better at discriminating between test chorales and temporally
reversed test chorales-challenging because Bach famously employed melodic reversal as a
compositional device. With larger models, parameter-trimming became state-trimming: An
average of 1.6 states were "pinched off" the 35-state models when all incoming transitions
were deleted.
While the conventionally estimated HMMs were wholly uninterpretable, in the entropically
estimated HMMs one can discern several basic musical structures (figure 1, right),
including self-transitioning states that output only tonic (C-E-G) or dominant (G-B-D)
triads, lower- or upper-register diatonic tones (C-D-E or F-G-A-B), and mordents (A-nGA). We also found chordal state sequences (F-A-C) and states that lead to the tonic (C) via
the mediant (E) or the leading tone (B).
Handwriting: We used 2D Gaussian-output HMMs to analyze handwriting data. Training
data, obtained from the UNIPEN web site [Reynolds, 1992], consisted of sequences of
normalized pen-position coordinates taken at 5msec intervals from 10 different individuals
writing the digits 0-9. The HMMs were estimated from identical data and initial conditions
(random upper-diagonal transition matrices; random output parameters). The diagrams
in Figure 2 depict transition graphs of two HMMs modeling the pen-strokes for the digit
"5," mapped onto the data. Ellipses indicate each state's output probability iso-contours
(receptive field); X s and arcs indicate state dwell and transition probabilities, respectively,
by their thicknesses. Entropic estimation induces an interpretable automaton that captures
essential structure and timing of the pen-strokes. 50 of the 80 original transition parameters
727
An Entropic Estimator for Structure Discovery
ConlUStOn MatrIX WIth 93 0% acct.JIlIcy
~~Y",
:,y
.
..
.
eonrus.on Matnll WIth 96 0%,accuracy
.
'.
:,"' .s -':
"
.,-!.:~ ,~_~~"
/
. -,
6
a. conventional
b. en tropic
c. conventional
d. en tropic
Figure 2: (a & b): State machines of conventionally and entropically estimated hidden
Markov models of writing "S." (c & d): Confusion matrices for all digits.
were trimmed. Estimation without the entropic prior results in a wholly opaque model, in
which none of the original dynamical parameters were trimmed. Model concision leads to
better classification-the confusion matrices show cumulative classificMion error over ten
trials with random initializations. Inspection of the parameters for the model in 2b showed
that all writers began in states 1 or 2. From there it is possible to follow the state diagram
to reconstruct the possible sequences of pen-strokes: Some writers start with the cap (state
1) while others start with the vertical (state 2); all loop through states 3-8 and some return
to the top (via state 10) to add a horizontal (state 12) or diagonal (state 11) cap.
Office activity: Here we demonstrate a model of human activity learned from mediumto long-term ambient video. By activity, we mean spatio-temporal patterns in the pose,
position, and movement of one's body. To make the vision tractable, we consider the
activity of a single person in a relatively stable visual environment, namely, an office.
We track the gross shape and position of the office occupant by segmenting each image
into foreground and background pixels. Foreground pixels are identified with reference
to an acquired statistical model of the background texture and camera noise. Their
ensemble properties such as motion or color are modeled via adaptive multivariate
Gaussian distributions, re-estimated in each frame.
A single bivariate Gaussian is
fitted to the foreground pixels and we record the associated ellipse parameters [mean x ,
meany, timean x , timean y, mass, timass, elongation, eccentricity]. Sequences of these
observation vectors are used to train and test the HMMs.
Approximately 30 minutes of data were taken at SHz from an SGI IndyCam. Data
was collected automatically and at random over several days by a program that started
recording whenever someone entered the room after it had been empty S+ minutes.
Backgrounds were re-Iearned during absences to accommodate changes in lighting and
room configuration. Prior to training, HMM states were initialized to tile the image
with their receptive fields, and transition probabilities were initialized to prefer motion
to adjoining tiles. Three sequences ranging from 1000 to 1900 frames in length were used
for entropic training of 12, 16,20, 2S, and 30-state HMMs.
Entropic training yielded a substantially sparsified model with an easily interpreted state
machine (see figure 3). Grouping of states into activities (done only to improve readability)
was done by adaptive clustering on a proximity matrix which combined Mahalonobis
distance and transition probability between states. The labels are the author's description
of the set of frames claimed by each state cluster during forward-backward analysis of
test data. Figure 4 illustrates this analysis, showing frames from a test sequence to which
specific states are strongly tuned. State S (figure 3 right) is particularly interesting-it has a
very non-specific receptive field, no self-transition, and an extremely low rate of occupancy.
Instead of modeling data, it serves to compress the model by summarizing transition
patterns that are common to several other states. The entropic model has proven to be
quite superior for segmented new video into activities and detecting anomalous behavior.
M Brand
728
-~ --. ~ ...".".
:~
irlitialization
..
.-.
tinalmochtl
.
~~
.
'
Figure 3: Top: The state machine found by en tropic training (left) is easily labeled and
interpreted. The state machine found by conventional training (right) is not, begin fully
connected. Bottom: Transition matrices after (1) initialization, (2) entropic training, (3)
conventional training, and (4 & 5) entropic training from larger initializations. The top row
indicates initial probabilities of each state; each subsequent row indicates the transition
probabilities out of a state. Color key: 0 = 0; ? = 1. The state machines above are
extracted from 2 & 3. Note that 4 & 5 show the same qualitative structure as 2, but sparser,
while 3 shows no almost no structure at all.
Figure 4: Some sample frames assigned high state-specific probabilities by the model. Note
that some states are tuned to velocities, hence the difference between states 6 and 11.
4
Related work
HMMs: The literature of structure-learning in HMMs is based almost entirely on generateand-test algorithms. These algorithms work by merging [Stokke and Omohundro, 1994]
or splitting [Takami and Sagayama, 1991] states, then retraining the model to see if any
advantage has been gained. Space constraints force us to summarize a recent literature
review: There are now more than 20 variations and improvements on these approaches, plus
some heuristic constructive algorithms (e.g., [Wolfertstetter and Ruske, 1995]). Though
these efforts use a variety of heuristic techniques and priors (including MDL) to avoid
detrimental model changes, much of the computation is squandered and reported run-times
often range from hours to days. Entropic estimation is exact. monotonic, and orders of
magnitude faster-only slightly longer than standard EM parameter estimation.
MDL: Description length minimization is typically done via gradient ascent or search via
model comparison; few estimators are known. Rissanen [1989] introduced an estimator for
binary fractions, from which Vovk [1995] derived an approximate estimator for Bernoulli
An Entropic Estimator for Structure Discovery
729
models over discrete sample spaces. It approximates a special case of our exact estimator,
which handles multinomial models in continuous sample spaces. Our framework provides
a unified Bayesian framework for two issues that are often treated separately in MDL:
estimating the number of parameters and estimating their values.
MaxEnt: Our prior has different premises and an effect opposite that of the "standard"
MaxEnt prior e- aD (9i1 9 o). Nonetheless, our prior can be derived via MaxEnt reasoning
from the premise that the expectation of the perplexity over all possible models is finite
[Brand, 1998]. More colloquially, we almost always expect there to be learnable structure.
Extensions: For simplicity of exposition (and for results that are independent of model
class), we have assumed prior independence of the parameters and taken H (8) to be the
combined parameter entropies of the model's component distributions. Depending on the
model class, we can also provide variants of eqns. 1-8 for H (8) =conditional entropy or
H (8) =entropy rate of the model. In Brand [1998] we present entropic MAP estimators
for spread and covariance parameters with applications to mixtures-of-Gaussians, radial
basis functions, and other popular models. In the same paper we generalize eqns. 1-8
with a temperature term, obtaining a MAP estimator that minimizes the free energy of the
model. This folds deterministic annealing into EM, turning it into a quasi-global optimizer.
It also provides a workaround for one known limitation of entropy minimization: It is
inappropriate for learning from data that is atypical of the source process.
Open questions: Our framework is currently agnostic w.r.t. two important questions: Is
there an optimal trimming policy? Is there a best entropy measure? Other questions
naturally arise: Can we use the entropy to estimate the peakedness of the posterior
distribution, and thereby judge the appropriateness of MAP models? Can we also directly
minimize the entropy of the hidden variables, thereby obtaining discriminant training?
5
Conclusion
Entropic estimation is highly efficient hillclimbing procedure for simultaneously estimating
model structure and parameters. It provides a clean Bayesian framework for minimizing all
entropies associated with modeling, and an E-MAP algorithm that brings the structure of a
randomly initialized model into alignment with hidden structures in the data via parameter
extinction . The applications detailed here are three of many in which entropically estimated
models have consistently outperformed maximum likelihood models in classification and
prediction tasks. Most notably, it tends to produce interpretable models that shed light on
the structure of relations between hidden variables and observed effects.
References
Brand, M. (1997). Structure discovery in conditional probability models via an entropic prior and
parameter extinction. NeuraL Computation To appear; accepted 8/98.
Brand, M. (1998). Pattern discovery via entropy minimization. To appear in Proc .. ArtificiaL
Intelligence and Statistics #7.
Merz, C. and Murphy, P. (1998). UCI repository of machine learning databases.
Rabiner, L. R. (1989). A tutorial on hidden Markov models and selected applications in speech
recognition. Proceedings of the IEEE, 77(2):257-286.
Reynolds, D. (1992). Handwritten digit data. UNIPEN web site, hUp:llhwr.nici.kun.nl/unipen/.
Donated by HP Labs, Bristol, England.
Rissanen, J. (1989). Stochastic CompLexit)' and StatisticaL Inquiry. World Scientific.
Stolcke, A. and Omohundro, S. (1994). Best-first model merging for hidden Markov model induction.
TR-94-003, International Computer Science Institute, U.c. Berkeley.
Takami, 1.-1. and Sagayama, S. (1991). Automatic generation of the hidden Markov model by
successive state splitting on the contextual domain and the temporal domain. TR SP91-88, IEICE.
Vovk, V. G. (1995). Minimum description length estimators under the optimal coding scheme. In
Vitanyi, P., editor, Proc. ComputationaL Learning Theory / Europe, pages 237-251. Springer-Verlag.
Wolfertstetter, F. and Ruske, G. (1995). Structured Markov models for speech recognition. In
InternationaL Conference on Acoustics. Speech. and SignaL Processing, volume I, pages 544-7.
| 1522 |@word trial:3 repository:2 compression:1 stronger:2 retraining:1 extinction:3 open:1 seek:1 mitsubishi:1 covariance:1 concise:2 thereby:2 tr:2 accommodate:1 initial:4 configuration:1 series:2 complexit:1 tuned:2 reynolds:2 current:1 com:1 contextual:1 chordal:1 numerical:1 partition:1 informative:1 subsequent:1 shape:1 mstep:1 remove:1 interpretable:3 depict:1 implying:1 intelligence:1 selected:1 device:1 tone:3 inspection:1 iso:1 record:1 detecting:1 provides:3 contribute:1 readability:1 successive:1 differential:1 qualitative:1 manner:1 introduce:1 acquired:1 notably:1 expected:1 behavior:1 multi:1 ijw:1 automatically:1 little:1 window:1 inappropriate:1 becomes:1 begin:1 estimating:3 notation:1 colloquially:1 mass:2 agnostic:1 kind:1 interpreted:2 minimizes:4 substantially:1 developed:1 dying:1 unified:1 extremum:1 sparsification:2 guarantee:1 assert:1 temporal:2 berkeley:1 donated:1 shed:1 superiority:1 appear:2 segmenting:1 understood:1 local:1 timing:1 tends:1 despite:1 approximately:1 plus:1 extinguishes:1 initialization:4 challenging:1 someone:1 hmms:15 range:1 practical:1 camera:1 testing:1 digit:4 procedure:1 logz:3 wholly:2 scant:1 radial:1 melodic:2 onto:1 operator:1 applying:2 writing:2 equivalent:1 map:12 conventional:5 missing:1 maximizing:1 deterministic:1 go:1 automaton:1 simplicity:1 splitting:2 subgraphs:1 estimator:18 rule:1 egn:1 handle:1 coordinate:1 variation:1 exact:2 iearned:1 velocity:1 satisfying:1 particularly:1 recognition:2 labeled:1 database:1 observed:2 bottom:1 capture:2 calculate:2 compressing:1 connected:1 triad:1 movement:1 removed:1 substantial:1 intuition:1 environment:1 gross:1 workaround:1 xio:2 reward:1 concision:1 dynamic:1 trained:1 weakly:3 predictive:1 writer:2 basis:1 accelerate:2 easily:3 train:1 mahalonobis:1 artificial:1 apparent:1 quite:1 larger:2 valued:1 heuristic:2 reconstruct:1 skews:1 statistic:4 descriptive:2 sequence:7 advantage:1 uci:2 loop:1 entered:1 mixing:1 description:3 inducing:1 sgi:1 convergence:1 empty:1 eccentricity:1 cluster:1 produce:1 converges:1 rotated:1 depending:1 pose:1 ij:4 strong:1 involves:1 indicate:3 judge:1 appropriateness:1 stochastic:1 human:2 premise:2 fix:1 extension:1 proximity:1 exp:1 predict:1 bump:1 matthew:1 optimizer:1 entropic:24 fh:1 estimation:11 proc:2 outperformed:1 label:1 currently:1 minimization:3 gaussian:3 always:1 avoid:1 office:3 derived:2 leapfrog:1 improvement:1 consistently:1 bernoulli:1 likelihood:4 indicates:2 contrast:2 summarizing:1 whence:1 posteriori:1 typically:2 lj:1 hidden:12 relation:2 quasi:1 i1:1 pixel:3 issue:1 classification:4 extraneous:1 special:1 field:3 elongation:1 identical:2 foreground:3 others:1 logx:2 simplify:1 few:1 inflates:1 randomly:1 simultaneously:1 individual:1 murphy:2 consisting:1 freedom:2 interest:1 trimming:9 message:1 highly:3 alignment:1 mdl:3 mixture:1 nl:1 light:1 adjoining:1 ambient:1 improbable:1 logarithm:2 initialized:3 re:4 maxent:3 overcomplete:1 fitted:1 merl:1 modeling:4 logp:2 maximization:1 deviation:1 predictor:1 reported:1 thickness:1 combined:2 person:1 international:2 discriminating:1 off:1 together:1 quickly:1 ilj:1 ambiguity:1 possibly:1 slowly:1 tile:2 derivative:1 leading:1 return:1 coding:2 register:1 ad:1 try:1 lab:1 analyze:1 start:2 bayes:1 minimize:2 accuracy:1 became:1 musical:1 ensemble:1 yield:2 identify:1 rabiner:2 generalize:1 bayesian:3 handwritten:1 lambert:2 lli:1 none:1 occupant:1 lighting:1 drive:1 bristol:1 stroke:3 simultaneous:1 inquiry:1 whenever:1 against:1 nonetheless:1 energy:1 naturally:1 associated:2 transposed:1 handwriting:3 sampled:1 dataset:1 concentrating:1 popular:1 color:2 cap:2 appears:1 day:2 follow:3 maximally:1 done:3 though:1 strongly:1 until:1 working:1 eqn:3 hand:1 web:2 horizontal:1 brings:1 aj:1 scientific:1 ieice:1 effect:2 consisted:1 multiplier:1 normalized:1 hence:1 assigned:1 laboratory:1 during:3 self:2 eqns:2 criterion:1 trying:1 omohundro:2 demonstrate:1 confusion:2 motion:2 temperature:1 reasoning:1 image:2 ranging:1 novel:1 began:1 common:1 superior:1 multinomial:4 volume:1 interpretation:1 approximates:1 cambridge:1 automatic:1 hp:1 had:1 stable:1 europe:1 longer:1 add:2 dominant:1 posterior:7 multivariate:1 showed:1 recent:1 perplexity:2 claimed:1 verlag:1 binary:1 minimum:1 lowerdimensional:1 impose:1 nici:1 employed:1 converge:1 maximize:1 signal:4 ii:2 branch:1 smoother:1 segmented:1 faster:1 england:1 cross:2 bach:5 long:1 coded:1 iog:1 ellipsis:1 calculates:1 converging:1 prediction:3 basic:1 anomalous:1 variant:1 vision:1 expectation:2 background:3 separately:1 interval:1 annealing:1 diagram:2 source:1 unlike:1 ascent:1 recording:1 tend:1 virtually:1 surviving:4 odds:4 peakedness:1 near:3 stolcke:1 indycam:1 variety:2 independence:1 identified:1 opposite:1 reduce:1 simplifies:2 utility:1 trimmed:4 effort:1 speech:3 compositional:1 detailed:1 listed:1 pinched:1 ten:2 takami:2 induces:1 tutorial:1 estimated:12 track:1 write:1 discrete:2 diatonic:1 sparsifying:1 key:2 rissanen:2 deleted:1 clarity:1 clean:1 backward:3 wasted:1 graph:1 fraction:1 sum:1 nga:1 baumwelch:1 inverse:1 run:1 uncertainty:2 opaque:1 discern:1 almost:3 prefer:1 uninterpretable:1 entirely:1 hi:2 dwell:1 vitanyi:1 accidental:1 fold:1 yielded:1 activity:6 constraint:1 extremely:1 relatively:1 structured:2 smaller:1 slightly:1 em:3 wi:2 constr:1 making:1 taken:3 equation:3 count:1 tractable:1 reversal:1 serf:1 gaussians:1 original:2 compress:1 top:3 remaining:1 ensure:1 clustering:1 iix:1 calculating:1 music:1 ellipse:1 question:3 olw:1 damage:1 receptive:3 diagonal:2 gradient:1 detrimental:1 subspace:1 reversed:1 pilj:2 mapped:1 separating:1 distance:1 hmm:7 collected:1 discriminant:1 toward:3 induction:1 length:5 modeled:1 minimizing:1 equivalently:1 kun:1 tropic:4 broadway:1 policy:1 negated:1 allowing:2 upper:2 vertical:1 observation:2 markov:6 arc:2 finite:2 truncated:1 sparsified:1 tonic:2 frame:5 introduced:1 namely:1 extensive:1 acoustic:1 learned:1 hour:1 bar:1 dynamical:1 pattern:3 summarize:1 program:1 including:2 video:3 event:1 treated:1 force:1 turning:1 representing:1 scheme:3 chorale:9 improve:1 occupancy:1 temporally:1 started:1 conventionally:3 logw:1 entropically:9 prior:14 literature:2 discovery:6 removal:1 review:1 relative:2 loss:1 fully:1 expect:1 interesting:1 limitation:1 generation:1 proven:1 versus:1 degree:1 editor:1 systematically:1 famously:1 pi:3 row:2 supported:3 surprisingly:1 free:1 bias:2 weaker:1 side:1 institute:1 taking:2 stokke:1 sparse:1 calculated:1 transition:17 stand:1 contour:1 t5:1 cumulative:1 forward:2 author:1 adaptive:2 world:1 excess:1 pruning:2 approximate:2 ml:4 global:1 incoming:1 assumed:1 spatio:1 search:1 iterative:2 pen:4 continuous:1 expanding:1 obtaining:2 improving:1 electric:1 sagayama:2 domain:2 spread:1 noise:3 arise:1 fair:1 body:1 site:2 en:5 position:3 msec:1 lie:1 pe:4 atypical:1 removing:1 minute:2 transitioning:1 xt:3 specific:3 showing:1 learnable:1 evidence:7 bivariate:1 essential:2 grouping:1 merging:2 gained:1 texture:1 magnitude:1 illustrates:2 sparser:1 entropy:21 simply:1 explore:1 visual:1 hillclimbing:1 lagrange:1 compressor:1 monotonic:1 springer:1 underflow:1 chance:1 extracted:1 ma:1 conditional:5 consequently:1 exposition:1 room:2 absence:1 content:1 change:2 vovk:2 accepted:2 oix:1 merz:2 brand:10 meaningful:1 ew:1 hup:1 constructive:1 correlated:1 |
573 | 1,523 | Optimizing Classifiers for Imbalanced
Training Sets
Grigoris Karakoulas
Global Analytics Group
Canadian Imperial Bank of Commerce
161 Bay St., BCE-ll,
Toronto ON, Canada M5J 2S8
Email: karakoulOcibc.ca
John Shawe-Taylor
Department of Computer Science
Royal Holloway, University of London
Egham, TW20 OEX
England
Email: jstOdcs.rhbnc.ac.uk
Abstract
Following recent results [9 , 8] showing the importance of the fatshattering dimension in explaining the beneficial effect of a large
margin on generalization performance, the current paper investigates the implications of these results for the case of imbalanced
datasets and develops two approaches to setting the threshold.
The approaches are incorporated into ThetaBoost, a boosting algorithm for dealing with unequal loss functions. The performance
of ThetaBoost and the two approaches are tested experimentally.
Keywords: Computational Learning Theory, Generalization, fat-shattering, large
margin, pac estimates, unequal loss, imbalanced datasets
1
Introduction
Shawe-Taylor [8] demonstrated that the output margin can also be used as an
estimate of the confidence with which a particular classification is made. In other
words if a new example has an output value we]) clear of the threshold we can
be more confident of the associated classification than when the output value is
closer to the threshold. The current paper applies this result to the case where
there are different losses associated with a false positive, than with a false negative.
If a significant number of data points are misclassified we can use the criterion
of minimising the empirical loss. If, however, the data is correctly classified the
empirical loss is zero for all correctly separating hyperplanes. It is in this case that
the approach can provide insight into how to choose the hyperplane and threshold.
In summary, the paper suggests ways in which a hyperplane should be optimised for
imbalanced datasets where the loss associated with misclassifying the less prevalent
class is higher.
254
2
G. Karakoulas and J Shawe-Taylor
Background to the Analysis
Definition 2.1 [3} Let F be a set of real-valued functions. We say that a set of
points X is ,-shattered by F if there are real numbers rx indexed by x E X such
that for all binary vectors b indexed by X, there is a function fb E F realising
dichotomy b with margin ,. The fat-shattering dimension Fat.:F of the set F is a
function from the positive real numbers to the integers which maps a value, to the
size of the largest ,-shattered set, if this is finite, or infinity otherwise.
In general we are concerned with classifications obtained by thresholding real-valued
functions. The classification values will be {-I, I} instead of the usual {O, I} in order to simplify some expressions. Hence, typically we will consider a set F of
functions mapping from an input space X to the reals. For the sake of simplifying
the presentation of our results we will assume that the threshold used for classification is O. The results can be extended to other thresholds without difficulty.
Hence we implicitly use the classification functions H = T(F) = {T(J) : f E F},
where T(f) is the function f thresholded at O. We will say that f has , margin on
the training set {(Xi, Yi) : i = 1, ... , m} , if minl<i<m{yd(Xi)} = ,. Note that a
positive margin implies that T(f) is consistent. - -
Definition 2.2 Given a real-valued function f : X --+ [-1, 1] used for classification
by thresholding at 0, and probability distribution P on X x {-I, I}, we use er p (f)
to denote the following probability erp(f) = P{(x, y) : yf(x) ::; O}. Further suppose
o ::; TJ ::; 1, then we use er p(fITJ) to denote the probability
er p (f ITJ) = P { ( x, y) : y f ( x) ::; 0 II f ( x ) I 2: TJ}?
The probability erp(fITJ) is the probability of misclassification of a randomly chosen
example given that it has a margin of TJ or more.
We consider the following restriction on the set of real-valued functions.
Definition 2.3 The real-valued function class F is closed under addition of constants if
TJ E
~,
f
E F :::>
f + TJ
E F.
Note that the linear functions (with threshold weights) used in perceptrons [9]
satisfy this property as do neural networks with linear output units. Hence, this
property applies to the Support Vector Machine, and the neural network examples.
We now quote a result from [8].
Theorem 2.4 [8} Let F be a class of real-valued functions closed under addition
of constants with fat-shattering dimension bounded by Fat.:Fh) which is continuous
from the right. With probability at least 1 - t5 over the choice of a random m sample
(Xi, Yi) drawn according to P the following holds. Suppose that for some f E F,
TJ > 0,
1. yd(xd
2: -TJ + 2,
for all (Xi, yd in the sample,
2. n
= I{i: yi/(xd 2: TJ + 2,}I,
3. n
2: 3J2m(2dln(288m) log2(12em) + In(32m 2jt5)),
=
Let d
Fat.:Fh /6). Then the probability that a new example with margin TJ
misclassified is bounded by
~ (2dlog2(288m)lOg2(12em) + log2
32;n2).
1S
255
Optimizing Classijers for Imbalanced Training Sets
3
Unequal Loss Functions
We consider the situation where the loss associated with an example is different
for misclassification of positive and negative examples. Let Lh(x, y) be the loss
associated with the classification function h on example (x, y). For the analysis
considered above the loss function is taken to be Lh(X, y) = Ih(x) - YI, that is 1 if
the point x is misclassified and 0 otherwise. This is also known as the discrete loss.
In this paper we consider a different loss function for classification functions.
Definition 3.1 The loss function L{3 is defined as L{3(x, y)
h (x) f. y, and 0, otherwise.
=
f3y
+ (1
- y), if
We first consider the classical approach of minimizing the empirical loss, that is the
loss on the training set. Since, the loss function is no longer binary the standard
theoretical results that can be applied are much weaker than for the binary case. The
algorithmic implications will, however, be investigated under the assumption we are
using a hyperplane parallel to the maximal margin hyperplane. The empirical risk
is given by ER(h) = 2:::1 Lf3(Xi, yd, for the training set {(Xi,Yi): i = 1, ... ,m}.
Assuming that the training set can be correctly classified by the hypothesis class
this criterion will not be able to distinguish between consistent hypotheses, hence
giving no reason not to choose the standard maximal margin choice. However,
there is a natural way to introduce the different losses into the maximal margin
quadratic programming procedure [1]. Here, the constraints given are specified as
Yi ((w . Xi) + 0) ~ 1, i ~ 1,2, ... , m. In order to force the hyperplane away from the
positive points which will incur greater loss, a natural heuristic is to set Yi = -1 for
negative examples and Yi = 1/f3 for positive points, hence making them further from
the decision boundary. In the case where consistent classification is possible, the
effect of this will be to move the hyperplane parallel to itself so that the margin on
the positive side is f3 times that on the negative side. Hence, to solve the problem
we simply use the standard maximal margin algorithm [1] and then replace the
threshold 0 with
1
b = 1 + f3[(w, x+)
+ f3(w? x-)),
(1)
where x+ (x-) is one of the closest positive (negative) points.
The alternative approach we wish to employ is to consider other movements of the
hyperplane parallel to itself while retaining consistency. Let
be the margin of the
maximal margin hyperplane. We consider a consistent hyperplane hI] with margin
+ "I to the positive examples, and
"I to the negative example. The basic
analytic tool is Theorem 2.4 which will be applied once for the positive examples
and once for the negative examples (note that classifications are in the set {-I, I}).
,0
,0
,0 -
Theorem 3.2 Let ho be the maximal margin hyperplane with margin ,0, while hI]
is as above with "I < ,0? Set
= (,0 + "I) /2 and ,- = (,0 - "I) /2. With probability
at least 1 - J over the choice of a random m sample (Xi, Yi) drawn according to P
the following holds. Suppose that for ho
,+
1. no
= I{i: YihO(xd
~ 2"1
+ ,o}1,
2. no ~ 3J2m(dln(288m) log2(12em)
+ In(8/J)),
Let d+ = FatF(r+ /6) and d- = FatF(r- /6). Then we can bound the expected loss
by
G. Karakoulas and 1. Shawe- Taylor
256
Proof: Using Theorem 2.4 we can bound the probability of error given that the
correct classification is positive in terms of the expression with the fat shattering
dimension d+ and n = no, while for a negative example we can bound the probability
of error in terms of the expression with fat shattering dimension d- and n = m.
Hence, the expected loss can be bounded by taking the maximum of the second
bound with n+ in place of m together with a factor {3 in front of the second log
term and the first bound multiplied by {3 ??
The bound obtained suggests a way of optimising the choice of "I, namely to minimise
the expression for the fat shattering dimension of linear functions [9]. Solving for "I
in terms of {a and {3 gives
"I
=
{a ((
W - 1) / ( W + 1) ) .
(2)
This choice of "I does not in general agree with that suggested by the choice of
the threshold b in the previous section. In a later section we report on initial
experiments for investigating the performance of these different choices.
4
The ThetaBoost Algorithm
The above idea for adjusting the margin in the case of unequal loss function can
also be applied to the AdaBoost algorithm [2] which has been shown to maximise
the margin on the training examples and hence the generalization can be bounded
in terms of the margin and the fat-shattering dimension of the functions that can
be produced by the algorithm [6]. We will first develop a boosting algorithm for
unequal loss functions and then extend it for adjustable margin. More specifically,
assume: (i) a set of training examples (Xl, yd, ... , (Xrn, Yrn) where Xi E X and
Y E Y = {-I, + I} j (ii) a weak learner that outputs hypotheses h : X -r {-I, + I}
and (iii) the unequal loss function L(3 (y) of Definition 3.1.
=
We assign initial weight Dl (i) = w+ to the n+ positive examples and Dt{ i)
wto the n- negative examples, where w+n+ + w- n- = 1. The values can be set so
that w+ /w- = {3 or they can be adjusted using a validation set. The generalization
of AdaBoost to the case of an unequal loss function is given as the AdaUBoost
algorithm in Figure 1. We adapt theorem 1 in [7] for this algorithm.
Theorem 4.1 Assuming the notation and algorithm of Figure 1, the following
bound holds on the training error of H
T
w+li: H(xd
#- Yi
11 + w-li:
=
H(xd
#- Yi
=
-11::;
IT Zt.
(3)
t=l
The choice of w+ and w- will force uneven probabilities of misclassification on
the training set, but to ensure that the weak learners concentrate on misclassified
positive examples we define Z (suppressing the subscript) as
(4)
i
Thus, to minimize training error we should seek to minimize Z with respect to Q'
(the voting coefficient) on each iteration of boosting. Following [7], we introduce
the notation W++, W_+, W+_ and W __ , where for Sl and S2 E {-I, +1}
D(i)
i :y, =31 ,h(x ,)=32
(5)
257
Optimizing Classifers for Imbalanced Training Sets
By equating to zero the first derivative of (4) with respect to a, Z'(a), and using (5)
we have - exp( -0'/ J3)W++/ ,6+exp(a/ ,6)W_+/,6+exp(a)W+_ -exp( -a)W__ = o.
Letting Y = exp(a) we get a polynomial in Y:
(6)
where C 1 = -W++/,6, C 2 = W_+/,6, C3 = W+_, and C4 = -W__ .
The root of this polynomial can be found numerically. Since Z" (a) > 0, Z' (a) can
have at most one zero and this gives the unique minimum of Z(a). The solution
for a from (6) is used (as at) when taking the distance of a training example from
the standard threshold on each iteration of the AdaUBoost algorithm in Figure 1
as well as when combining the weak learners in H(x).
The ThetaBoost algorithm searches for a positive and a negative support vector
(SV) point such that the hyperplane separating them has the largest margin. Once
these SV points are found we can then apply the formulas (1) and (2) of Sections
3.1 and 3.2 respectively to compute values for adjusting the threshold. See Figure
2 for the complete algorithm.
Algorithm AdaUBoost(X, Y, (3)
1. Initialize Dt{i) as described above.
2. For t
?
?
?
?
?
= 1, ... , T
train weak learner using distribution Dt;
get weak hypothesis h t ;
choose at E lR ;
update: Dt+l(i) = Dt(i) exp[-at(3iYih(xdl/Zt
where (3i = 1/(3 if Yi = 1 and 1 if otherwise, and Zt is a normalization
factor such that Li Dt+1(i) = 1;
3. Output the final hypothesis: H(x) = sgn (L'f=l atht(x)).
Algorithm ThetaBoost(X, Y, (3, 6M
1. H(x)
)
= AdaUBoost(X, Y, ,6);
2. Remove from the training dataset the false positive and borderline points;
3. Find the smallest H(x+) and mark this as the SV+; and remove any negative points with value greater than H(SV+);
4. Find the first negative point that is next in ranking to the SV+ and mark
this as SV_; and compute the margin as the sum of distances, d+ and d_,
of SV+ and SV_ from the standard threshold;
5. Check for candidate SV_ 's that are near to the current one and change the
margin by at least 6M ;
6. Use SV+ and SV_ to compute the theta threshold from Eqn (1) and (2);
7. Output the final hypothesis: H(x)
= sgn (L'f=l atht(x) -
e)
Figure 1: The AdaUBoost and Theta-Boost algorithms.
258
5
G. Karakoulas and J Shawe-Taylor
Experiments
The purpose of the experiments reported in this section is two-fold:
(i) to compare the generalization performance of AdaUBoost against that of
standard Adaboost on imbalanced datasetsj
(ii) to examine the two formulas for choosing the threshold in ThetaBoost and
evaluate their effect on generalization performance.
For the evaluations in (i) and (ii) we use two performance measures: the average
Li3 and the geometric mean of accuracy (g-mean) [4]. The latter is defined as
9 = Jprecision . recall, where
..
preClSlOn =
# positives correct
# posItIves
. . pre d?Icte d j
_
recaII -
# positives correct
# true POSI.tJ.ves .
The g-mean has recently been proposed as a performance measure that, in contrast
to accuracy, can capture the "specificity" trade-off between false positives and true
positives in imbalanced datasets [4]. It is also independent of the distribution of
examples between classes.
For our initial experiments we used the satimage dataset from the UCI repository
[5] and used a uniform D 1 ? The dataset is about classifying neigborhoods of pixels
in a satelite image. It has 36 continuous attributes and 6 classes. We picked class
4 as the goal class since it is the less prevalent one (9.73% of the dataset). The
dataset comes in a training (4435 examples) and a test (2000 examples) set.
Table 1 shows the performance on the test set of AdaUBoost, AdaBoost and C4.5
for different values of the beta parameter. It should be pointed out that the latter
two algorithms minimize the total error assuming an equal loss function (13 = 1). In
the case of equal loss AdaUBoost simply reduces to AdaBoost. As observed from the
table the higher the loss parameter the bigger the improvement of AdaUBoost over
the other two algorithms. This is particularly apparent in the values of g-mean.
f3 values
1
2
4
8
16
AdaUBoost
avgLoss g-mean
0.0545
0.773
0.0895
0.865
0.13
0.889
0.1785
0.898
0.267
0.89
AdaBoost
avgLoss g-mean
0.0545
0.773
0.0831
0.773
0.1662
0.773
0.3324
0.773
0.664
0.773
C4.5
avgLoss g-mean
0.724
0.0885
0.136
0.724
0.231
0.724
0.724
0.421
0.801
0.724
Table 1: Generalization performance in the SatImage dataset.
Figure 2 shows the generalization performance of ThetaBoost in terms of average
loss (13 = 2) for different values of the threshold (). The latter ranges from the largest
margin of negative examples that corresponds to SV_ to the smallest margin of
positive examples that corresponds to SV+. This range includes the values of band
TJ given by formulas (I) and (2). In this experiment,sM was set to 0.2. As depicted in
the figure , the margin defined by b achieves better generalization performance than
the margin defined by TJ. In particular, b is closer to the value of () that gives the
minimum loss on this test set. In addition, ThetaBoost with b performs better than
AdaUBoost on this test set. We should emphasise, however, that the differences
are not significant and that more extensive experiments are required before the two
approaches can be ranked reliably.
Optimizing Classifers for Imbalanced Training Sets
259
0.2.----------.----------.,-----------,
0.18
0.16
en
en
.3
~0.14
l!!Q)
~
0.12
0.1
0.08L--------L--------'----------'
-50
o
50
100
Threshold e
Figure 2: Average Loss L{3 (13 = 2) on test set as a function of ()
6
Discussion
In the above we built a theoretical framework for optimaIly setting the margin
given an unequal loss function. By applying this framework to boosting we developed AdaUBoost and ThetaBoost that generalize Adaboost, a weIl known boosting
algorithm, for taking into account unequal loss functions and adjusting the margin
in imbalanced datasets. Initial experiments have shown that both these factors
improve the generalization performance of the boosted classifier.
References
[lJ Corinna Cortes and Vladimir Vapnik, Machine Learning, 20, 273- 297, 1995.
[2J Yoav Freund and Robert Schapire, pages 148-156 in Proceedings of the International Conference on Machine Learning, ICML '96, 1996.
[3] Michael J. Kearns and Robert E. Schapire, pages 382- 391 in Proceedings of the
31st Symposium on the Foundations of Computer Science, FOCS'90, 1990.
[4J Kubat, M., Holte, R. and Matwin, S., Machine Learning, 30, 195-215, 1998.
[5] Merz, C.J. and Murphy, P.M. (1997). UCI repository of machine learning
databases. http://www.ics.uci.edu/ mlearn/MLRepository.html.
[6] R. Schapire, Y. Freund , P. Bartlett, W. Sun Lee, pages 322- 330 in Proceedings
of International Conference on Machine Learning, ICML '97, 1997.
[7] Robert Schapire and Yoram Singer, in Proceedings of the Eleventh Annual
Conference on Computational Learning Theory, COLT'98, 1998.
[8] John Shawe-Taylor, Algorithmica, 22,157-172,1998.
[9J John Shawe-Taylor, Peter Bartlett, Robert Williamson and Martin Anthony,
IEEE Trans. Inf. Theory, 44 (5) 1926-1940, 1998.
| 1523 |@word repository:2 polynomial:2 seek:1 simplifying:1 initial:4 suppressing:1 current:3 john:3 analytic:1 remove:2 update:1 lr:1 boosting:5 toronto:1 hyperplanes:1 beta:1 symposium:1 focs:1 eleventh:1 introduce:2 li3:1 expected:2 examine:1 bounded:4 notation:2 kubat:1 developed:1 voting:1 xd:5 fat:10 classifier:2 uk:1 unit:1 positive:21 maximise:1 before:1 optimised:1 subscript:1 yd:5 equating:1 suggests:2 analytics:1 range:2 commerce:1 unique:1 borderline:1 procedure:1 tw20:1 empirical:4 confidence:1 word:1 pre:1 specificity:1 get:2 risk:1 applying:1 xrn:1 restriction:1 www:1 map:1 demonstrated:1 insight:1 suppose:3 programming:1 hypothesis:6 particularly:1 database:1 observed:1 capture:1 sun:1 movement:1 trade:1 solving:1 incur:1 learner:4 matwin:1 train:1 london:1 dichotomy:1 choosing:1 apparent:1 heuristic:1 valued:6 solve:1 say:2 otherwise:4 itself:2 final:2 maximal:6 karakoulas:4 uci:3 combining:1 develop:1 ac:1 keywords:1 implies:1 come:1 concentrate:1 correct:3 attribute:1 sgn:2 icte:1 assign:1 generalization:10 d_:1 adjusted:1 hold:3 considered:1 ic:1 exp:6 mapping:1 algorithmic:1 achieves:1 smallest:2 fh:2 purpose:1 wto:1 quote:1 largest:3 tool:1 boosted:1 improvement:1 prevalent:2 check:1 contrast:1 shattered:2 typically:1 lj:1 misclassified:4 pixel:1 classification:12 html:1 colt:1 retaining:1 initialize:1 equal:2 once:3 f3:5 shattering:7 optimising:1 icml:2 report:1 develops:1 simplify:1 employ:1 randomly:1 bce:1 murphy:1 algorithmica:1 lf3:1 evaluation:1 tj:12 implication:2 closer:2 lh:2 indexed:2 taylor:7 oex:1 theoretical:2 yoav:1 uniform:1 front:1 reported:1 sv:8 confident:1 st:2 international:2 lee:1 off:1 j2m:2 michael:1 together:1 itj:1 posi:1 choose:3 derivative:1 li:3 account:1 w_:3 includes:1 coefficient:1 satisfy:1 ranking:1 fatshattering:1 later:1 root:1 picked:1 closed:2 parallel:3 minimize:3 accuracy:2 fitj:2 generalize:1 weak:5 produced:1 rx:1 classified:2 mlearn:1 fatf:2 email:2 definition:5 against:1 associated:5 proof:1 dataset:6 adjusting:3 recall:1 higher:2 dt:6 adaboost:7 eqn:1 yf:1 yrn:1 effect:3 true:2 hence:8 ll:1 mlrepository:1 criterion:2 complete:1 performs:1 image:1 recently:1 s8:1 extend:1 classifers:2 numerically:1 significant:2 consistency:1 pointed:1 shawe:7 longer:1 closest:1 imbalanced:10 recent:1 optimizing:4 inf:1 binary:3 yi:12 minimum:2 greater:2 holte:1 minl:1 ii:4 reduces:1 england:1 adapt:1 minimising:1 bigger:1 j3:1 basic:1 iteration:2 normalization:1 background:1 addition:3 xdl:1 integer:1 near:1 canadian:1 iii:1 concerned:1 idea:1 minimise:1 expression:4 rhbnc:1 bartlett:2 f:1 peter:1 clear:1 band:1 schapire:4 http:1 sl:1 misclassifying:1 dln:2 correctly:3 discrete:1 group:1 threshold:16 imperial:1 drawn:2 erp:2 thresholded:1 sum:1 place:1 decision:1 investigates:1 bound:7 hi:2 distinguish:1 fold:1 quadratic:1 annual:1 infinity:1 constraint:1 sake:1 martin:1 department:1 according:2 beneficial:1 em:3 making:1 taken:1 agree:1 singer:1 letting:1 multiplied:1 apply:1 away:1 egham:1 alternative:1 corinna:1 ho:2 ensure:1 log2:4 yoram:1 giving:1 classical:1 move:1 usual:1 distance:2 separating:2 reason:1 assuming:3 minimizing:1 vladimir:1 robert:4 negative:13 reliably:1 zt:3 adjustable:1 datasets:5 sm:1 finite:1 situation:1 extended:1 incorporated:1 canada:1 namely:1 required:1 specified:1 c3:1 extensive:1 c4:3 unequal:9 boost:1 trans:1 able:1 suggested:1 built:1 royal:1 misclassification:3 difficulty:1 natural:2 force:2 ranked:1 improve:1 theta:2 geometric:1 freund:2 loss:32 validation:1 foundation:1 consistent:4 thresholding:2 bank:1 classifying:1 summary:1 side:2 weaker:1 explaining:1 taking:3 emphasise:1 boundary:1 dimension:7 fb:1 t5:1 made:1 implicitly:1 dlog2:1 dealing:1 global:1 investigating:1 xi:9 continuous:2 search:1 bay:1 table:3 ca:1 williamson:1 investigated:1 anthony:1 s2:1 n2:1 atht:2 realising:1 en:2 wish:1 xl:1 candidate:1 theorem:6 formula:3 showing:1 pac:1 er:4 cortes:1 dl:1 ih:1 false:4 vapnik:1 importance:1 margin:31 depicted:1 simply:2 applies:2 corresponds:2 goal:1 presentation:1 satimage:2 replace:1 experimentally:1 change:1 specifically:1 hyperplane:11 kearns:1 total:1 merz:1 perceptrons:1 holloway:1 uneven:1 support:2 mark:2 latter:3 evaluate:1 tested:1 |
574 | 1,525 | Batch and On-line Parameter Estimation of
Gaussian Mixtures Based on the Joint Entropy
Yoram Singer
AT&T Labs
singer@research.att.com
Manfred K. Warmuth
University of California, Santa Cruz
manfred@cse.ucsc.edu
Abstract
We describe a new iterative method for parameter estimation of Gaussian mixtures. The new method is based on a framework developed by
Kivinen and Warmuth for supervised on-line learning. In contrast to gradient descent and EM, which estimate the mixture's covariance matrices,
the proposed method estimates the inverses of the covariance matrices.
Furthennore, the new parameter estimation procedure can be applied in
both on-line and batch settings. We show experimentally that it is typically faster than EM, and usually requires about half as many iterations
as EM.
1 Introduction
Mixture models, in particular mixtures of Gaussians, have been a popular tool for density
estimation, clustering, and un-supervised learning with a wide range of applications (see
for instance [5, 2] and the references therein). Mixture models are one of the most useful
tools for handling incomplete data, in particular hidden variables. For Gaussian mixtures
the hidden variables indicate for each data point the index of the Gaussian that generated it.
Thus, the model is specified by ajoint density between the observed and hidden variables.
The common technique used for estimating the parameters of a stochastic source with hidden variables is the EM algorithm. In this paper we describe a new technique for estimating
the parameters of Gaussian mixtures. The new parameter estimation method is based on a
framework developed by Kivinen and Warmuth [8] for supervised on-line learning. This
framework was successfully used in a large number of supervised and un-supervised problems (see for instance [7, 6, 9, 1]).
Our goal is to find a local minimum of a loss function which, in our case, is the negative
log likelihood induced by a mixture of Gaussians. However, rather than minimizing the
Parameter Estimation of Gaussian Mixtures
579
loss directly we add a tenn measuring the distance of the new parameters to the old ones.
This distance is useful for iterative parameter estimation procedures. Its purpose is to keep
the new parameters close to the old ones. The method for deriving iterative parameter
estimation can be used in batch settings as well as on-line settings where the parameters
are updated after each observation. The distance used for deriving the parameter estimation
method in this paper is the relative entropy between the old and new joint density of the
observed and hidden variables. For brevity we tenn the new iterative parameter estimation
method the joint-entropy (JE) update.
The JE update shares a common characteristic with the Expectation Maximization [4, 10]
algorithm as it first calculates the same expectations. However, it replaces the maximization
step with a different update of the parameters. For instance, it updates the inverse of the
covariance matrix of each Gaussian in the mixture, rather than the covariance matrices
themselves. We found in our experiments that the JE update often requires half as many
iterations as EM. It is also straightforward to modify the proposed parameter estimation
method for on-line setting where the parameters are updated after each new observation.
As we demonstrate in our experiments with digit recognition, the on-line version of the
JE update is especially useful in situations where the observations are generated by a nonstationary stochastic source.
2 Notation and preliminaries
Let S be a sequence of training examples (Xl, X2, ..? , XN) where each Xi is a ddimensional vector in ~d. To model the distribution of the examples we use m ddimensional Gaussians. The parameters of the i-th Gaussian are denoted by 8 i and they
include the mean-vector and the covariance matrix
The density function of the ith Gaussian, denoted P(xI8d, is
We denote the entire set of parameters of a Gaussian mixture by 8 = {8i }:: 1 =
{Wi, Pi' C i }::l where w = (WI, ... , w m ) is a non-negative vector of mixture coefficients
such that 2:::1 Wi = 1. We denote by P(xI8) = 2:;:1 w;P(xI8d the likelih~od of an
observation x according to a Gaussian mixture with parameters_8. Let 8 i and 8 i be two
Gaussian distributions. For brevity, we d~note by E; (Z) and E j (Z) the expectation of a
random variable Z with respect to 8i and 8 j ? Let f be a parametric function whose parameters constitute a matrix A = (a;j). We denote by {) f / {)A the matrix of partial derivatives
of f with respect to the elements in A. That is, the ij element of {) f / {)A is {) f / {)aij.
Similarly, let B = (bij(x)) a matrix whose elements are functions of a scalar x. Then, we
denote by dB / dx the matrix of derivatives of the elements in B with respect to x, namely,
the ij element of dB / dx is dbij (x) / dx.
3 The framework for deriving updates
Kivinen and Warmuth [8] introduced a general framework for deriving on-line parameter
updates. In this section we describe how to apply their framework for the problem of
Y. Singer and M. K. Warmuth
580
parameter estimation of Gaussian mixtures in a batch setting. We later discuss how a
simple modification gives the on-line updates.
Given a set of data points S in ~d and a number m, the goal is to find a set of m
Gaussians that minimize the loss on the data, denoted as loss(SI8). For density estimation the natural loss function is the negative log-likelihood of the data loss(SI8) =
-(I/ISI) In P(SI8) ~f -(I/ISI) L: xES In P(xI8). The best parameters which minimize
the above loss cannot be found analytically. The common approach is to use iterative methods such as EM [4, 10] to find a local minimizer of the loss.
In an iterative parameter estimation framework we are given the old set of parameters 8 t
and we need to find a set of new parameters 8 t +1 that induce smaller loss. The framework
introduced by Kivinen and Warmuth [8] deviates from the common approaches as it also
requires to the new parameter vector to stay "close" to the old set of parameters which
incorporates all that was learned in the previous iterations. The distance of the new parameter setting 8 t +1 from the old setting 8 t is measured by a non-negative distance function
Ll(8 t +1 , 8 t ). We now search for a new set of parameters 8 t +1 that minimizes the distance
summed with the loss multiplied by 17. Here 17 is a non-negative number measuring the relative importance of the distance versus the loss. This parameter 17 will become the learning
rate of the update. More formally, the update is found by setting 8 t+1 = arg mineUt(8)
whereU t (8) = Ll(8,8 t ) + 17loss(SI8) + A(L:::1 Wi -1). (We use a Lagrange multiplier A to enforce the constraint that the mixture coefficients sum to one.) By choosing the
apropriate distance function and 17 = 1 one can show that EM becomes the above update.
For most distance functions and learning rates the minimizer of the function U t (8) cannot be found analytically as both the distance function and the log-likelihood are usually
non-linear in 8. Instead, we expand the log-likelihood using a first order Taylor expansion around the old parameter setting. This approximation degrades the further the new
parameter values are from the old ones, which further motivates the use of the distance
function Ll(8, 8 t ) (see also the discussion in [7]). We now seek a new set of parameters
8 t + 1 = argmineVt(8) where
m
Vt(8) =
~(8, 0 t ) + '7 (loss(510 t ) + (8 - 0 t ) . V' e l0ss(510t)) + A(L w.
- 1) .
(1)
.=1
Here V' eloss(SI8t) denotes the gradient of the loss at 8 t . We use the above method
Eq. (1) to derive the updates of this paper. For density estimation, it is natural to use the
relative entropy between the new and old density as a distance. In this paper we use the
joint density between the observed (data points) and hidden variables (the indices of the
Gaussians). This motivates the name joint-entropy update.
4
Entropy based distance functions
We first consider the relative entropy between the new and old parameter parameters of a
single Gaussian. Using the notation introduced in Sec. 2, the relative entropy between two
Gaussian distributions denoted by 8i , 8i is
~(8., 8i)
def
=
[
P(xI0.)
JXE~d P(xI0i) In P(xI8.) dx
zI}n
le.1
le.1
-z- -
I
IE-((
-)Te--))
z
' X -I-'i
? (X -I-'i
1-(( X
+ zEi
-
I
I-'i) Te? ( x-I-'. ))
Parameter Estimation of Gaussian Mixtures
581
Using standard (though tedious) algebra we can rewrite the expectations as follows:
A(8-- i, 8)
-i
U
ICil = 2"1] n -;;:;ICil
-d
2
1 (C-1C-)
+ 2"tr
i
i + 2"1(J.li -
J.l )T C i-1(J.li - J.li ) .
(2)
The relative entropy between the new and the old mixture models is the following
-
~(0,0)
f
def
-
f
P(xI8)
= ix P(xI0) In P(xI0)dx = ix
~- L~1 w.P(xI8.)
7::
w.P(xI0.)ln ~:1 w.P(xI0./ x .
(3)
Ideally, we would like to use the above distance function in V t to give us an update of
in terms of 8. However, there isn't a closed form expression for Eq. (3). Although the
relative entropy between two Gaussians is a convex function in their parameters, the relative
entropy between two Gaussian mixtures is non-convex. Thus, the loss function V t (e) may
have multiple minima, making the problem of finding arg mine V t (e) difficult.
e
In order to sidestep this problem we use the log-sum inequality [3] to obtain an upper bound
for the distance function ~(e, 8). We denote this upper bound as Li(e, 8).
=L:
m
-
W,
In ;;;,
w,
_=1
+
L:
m
w,
j -
,
p(xle ,) In p(xle
I ) dx =
P(x e,l
_=1
x
L:
m
-
W,
In -;;;,
w,
+
,=1
L:
m
-
-
WI~(e"
e,l .
(4)
1=1
We call the new distance function Li(e, 8) the joint-entropy distance. Note that in this
distance the parameters of Wi and Wi are "coupled" in the sense that it is a convex combination of the distances 6.(8 i , 8d. In particular, Li(8 , 8) as a function of the parameters
Wi, Pi' Ci does not remain constant any more when the parameters of the individual Gaussians are permuted. Furthermore, Li (e, 8) is also is sufficiently convex so that finding the
minimizer of V t is possible (see below).
5 The updates
We are now ready to derive the new parameter estimation scheme. This is done by setting
the partial derivatives of V t , with respect to
to O. That is, our problem consists of solving
the following equations
e,
a~(e, e)
_
1)
-
151
aw,
a In p(5Ie)
+>- =
aw,
0,
a~(e, e)
_
-
1)
151
aJ.L,
a In P(5Ie)
=
aJ.L ,
0,
a~(e , e)
ac,
1)
aln p(5Ie)
151
ac,
-
= o.
We now use the fact that Ci and thus C;l is symmetric. The derivatives of Li(e, 8), as
defined by Eq. (4) and Eq. (2), with respect to Wi, Pi and C\, are
In -W.
w.
ICd
+ 1 + -21 In -ICd
aE(0,0)
ac.
1 ( -1 - )
1()TC-1 ()
+ -tr
2 C Ci + -2"",,,,,,,,,,,"", (5)
ll . -
II
.
II
-"
t
(6)
alii
aE(0,0)
- d
-2
__
1 - (C-- 1 C-1)
2 Wi i
+ .
.
(7)
Y Singer and M. K. Warmuth
582
To simplify the notation throughout the rest of the paper we define the following variables
f3.(x)
d ef
=
P(xI0i)
( d ef wi P (xI0i)
P(x\0) and (X i x) = P(x\0)
= P ('1t x, 0i) = wif3i(X ) .
The partial derivatives of the log-likelihood are computed similarly:
oln P(SI0)
=
OWi
OC.
=
~a ( )
L.; P' x
X? s
(8)
~ w.P(xI0.) -1 (
L.; P(xI0) C i X-I-I.)
oln P(SI0)
01-1.
oIn P(SI0)
~ P(xI0i)
L.; P(xI0)
X?S
=
x?s
=
(9)
~s
_l ~ wiP(xI0.) (C:- 1
2
~
-1 (
L.;(X.(x)C. X-I-Ii)
L.;
P(xI0)
?
_
C:- 1 (
?
x
_
.)(
1-1.
x
_
1-1.
)TC:-1)
?
x?s
-t L(X,(x)(Ci
1 -
C i 1 (x-l-li)(X-I-I.f c ;-t).
(10)
x?s
We now need to decide on an order for updating the parameter classes Wi, Pi ' and C i . We
use the same order that EM uses, namely, Wi, then Pi' and finally, C i . (After doing one
pass over all three groups we start again using the same order.) Using this order results in
a simplified set of equations as several terms in Eq. (5) cancel out. Denote the size of the
sample by N = lSI. We now need to sum the derivatives from Eq. (5) and Eq. (8) while
using the fact that the Lagrange multiplier). simply assures that the new weight Wi sum to
one. By setting the result to zero, we get that
w. t-
E:l
WJ
exp (-N Ex? s f3i(X?)
(11)
Similarly, we sum Eq. (6) and Eq. (9), set the result to zero, and get that
I-li t-I-I.
+ ~ Lf3i(X) (x -I-Ii)'
(12)
x?s
Finally, we do the same for C i . We sum Eq. (7) and Eq. (10) using the newly obtained Pi'
Cit t- Ci 1 + ~ Lf3.(x) (Cit - C;-l(X -I-I.)(x -l-lifCi1) .
x?s
(13)
We call the new iterative parameter estimation procedure the joint-entropy (JE) update.
To summarize, the JE update is composed of the following alternating steps: We first calculate for each observation x the value !3i(X) = P(xI8;}j P(xI8) and then update the
parameters as given by Eq. (11), Eq. (12), and Eq. (13). The JE update and EM differ in
several aspects. First, EM uses a simple update for the mixture we!ghts w . Second, EM
uses the expectations (with respect to the current parameters) of the sufficient statistics [4]
for Pi and C; to find new sets of mean vectors and covariance matrices. The JE uses a
(slightly different) weighted average of the observation and, in addition, it adds the old
parameters. The learning rate TJ determines the proportion to be used in summing the old
parameters and the newly estimated parameters. Last, EM estimates the covariance matrices Ci whereas the new update estimates the inverses, C;l, of these matrices. Thus, it is
potentially be more stable numerically in cases where the covariance matrices have small
condition number.
To obtain an on-line procedure we need to update the parameters after each new observation
at a time. That is, rather than summing over all xES, for a new observation Xt, we update
Parameter Estimation of Gaussian Mixtures
-3.0
I
II
I
JE ot8_1 .9
-3.1
J
. "!/
/018=1.5
.:
/
ot8:1 .1../ / ota=l .OS
~~
I
~ -32
--::::_________
'i
/ /.
_~" EM
__-0':;:;--<
S
r'"
! -0170
- 0 171
~
l-//' ./'
~-3.3
583
EM
,
,
r
EU
,,
._...-
~----7---!----:-,--7---:----!---'
/( ,/
-3.4
-<),
lo,
rr"
.....
BJ
...............
.9'"
-0'
o
50
100
150
200
Number 01 iterations
250
300
10
15
~
._-
~
~
................. ..........
E'"
~
~
~
~
Figure 1: Left: comparison of the convergence rate of EM and the JE update with different
learning rates. Right: example of a case where EM initially increases the likelihood faster
than the JE update.
the parameters and get a new set of parameters 8 t +1 using the current parameters 8 t ? The
new parameters are then used for inducing the likelihood of the next observation Xt+ 1. The
on-line parameter estimation procedure is composed of the following steps:
Xj e,
1. Set: (3i (Xt ) = PP(Xj
e) .
2. Parameter updates:
(a) Wj
f- Wj
exp (-1]t(3j (xt)) /
+
(b) J,lj
f- J,lj
1]t
1
(c) Ci f- Ci 1
I:j=1 Wj exp ( -1]t(3j (xt))
(3j (xt) (Xt - J,lj)
+ 1]t (3j(xt) (Cil - Ci 1 (Xt - J,lj)(Xt - J,lj)TCi1).
To guarantee convergence of the on-line update one should use a diminishing learning rate,
that is 1]t -t 0 as t -t 00 (for further motivation see [lID.
6 Experiments
We conducted numerous experiments with the new update. Due to the lack of space we describe here only two. In the first experiment we compared the JE update and EM in batch
settings. We generated data from Gaussian mixture distributions with varying number of
components (m
2 to 100) and dimensions (d
2 to 20). Due to the lack of space
we describe here results obtained from only one setting. In this setting the examples were
generated by a mixture of 5 components with w = (0.4 , 0.3,0.2,0.05,0.05). The mean
vectors were the 5 standard unit vectors in the Euclidean space 1R5 and we set all of covariances matrices to the identity matrix. We generated 1000 examples. We then run EM and
the JE update with different learning rates (1]
1.9,1.5,1.1,1.05). To make sure that all
the runs will end in the same local maximum we fist performed three EM iterations. The
results are shown on the left hand side of Figure 1. In this setting, the JE update with high
learning rates achieves much faster convergence than EM. We would like to note that this
behavior is by no means esoteric - most of our experiments data yielded similar results.
=
=
=
We found a different behavior in low dimensional settings. On the right hand side of Figure 1 we show convergence rate results for a mixture containing two components each of
which is a single dimension Gaussians. The mean of the two components were located
Y. Singer and M. K. Warmuth
584
at 1 and -1 with the same variance of 2. Thus, there is a significant "overlap" between
the two Gaussian constituting the mixture. The mixture weight vector was (0 .5,0 .5). We
generated 50 examples according to this distribution and initialized the parameters as fol0.01,1-'2
-0.01, 0"1
0"2
2, WI
W2
0.5 We see that initially
lows: 1-'1
EM increases the likelihood much faster than the JE update. Eventually, the JE update
convergences faster than EM when using a small learning rate (in the example appearing in
Figure 1 we set 'rJ = 1.05). However, in this setting, the JE update diverges when learning
rates larger than 'rJ
1.1 are used. This behavior underscores the advantages of both methods. EM uses a fixed learning rate and is guaranteed to converge to a local maximum of the
likelihood, under conditions that typically hold for mixture of Gaussians [4, 12]. the JE update, on the other hand, encompasses a learning rate and in many settings it converges much
faster than EM. However, the superior performance in high dimensional cases demands its
price in low dimensional "dense" cases. Namely, a very conservative learning rate, which
is hard to tune, need to be used. In these cases, EM is a better alternative, offering almost
the same convergence rate without the need to tune any parameters.
=
=
=
=
=
=
=
Acknowledgments Thanks to Duncan Herring for careful proof reading and providing
us with interesting data sets.
References
[1] E. Bauer, D. Koller, and Y. Singer. Update rules for parameter estimation in Bayesian
networks. In Proc. of the 13th Annual Con! on Uncertainty in AI, pages 3-13, 1997.
[2] C.M. Bishop. Neural Networks and Pattern Recognition. Oxford Univ. Press, 1995.
[3] Thomas M. Cover and Joy A Thomas. Elements of Information Theory. Wiley, 1991.
[4] AP. Dempster, N.M. Laird, and D.B. Rubin. Maximum-likelihood from incomplete
data via the EM algorithm. Journal of the Royal Statistical Society, B39:1-38, 1977.
[5] R.O. Duda and P.E. Hart. Pattern Classification and Scene Analysis. Wiley, 1973.
[6] D. P. Helmbold, J. Kivinen, and M.K. Warmuth. Worst-case loss bounds for sigmoided neurons. In Advances in Neural Information Processing Systems 7, pages
309-315, 1995.
[7] D.P. Helmbold, R.E. Schapire, Y.Singer, and M.K. Warmuth. A comparison of new
and old algorithms for a mixture estimation problem. Machine Learning, Vol. 7, 1997.
[8] J. Kivinen and M.K. Warmuth. Additive versus exponentiated gradient updates for
linear prediction. Information and Computation, 132(1): 1-64, January 1997.
[9] J. Kivinen and M.K. Warmuth. Relative loss bounds for multidimensional regression
problems. In Advances in Neural Information Processing Systems 10, 1997.
[10] R.A Redner and H.E Walker. Mixture densities, maximum likelihood and the EM
algorithm. SIAM Review, 26(2), 1984.
[11] D.M. Titterington, A.EM. Smith, and U.E. Makov. Statistical Analysis of Finite Mixture Distributions. Wiley, 1985.
[12] C.E Wu. On the convergence properties of the EM algorithm. Annals of Stat., 11 :95103, 1983.
| 1525 |@word version:1 proportion:1 duda:1 tedious:1 si8:4 seek:1 covariance:9 b39:1 tr:2 tci1:1 att:1 offering:1 current:2 com:1 od:1 dx:6 herring:1 cruz:1 additive:1 ota:1 update:38 joy:1 half:2 tenn:2 warmuth:12 ith:1 smith:1 manfred:2 cse:1 ucsc:1 become:1 consists:1 behavior:3 themselves:1 isi:2 becomes:1 estimating:2 notation:3 aln:1 minimizes:1 developed:2 titterington:1 finding:2 guarantee:1 multidimensional:1 unit:1 local:4 modify:1 oxford:1 ap:1 therein:1 range:1 acknowledgment:1 digit:1 procedure:5 induce:1 get:3 cannot:2 close:2 straightforward:1 convex:4 helmbold:2 rule:1 deriving:4 updated:2 annals:1 us:5 element:6 recognition:2 updating:1 dbij:1 located:1 observed:3 worst:1 calculate:1 wj:4 eu:1 oln:2 dempster:1 ideally:1 mine:1 rewrite:1 solving:1 algebra:1 joint:7 univ:1 describe:5 choosing:1 whose:2 larger:1 furthennore:1 statistic:1 laird:1 sequence:1 rr:1 advantage:1 xi8:7 ajoint:1 inducing:1 convergence:7 diverges:1 converges:1 derive:2 ac:3 stat:1 measured:1 ij:2 eq:14 ddimensional:2 indicate:1 differ:1 stochastic:2 preliminary:1 hold:1 around:1 sufficiently:1 exp:3 bj:1 achieves:1 purpose:1 estimation:22 proc:1 si0:3 successfully:1 tool:2 weighted:1 gaussian:20 rather:3 varying:1 likelihood:11 underscore:1 contrast:1 sense:1 typically:2 entire:1 lj:5 initially:2 hidden:6 diminishing:1 koller:1 expand:1 arg:2 classification:1 denoted:4 summed:1 f3:1 r5:1 cancel:1 simplify:1 composed:2 individual:1 lf3:1 mixture:30 tj:1 partial:3 incomplete:2 old:14 taylor:1 euclidean:1 initialized:1 instance:3 cover:1 measuring:2 maximization:2 conducted:1 aw:2 thanks:1 density:9 siam:1 ie:4 stay:1 again:1 containing:1 derivative:6 sidestep:1 li:10 makov:1 sec:1 coefficient:2 later:1 performed:1 lab:1 closed:1 doing:1 start:1 minimize:2 variance:1 characteristic:1 bayesian:1 pp:1 proof:1 con:1 newly:2 popular:1 redner:1 supervised:5 done:1 though:1 furthermore:1 hand:3 o:1 lack:2 aj:2 name:1 multiplier:2 analytically:2 alternating:1 symmetric:1 ll:4 oc:1 demonstrate:1 ef:2 common:4 superior:1 permuted:1 sigmoided:1 xi0:10 numerically:1 significant:1 ai:1 similarly:3 stable:1 add:2 inequality:1 vt:1 minimum:2 converge:1 ii:5 fist:1 multiple:1 rj:2 faster:6 hart:1 calculates:1 prediction:1 regression:1 ae:2 expectation:5 iteration:5 addition:1 whereas:1 walker:1 source:2 w2:1 rest:1 sure:1 induced:1 db:2 incorporates:1 call:2 nonstationary:1 xj:2 zi:1 f3i:1 icd:2 expression:1 constitute:1 useful:3 santa:1 tune:2 cit:2 schapire:1 lsi:1 estimated:1 vol:1 group:1 sum:6 run:2 inverse:3 uncertainty:1 throughout:1 almost:1 decide:1 wu:1 duncan:1 def:2 bound:4 guaranteed:1 replaces:1 yielded:1 annual:1 constraint:1 x2:1 scene:1 aspect:1 according:2 combination:1 smaller:1 remain:1 em:29 slightly:1 wi:15 lid:1 modification:1 making:1 handling:1 ln:1 equation:2 assures:1 discus:1 eventually:1 singer:7 end:1 gaussians:9 multiplied:1 apply:1 enforce:1 appearing:1 batch:5 alternative:1 thomas:2 denotes:1 clustering:1 include:1 yoram:1 especially:1 society:1 parametric:1 degrades:1 gradient:3 distance:19 index:2 providing:1 minimizing:1 difficult:1 potentially:1 negative:5 motivates:2 upper:2 observation:9 neuron:1 finite:1 descent:1 january:1 situation:1 introduced:3 namely:3 specified:1 xle:2 california:1 learned:1 usually:2 below:1 pattern:2 reading:1 summarize:1 encompasses:1 royal:1 overlap:1 natural:2 kivinen:7 scheme:1 numerous:1 ready:1 icil:2 coupled:1 isn:1 alii:1 deviate:1 review:1 relative:9 loss:17 zei:1 interesting:1 versus:2 sufficient:1 rubin:1 share:1 pi:7 lo:1 last:1 aij:1 side:2 exponentiated:1 wide:1 bauer:1 dimension:2 xn:1 simplified:1 constituting:1 keep:1 summing:2 xi:1 un:2 iterative:7 search:1 expansion:1 oin:1 dense:1 owi:1 motivation:1 je:18 cil:1 wiley:3 xl:1 ix:2 bij:1 xt:10 bishop:1 wip:1 x:2 importance:1 ci:9 te:2 demand:1 entropy:13 tc:2 simply:1 lagrange:2 scalar:1 minimizer:3 determines:1 goal:2 identity:1 careful:1 price:1 experimentally:1 hard:1 conservative:1 pas:1 formally:1 brevity:2 ex:1 |
575 | 1,526 | Recurrent Cortical Amplification Produces
Complex Cell Responses
Frances S. Chance~ Sacha B. Nelson~ and L. F. Abbott
Volen Center and Department of Biology
Brandeis University
Waltham, MA 02454
Abstract
Cortical amplification has been proposed as a mechanism for enhancing
the selectivity of neurons in the primary visual cortex. Less appreciated
is the fact that the same form of amplification can also be used to de-tune
or broaden selectivity. Using a network model with recurrent cortical
circuitry, we propose that the spatial phase invariance of complex cell
responses arises through recurrent amplification of feedforward input.
Neurons in the network respond like simple cells at low gain and complex ceUs at high gain. Similar recurrent mechanisms may playa role
in generating invariant representations of feedforward input elsewhere in
the visual processing pathway.
1 INTRODUCTION
Synaptic input to neurons in the primary visual cortex is primarily recurrent, arising from
other cortical cells. The dominance of this type of connection suggests that it may play an
important role in cortical information processing. Previous studies proposed that recurrent
connections amplify weak feedforward input to the cortex (Douglas et aI., 1995) and selectively amplify tuning for specific stimulus characteristics, such as orientation or direction
of movement (Douglas et aI., 1995; Ben-Yishai et aI., 1995; Somers et aI., 1995; Sompolinsky and Shapley, 1997). Cortical cooling and shocking experiments provide evidence that
there is cortical amplification through recurrent connections, but they do not show increases
in orientation or direction selectivity as a result of this amplification (Ferster et aI., 19%;
Chung and Ferster, 1998). Recurrent connections can also decrease neuronal selectivity
through the same form of amplification, generating responses that are insensitive to certain
stimulus features. Although the ability to sharpen tuning may be an important feature in
cortical processing, the capacity to broaden tuning for particular stimulus attributes is also
desirable.
Neurons in the primary visual cortex can be divided into two classes based on their re-
Recurrent Cortical Amplification Produces Complex Cell Responses
91
sponses to visual stimuli such as counterphase and drifting sinusoidal gratings. Simple cells show tuning for orientation, spatial frequency, and spatial phase of a grating
(Movshon et aI., 1978a). Complex cells exhibit orientation and spatial frequency tuning, but are insensitive to spatial phase (Movshon et aI., 1978b). A counterphase grating,
s(x, t) = cos(Kx - cp) cos(wt), is one in which the spatial phase, CP,and spatial frequency,
K, are held constant but the contrast, s(x, t), varies sinusoidally in time at some frequency
w. In response to a counterphase grating, the activity of a simple cell oscillates at the same
frequency as the stimulus, w. A complex cell response is modulated at twice the frequency,
2w. To create a drifting grating of frequency 1/, s( x, t) = cos( K x - I/t), the spatial phase
and spatial frequency are held constant but the grating is moved at velocity 1// K. A simple
cell response to a drifting grating is highly modulated at frequency 1/, while a complex
cell response to a drifting grating is elevated but relatively unmodulated. The differences
between complex and simple cell responses are a direct consequence of the complex cell
spatial phase insensitivity.
Previous models of complex cells generate spatial-phase invariant responses through converging sets of feedforward inputs with a wide range of spatial phase preferences but similar
orientation and spatial frequency selectivities (Hubel and Wiesel, 1962; Mel et aI., 1998).
These models do not incorporate recurrent connections between complex cells, which are
known to be particularly strong (Toyama et al., 1981). We propose that the spatial phase
invariance of complex cell responses can arise from a broadening of spatial phase tuning
by cortical amplification (Chance et aI., 1998). The model neurons exhibit simple cell behavior when weakly coupled and complex cell behavior when strongly coupled, suggesting
that the two classes of neurons in the primary visual cortex may arise from the same basic
cortical circuit.
2
THE MODEL
The activity of neuron i in the model network is characterized by a firing rate rio Each
neuron sums feedforward and recurrent input and responds as described by the standard
rate-model equation
dri
Trdi
~
= Ii + L.J Wijrj - rio
Ii represents the feedforward input to cell i, W ij is the weight of the synapse from neuron
j to neuron i, and Tr is a time constant. Previous studies have suggested that, for a neuron
receiving many inputs, Tr is small, closer to a synaptic time constant than the membrane
time constant (Ben-Yishai et al., 1995; Treves, 1993). Thus we choose Tr = 1 ms.
The feedforward input describes the response of a simple cell with a Gabor receptive field
Ii = [/ dxGi(x)
1
00
dt' H(t')s(x, t - t')] + '
where s(x, t) represents the contrast function of the visual stimulus and the notation [ 1+
indicates rectification. The temporal response function is (Adelson and Bergen, 1985)
(at')5
(at')7)
H(t') = exp(-at') ( - - - - 5!
where we use a
7!'
= l/ms. The spatial filter is a Gabor function,
G= exp ( - 2:; ) cos(kix -
<Pi)'
where Ui determines the spatial extent of the receptive field, ki is the preferred spatial
frequency, and <Pi is the preferred spatial phase. The values of <Pi are equally distributed
F S. Chance, S. B. Nelson and L. F Abbott
92
over the interval [-180 0 , 180 0 ) . To give the neurons a realistic bandwidth, (j i is chosen
such that ki(ji = 2.5 . Initially we consider a simplified case in which k i = 1 for all cells.
Later we consider the spatial frequency selectivity of neurons in the network and allow the
value of k i to range from 0 to 3.5 cycles/deg.
In this paper we assume that the model network describes one orientation column of the
primary visual cortex, and thus all neurons have the same orientation tuning. All stimuli
are of the optimal orientation for the network.
Spatial phase tuning is selectively broadened in the model because the strength of a recurrent connection between two neurons is independent of the spatial phase selectivities of
their feedforward inputs. In the model with all k i = 1, the recurrent input is determined by
9
Wij
= (N -1) '
for all i =I j. N is the number of cells in the network, and 0 ~ 9 < gmax, where gmax is
the largest value of 9 for which the network remains stable. In this case gmax = 1.
3 RESULTS
The steady-state solution of the rate-model equation is given by Ti = Ii + L WijTj . To
solve this equation, we express the rates and feedfoward inputs in terms of a complete set
of eigenvectors ~r of the recurrent weight matrix, L Wij~r = ~/.l~r for I-L = 1,2, ... , N,
where ~/.l are the eigenvalues. The solution is then
This equation displays the phenomenon of cortical amplification if one or more of the
eigenvalues is near one. If we assume only one eigenvalue, ~1 , is close to one, the factor 1~1 in the denominator causes the I-L = 1 term to dominate and we find Ti ~ ~t L Ij~J (1 ~1) -1. The input combination L Ij~J dominates the response, determining selectivity,
and this mode is amplified by a factor 1/(1 - ~1)' We refer to this amplification factor as
the cortical gain.
In the case where W ij = g/(N - 1) for i =I j, the largest eigenvalue is ~1 = 9 and the
corresponding eigenvector has all components equal to each other. For 9 near one, the recurrent input to neuron i is then proportional to Lj [cos(~-1>j)]+ which, for large numbers
of cells with uniformly placed preferred spatial phases 1>i, is approximately independent of
~, the spatial phase of the stimulus. When 9 is near zero, the network is at low gain and
the response of neuron i is roughly proportional to its feedforward input, [cos(~ - 1>j)]+,
and is sensitive to spatial phase.
The response properties of simple and complex cells to drifting and counterphase gratings
are duplicated by the model neuron, as shown in figure 1. For low gain (gain = 1, top panels
of figures lA and IB), the neuron acts as a simple cell and its activity is modulated at the
same frequency as the stimulus (w for counterphase gratings and v for drifting gratings).
At high gain (gain = 20), the neuron responds like a complex cell, exhibiting frequency
doubling in the response to a counterphase grating (bottom panel of Figure 1A) and an
elevated DC response to a drifting grating (bottom panel, Figure IB). Intermediate gain
(gain =5) produces intermediate behavior (middle panels).
The basis of this model is that the amplified mode is independent of spatial phase. If
the amplified mode depends on spatial frequency or orientation, neurons at high gain can
be selective for these attributes. To show that the model can retain selectivity for other
93
Recurrent Cortical Amplification Produces Complex Cell Responses
500
1000
500
1000
500
1000
1~~1VY)!\<
o
500
1000
time (ms)
time (ms)
Figure 1: The effects of recurrent input on the responses of a neuron in the model network.
The responses of one neuron to a 2 Hz counterphase grating (A) and to a 2 Hz drifting
grating (B) are shown for different levels of network gain. From top to bottom in A and B,
the gain of the network is one, five, and twenty.
stimulus characteristics while maintaining spatial phase insensitivity, we allowed the spatial
frequency selectivity which each neuron receives from its feedforward input, ki' to vary
from neuron to neuron and also modified the recurrent weight matrix so that the strength
of the connection between two neurons, i and j, depends on k i - k j . The dependence is
modeled as a difference of Gaussians, so the recurrent weight matrix is now
9
[
((k i - kj
W ij = (N -1) 2exp 20'~
)2) -
((k i - kj
exp 20';
)2)] .
Thus neurons that receive feedforward input tuned for similar spatial frequencies excite
each other and neurons that receive very differently tuned feedforward input inhibit each
other. This produces complex cells that are tuned to a variety of spatial frequencies, but are
still insensitive to spatial phase (see figure 2). The spatial frequency tuning curve width is
primarily determined by 0' c = 0.5 cycle/deg and 0' s = 1 cycle/deg.
Cells within the same network do not have to exhibit the same level of gain. In previous figures, the gain of the network was determined by a parameter 9 that described the
strength of all the connections between neurons. In figure 3, the recurrent input to cell i
is determined by W ij = gi/(N - 1), where the values of gi are chosen randomly within
the allowed range. The gain of each neuron depends on the value of gi for that neuron. As
shown in figure 3, a range of complex and simple cell behaviors now coexist within the
same network.
4 DISCUSSION
In the recurrent model we have presented, as in Hubel and Wiesel's feedforward model, the
feedforward input to a complex cell arises from simple cells. Measurements by Alonso and
F S. Chance, S. B. Nelson and L. F Abbott
94
A c:~
100
~
50
o
a.
(/)
xctS
E
#.
O-+-~r----r--~----'
-180
-90
o
90
phase (deg)
180
o
1
2
3
spatial frequency (cyc/deg)
Figure 2: Neurons in a high-gain network can be selective for spatial frequency while remaining insensitive to spatial phase. Both spatial phase and spatial frequency tuning are
included in the feedforward input. A) The spatial phase tuning curves of three representative neurons from a high-gain network. B) The spatial frequency tuning curves of the same
three neurons as in A.
Martinez (1998) support this circuitry. However, direct excitatory input to complex cells
arising from the LGN has also been reported (Hoffman and Stone, 1971; Singer et aI., 1975;
Ferster and Lindstrom, 1983). Supporting these measurements is evidence that certain
stimuli can excite complex cells without strong excitation of simple cells (Hammond and
Mackay, 1975, 1977; Movshon, 1975) and also that complex cells still respond when simple
cells are silenced (Malpeli, 1983; Malpeli et ai, 1986; Mignard and Malpeli, 1991). In
accordance with this, the weak feedforward simple cell input in the recurrent model could
probably be replaced by direct LGN input, as in the feedforward model of Mel et al. (1998).
The proposed model makes definite predictions about complex cell responses. If the phaseinvariance of complex cell responses is due to recurrent interactions, manipulations that
modify the balance between feedforward and recurrent drive should change the nature of
the responses in a predictable manner. The model predicts that blocking local excitatory
connections should turn complex cells into simple cells. Conversely, manipulations that
increase cortical gain should make simple cells act more like complex cells. One way to
increase cortical gain may be to block or partially block inhibition since this increases the
influence of excitatory recurrent connections. Experiments along these lines have been
performed, and blockade of inhibition does indeed cause simple cells to take on complex
cell properties (Sillito, 1975; Shulz et aI., 1993).
In a previous study, Hawken, Shapley, and Grosof (1996) noted that the temporal frequency
tuning curves for complex cells are narrower for counterphase stimuli than for drifting
stimuli. The recurrent model reproduces this result as long as the integration of synaptic
inputs depends on temporal frequency. Such a dependence is provided, for example, by
short-term synaptic depression (Chance et aI., 1998). Hubel and Wiesel's feedforward
model (1962) does not reproduce this effect, even with synaptic depression at the synapses.
We have presented a model of primary visual cortex in which complex cell response characteristics arise from recurrent amplification of simple cell responses. The complex cell
responses in the high gain regime arise because recurrent connections selectively deamplify selectivity for spatial phase. Thus recurrent connections can act to generate invariant
representation of input data. A similar mechanism could be used to produce responses that
are independent of other stimulus attributes, such as size or orientation. Given the ubiquity
of invariant representations in the visual pathway, this mechanism may have widespread
use.
95
Recurrent Cortical Amplification Produces Complex Cell Responses
1~~~
o
~1001JV\M
eft. 50
o -+-------,.--------.
o
500
time (ms)
1000
500
1000
1~~~
o
500
1000
time (ms)
Figure 3: Responses to a 4 Hz drifting grating of four neurons from a large network consisting of a mixture of simple and complex cells. The two traces on the left represent simple
cells and the two traces on the right represent complex cells.
Acknowledgements
Research supported by the Sloan Center for Theoretical Neurobiology at Brandeis University, the National Science Foundation (DMS-95-03261), the W.M. Keck Foundation, the
National Eye Institute (EY-11116), and the Alfred P. Sloan Foundation.
References
Adelson, E. H. & Bergen, J. R. Spatiotemporal energy models for the perception of motion.
1. Opt. Soc. Am. A. 2,284-299 (1985)
Alonso, J-M. & Martinez, L. M. Functional connectivity between simple cells and complex
cells in cat striate cortex. Nature Neuroscience 1,395-403 (1998)
Ben-Yishai, R., Bar-Or, L. & Sompolinsky, H. Theory of orientation tuning in visual cortex.
Proc. Natl. Acad. Sci. USA 92,3844-3848 (1995)
Chance F. S., Nelson S. B. & Abbott L. F. Complex cells as cortically amplified simple
cells. (submitted)
Chung, S. & Ferster, D. Strength and orientation tuning of the thalamic input to simple
cells revealed by electrically evoked cortical suppression. Neuron 20,1177-1189 (1998)
Douglas, R. J., Koch, c., Mahowald, M., Martin , K. A. C. & Suarez, H. H. Recurrent
excitation in neocortical circuits. Science 269 981-985 (1995)
Ferster, D., Chung, S. & Wheat, H. Orientation selectivity ofthalamic input to simple cells
of cat visual cortex. Nature 380,249-252 (1996)
Ferster, D. & Lindstrom, S. An intracellular analysis of geniculo-cortical connectivity in
area 17 of the cat. 1. Physiol . (Lond) 342,181-215 (1983)
Hammond, P. & MacKay, D. M. Differential responses of cat visual cortical cells to textured stimuli. Exp. Brain Res. 22,427-430 (1975)
96
F S. Chance, S. B. Nelson and L. F Abbott
Hammond, P. & MacKay, D. M. Differential responsiveness of simple and complex cells
in cat striate cortex to visual texture. Exp. Brain Res. 30, 275-296 (1977)
Hawken, M. J., Shapley, R. M. & Grosof, D. H. Temporal-frequency selectivity in monkey
visual cortex. Vis. Neurosci. 13477-492 (1996)
Hoffman, K. P. & Stone, J. Conduction velocity of afferents of cat visual cortex: a correlation with cortical receptive field properties. Brain Res. 32,460-466 (1971)
Hubel, D. H. & Wiesel, T. N. Receptive fields, binocular interaction and functional architecture in the cat's visual cortex. 1. Physiol.160, 106-154 (1962)
Malpeli, J. G. Activity of cells in area 17 of the cat in absence of input from layer A of
lateral geniculate nucleus. 1. Neurophysiol. 49,595-610 (1983)
Malpeli, J. G., Lee, C., Schwark, H. D. & Weyand, T. G. Cat area 17.1. Pattern of thalamic
control of cortical layers. 1. Neurophysiol. 56,1062-1073 (1986)
Mel, B. W., Ruderman, D. L. & Archie, K. A. Translation-invariant orientation tuning in
visual complex cells could derive from intradendritic computations. 1. Neurosci. 1843254334 (1998)
Mignard, M. & Malpeli, J. G. Paths of information flow through visual cortex. Science 251,
1249-1251 (1991)
Movshon, J. A. The velocity tuning of single units in cat striate cortex. 1. Physiol. 249,
445-468 (1975)
Movshon, J., Thompson, I. & Tolhurst, D. Spatial summation in the receptive fields of
simple cells in the cat's striate cortex. 1. Physiol. (Lond) 283, 53-77 (1978)
Movshon, J., Thompson, I. & Tolhurst, D. Receptive field organization of complex cells in
cat's striate cortex. 1. Physiol. (Lond) 283,79-99 (1978)
Shulz, D. E., Bringuier, B. & Fregnac, Y. A complex-like structure of simple visual cortical
receptive fields is masked by GABA-A intracortical inhibition. Soc .jor Neurosci . Abs. 19,
628 (1993)
Sillito, A. M. The contribution of inhibitory mechanisms to the receptive field properties of
neurones in the striate cortex of the cat. 1. Physiol. (Lond) 250,305-329 (1975)
Singer, W., Tretter, F. & Cynader, M. Organization of cat striate cortex: a correlation of
receptive-field properties with afferent and efferent connections. 1. Neurophysiol. 38, 10801098 (1975)
Somers, D. C., Nelson, S. B. & Sur, M. An emergent model of orientation selectivity in cat
visual cortical simple cells. 1. Neurosci. 15,5448-5465 (1995)
Sompolinsky, H. & Shapley, R. New perspectives on the mechanisms for orientation selectivity. Current Opinion in Neurobiology 7, 514-522 (1997)
Toyama, K., Kimura, M. & Tanaka, K. Organization of cat visual cortex as investigated by
cross-correlation technique. 1. Neurophysiol. 46,202-214 (1981)
Treves, A. Mean-field analysis of neuronal spike dynamics. Network 4, 259-284 (1993)
| 1526 |@word middle:1 wiesel:4 tr:3 tuned:3 current:1 physiol:6 realistic:1 feedfoward:1 short:1 tolhurst:2 preference:1 kix:1 five:1 along:1 direct:3 differential:2 pathway:2 shapley:4 manner:1 indeed:1 roughly:1 behavior:4 brain:3 provided:1 notation:1 circuit:2 panel:4 eigenvector:1 monkey:1 kimura:1 temporal:4 toyama:2 ti:2 act:3 oscillates:1 control:1 unit:1 broadened:1 unmodulated:1 accordance:1 modify:1 local:1 consequence:1 acad:1 firing:1 path:1 approximately:1 twice:1 evoked:1 suggests:1 conversely:1 co:6 range:4 block:2 definite:1 area:3 gabor:2 amplify:2 close:1 coexist:1 influence:1 center:2 thompson:2 dominate:1 weyand:1 play:1 velocity:3 particularly:1 cooling:1 predicts:1 blocking:1 bottom:3 role:2 suarez:1 wheat:1 cycle:3 sompolinsky:3 movement:1 decrease:1 inhibit:1 predictable:1 ui:1 dynamic:1 weakly:1 cynader:1 basis:1 textured:1 neurophysiol:4 differently:1 emergent:1 cat:16 solve:1 ability:1 gi:3 eigenvalue:4 propose:2 interaction:2 malpeli:6 insensitivity:2 amplified:4 amplification:14 moved:1 keck:1 produce:7 generating:2 ben:3 derive:1 recurrent:31 ij:6 grating:16 soc:2 strong:2 waltham:1 exhibiting:1 direction:2 attribute:3 filter:1 opinion:1 opt:1 summation:1 koch:1 exp:6 circuitry:2 vary:1 proc:1 geniculate:1 sensitive:1 largest:2 create:1 hoffman:2 modified:1 indicates:1 contrast:2 suppression:1 am:1 rio:2 bergen:2 lj:1 initially:1 wij:2 france:1 selective:2 lgn:2 reproduce:1 orientation:16 spatial:42 integration:1 mackay:3 field:10 equal:1 biology:1 represents:2 adelson:2 stimulus:15 primarily:2 shulz:2 randomly:1 national:2 replaced:1 phase:24 consisting:1 ab:1 organization:3 highly:1 mixture:1 natl:1 yishai:3 held:2 closer:1 gmax:3 re:4 theoretical:1 column:1 sinusoidally:1 mahowald:1 masked:1 sacha:1 archie:1 reported:1 conduction:1 varies:1 spatiotemporal:1 retain:1 lee:1 receiving:1 fregnac:1 connectivity:2 choose:1 chung:3 suggesting:1 sinusoidal:1 de:1 intracortical:1 sloan:2 afferent:2 depends:4 vi:1 later:1 performed:1 thalamic:2 contribution:1 characteristic:3 weak:2 intradendritic:1 hammond:3 drive:1 submitted:1 synapsis:1 synaptic:5 energy:1 frequency:26 dm:1 efferent:1 gain:21 duplicated:1 dt:1 response:31 synapse:1 strongly:1 binocular:1 correlation:3 receives:1 ruderman:1 widespread:1 mode:3 usa:1 effect:2 blockade:1 width:1 noted:1 mel:3 steady:1 excitation:2 m:6 stone:2 complete:1 neocortical:1 cp:2 motion:1 functional:2 ji:1 insensitive:4 elevated:2 eft:1 refer:1 measurement:2 ai:13 tuning:17 sharpen:1 stable:1 cortex:21 inhibition:3 playa:1 perspective:1 manipulation:2 selectivity:15 certain:2 responsiveness:1 ey:1 ii:4 desirable:1 characterized:1 cross:1 long:1 divided:1 equally:1 converging:1 prediction:1 basic:1 denominator:1 enhancing:1 represent:2 cell:68 receive:2 interval:1 probably:1 hz:3 dri:1 flow:1 near:3 feedforward:19 intermediate:2 revealed:1 variety:1 geniculo:1 architecture:1 bandwidth:1 movshon:6 cyc:1 neurones:1 cause:2 depression:2 eigenvectors:1 tune:1 generate:2 vy:1 inhibitory:1 neuroscience:1 arising:2 alfred:1 express:1 dominance:1 four:1 jv:1 douglas:3 abbott:5 sum:1 respond:2 somers:2 hawken:2 ki:3 layer:2 display:1 activity:4 strength:4 lond:4 relatively:1 martin:1 department:1 combination:1 electrically:1 gaba:1 membrane:1 describes:2 invariant:5 rectification:1 equation:4 remains:1 turn:1 mechanism:6 singer:2 gaussians:1 ubiquity:1 drifting:10 broaden:2 top:2 remaining:1 maintaining:1 spike:1 receptive:9 primary:6 dependence:2 striate:7 responds:2 exhibit:3 sci:1 capacity:1 lateral:1 alonso:2 nelson:6 extent:1 sur:1 modeled:1 balance:1 trace:2 twenty:1 neuron:37 supporting:1 neurobiology:2 volen:1 dc:1 treves:2 connection:13 tanaka:1 suggested:1 bar:1 perception:1 pattern:1 regime:1 eye:1 coupled:2 kj:2 acknowledgement:1 determining:1 proportional:2 foundation:3 nucleus:1 pi:3 translation:1 elsewhere:1 excitatory:3 placed:1 supported:1 appreciated:1 allow:1 institute:1 wide:1 distributed:1 curve:4 cortical:24 jor:1 simplified:1 brandeis:2 preferred:3 deg:5 reproduces:1 hubel:4 excite:2 sillito:2 nature:3 broadening:1 investigated:1 complex:39 intracellular:1 neurosci:4 arise:4 martinez:2 allowed:2 neuronal:2 silenced:1 representative:1 cortically:1 ib:2 lindstrom:2 specific:1 evidence:2 dominates:1 sponses:1 texture:1 kx:1 visual:22 partially:1 doubling:1 chance:7 determines:1 ma:1 narrower:1 ferster:6 absence:1 change:1 included:1 determined:4 uniformly:1 wt:1 invariance:2 la:1 selectively:3 support:1 arises:2 modulated:3 incorporate:1 phenomenon:1 |
576 | 1,527 | Facial Memory is Kernel Density Estimation
(Almost)
Matthew N. Dailey
Garrison W. Cottrell
Department of Computer Science and Engineering
U.C. San Diego
La Jolla, CA 92093-0114
{mdailey,gary}@cs.ucsd.edu
Thomas A. Busey
Department of Psychology
Indiana University
Bloomington, IN 47405
busey@indiana.edu
Abstract
We compare the ability of three exemplar-based memory models, each
using three different face stimulus representations, to account for the
probability a human subject responded "old" in an old/new facial memory experiment. The models are 1) the Generalized Context Model, 2)
SimSample, a probabilistic sampling model, and 3) MMOM, a novel
model related to kernel density estimation that explicitly encodes stimulus distinctiveness. The representations are 1) positions of stimuli in
MDS "face space," 2) projections of test faces onto the "eigenfaces" of
the study set, and 3) a representation based on response to a grid of Gabor
filter jets. Of the 9 model/representation combinations, only the distinctiveness model in MDS space predicts the observed "morph familiarity
inversion" effect, in which the subjects' false alarm rate for morphs between similar faces is higher than their hit rate for many of the studied
faces. This evidence is consistent with the hypothesis that human memory for faces is a kernel density estimation task, with the caveat that distinctive faces require larger kernels than do typical faces.
1 Background
Studying the errors subjects make during face recognition memory tasks aids our understanding of the mechanisms and representations underlying memory, face processing, and
visual perception. One way of evoking such errors is by testing subjects' recognition of
new faces created from studied faces that have been combined in some way (e.g. Solso and
McCarthy, 1981; Reinitz, Lammers, and Cochran 1992). Busey and Tunnicliff (submitted) have recently examined the extent to which image-quality morphs between unfamiliar
faces affect subjects' tendency to make recognition errors.
Their experiments used facial images of bald males and morphs between these images (see
? \i<"
Facial Memory Is Kernel Density Estimation (Almost)
25
..
... "
<: . .
\ ,';'.>,"i'.. >
"
I'
'.'
",.,,",
.,. :- . ?... '
.'~
,.;;.'
'
..
.
.'...
.
.
Il"
..
r
::;;1;f~
Figure 1: Three normalized morphs from the database.
Figure 1) as stimuli. In one study, Busey (in press) had subjects rate the similarity of all
pairs in a large set of faces and morphs, then performed a multidimensional scaling (MDS)
of these similarity ratings to derive a 6~dimensional "face space" (Valentine and Endo,
1992). In another study, "Experiment 3" (Busey and Tunnicliff, submitted), 179 subjects
studied 68 facial images, including 8 similar pairs and 8 dissimilar pairs, as determined in a
pilot study. These pairs were included in order to study how morphs between similar faces
and dissimilar faces evoke false alanns. We call the pair of images from which a morph are
derived its "parents," and the morph itself as their "child." In the experiment's test phase,
the subjects were asked to make new/old judgments in response to 8 of the 16 morphs, 20
completely new distractor faces, the 36 non-parent targets and one of the parents of each of
the 8 morphs. The results were that, for many of the morphlparent pairs, subjects responded
"old" to the unstudied morph more often than to its studied parent. However, this effect (a
morphfamiliarity inversion) only occurred for the morphs with similar parents. It seems
that the similar parents are so similar to their "child" morphs that they both contribute
toward an "old" (false alann) response to the morpho
Researchers have proposed many models to account for data from explicit memory experiments. Although we have applied other types of models to Busey and Tunnicliff's
data with largely negative results (Dailey et al., 1998), in this paper, we limit discussion
to exemplar-based models, such as the Generalized Context Model (Nosofsky, 1986) and
SAM (Gillund and Shiffrin, 1984). These models rely on the assumption that subjects
explicitly store representations of each of the stimuli they study. Busey and Tunnicliff applied several exemplar-based models to the Experiment 3 data, but none of these models
have been able to fully account for the observed similar morph familiarity inversion without positing that the similar parents are explicitly blended in memory, producing prototypes
near the morphs.
We extend Busey and Tunnicliff's (submitted) work by applying two of their exemplar
models to additional image-based face stimulus representations, and we propose a novel
exemplar model that accounts for the similar morphs' familiarity inversion. The results are
consistent with the hypothesis that facial memory is a kernel density estimation (Bishop,
1995) task, except that distinctive exemplars require larger kernels. Also, on the basis of
our model, we can predict that distinctiveness with respect to the study set is the critical
factor influencing kernel size, as opposed to a context-free notion of distinctiveness. We
can easily test this prediction empirically.
2 Experimental Methods
2.1 Face Stimuli and Normalization
The original images were 104 digitized 560x662 grayscale images of bald men, with consistent lighting and background and fairly consistent position. The subjects varied in race
and extent of facial hair. We automatically located the left and right eyes on each face using
a simple template correlation technique then translated, rotated, scaled and cropped each
image so the eyes were aligned in each image. We then scaled each image to 114x 143 to
speed up image processing. Figure 1 shows three examples of the normalized morphs (the
original images are copyrighted and cannot be published) .
26
M N. Dailey, G. W Cottrell and T. A. Busey
2.2 Representations
Positions in multidimensional face space Many researchers have used a multidimensional scaling approach to model various phenomena in face processing (e.g. Valentine and
Endo, 1992). Busey (in press) had 343 subjects rate the similarity of pairs of faces in the
test set and performed a multidimensional scaling on the similarity matrix for 100 of the
faces (four non-parent target faces were dropped from this analysis). The process resulted
in a 6-dimensional solution with r2 = 0.785 and a stress of 0.13. In the MDS modeling
results described below, we used the 6-dimensional vector associated with each stimulus as
its representation.
Principal component projections "Eigenfaces," or the eigenvectors of the covariance
matrix for a set of face images, are a common basis for face representations (e.g. Turk and
Pentland, 1991). We performed a principal components analysis on the 68 face images used
in the study set for Busey and Tunnicliff's experiment to get the 67 non-zero eigenvectors
of their covariance matrix. We then projected each of the 104 test set images onto the 30
most significant eigenfaces to obtain a 30-dimensional vector representing each face. l
Gabor filter responses von der Malsburg and colleagues have made effective use of
banks of Gabor filters at various orientations and spatial frequencies in face recognition systems. We used one form of their wavelet (Buhmann, Lades, and von der Malsburg, 1990) at
five scales and 8 orientations in an 8x8 square grid over each normalized face image as the
basis for a third face stimulus representation. However, since this representation resulted
in a 2560-dimensional vector for each face stimulus, we performed a principal components
analysis to reduce the dimensionality to 30, keeping this representation's dimensionality the
same as the eigenface representation's. Thus we obtained a 30-dimensional vector based
on Gabor filter responses to represent each test set face image.
2.3 Models
The Generalized Context Model (GCM) There are several different flavors to the GCM.
We only consider a simple sum-similarity form that will lead directly to our distinctivenessmodulated density estimation model. Our version of GCM's predicted P(old), given a
representation y of a test stimulus and representations x E X of the studied exemplars, is
predy
= a + {3 L
e- c (d x ?y )2
xEX
where a and {3linearly convert the probe's summed similarity to a probability, X is the set
of representations of the study set stimuli; c is used to widen or narrow the width of the
similarity function, and dx,y is either Ilx - yll, the Euclidean distance between x and y
or the weighted Euclidean distance VLk Wk(Xk - Yk)2 where the "attentional weights"
Wk are constants that sum to 1. Intuitively, this model simply places a Gaussian-shaped
function over each of the studied exemplars, and the predicted familiarity of a test probe is
simply the summed height of each of these surfaces at the probe's location.
Recall that two of our representations, PC projection space and Gabor filter space, are
30-dimensional, whereas the other, MDS, is only 6-dimensional. Thus allowing adaptive
weights for the MDS representation is reasonable, since the resulting model only uses 8
parameters to fit 100 points, but it is clearly unreasonable to allow adaptive weights in
PC and Gabor space, where the resulting models would be fitting 32 parameters to 100
points. Thus, for all models, we report results in MDS space both with and without adaptive
weights, but do not report adaptive weight results for models in PC and Gabor space.
SimSample Busey and Tunnicliff (submitted) proposed SimSample in an attempt to remedy the GCM's poor predictions of the human data. It is related to both GCM, in that it
1We used 30 eigenfaces because with this number, our theoretical "distinctiveness" measure was
best correlated with the same measure in MDS space.
27
Facial Memory Is Kernel Density Estimation (Almost)
uses representations in MDS space, and SAM (Gillund and Shiffrin, 1984), in that it involves sampling exemplars. The idea behind the model is that when a subject is shown
a test stimulus, instead of a summed comparison to all of the exemplars in memory, the
test probe probabilistically samples a single exemplar in memory, and the subject responds
"old" if the probe's similarity to the exemplar is above a noisy criterion. The model has
a similarity scaling parameter and two parameters describing the noisy threshold function.
Due to space limitations, we cannot provide the details of the model here.
Busey and Tunnicliff were able to fit the human data within the SimS ample framework,
but only when they introduced prototypes at the locations of the morphs in MDS space and
made the probability of sampling the prototype proportional to the similarity of the parents.
Here, however, we only compare with the basic version that does not blend exemplars.
Mixture Model of Memory (MMOM) In this model, we assume that subjects, at study
time, implicitly create a probability density surface corresponding to the training set. The
subjects' probability of responding "old" to a probe are then taken to be proportional to the
height of this surface at the point corresponding to the probe. The surface must be robust
in the face of the variability or noise typically encountered in face recognition (lighting
changes, perspective changes, etc.) yet also provide some level of discrimination support
(i.e. even when the intervals of possible representations for a single face could overlap
due to noise, some rational decision boundary must still be constructed). If we assume
a Gaussian mixture model, in which the density surface is built from Gaussian "blobs"
centered on each studied exemplar, the task is a form of kernel density estimation (Bishop,
1995).
We can fonnulate the task of predicting the human subjects' P(old) in this framework, then,
as optimizing the priors and widths of the kernel functions to minimize the mean squared
error of the prediction. However, we also want to minimize the number of free parameters
in the model - parsimonious methods for setting the priors and kernel function widths
potentially lead to more useful insights into the principles underlying the human data. If
the priors and widths were held constant, we would have a simple two parameter model
predicting the probability a subject responds "old" to a test stimulus y:
predy =
L
oe-
I!x_~1!2
2 ..
xEX
where a folds together the uniform prior and normalization constants, and (7 is the standard deviation of the Gaussian kernels. If we ignore the constants, however, this model
is essentially the same as the version of the GCM described above. As the results section
will show, this model cannot fully account for the human familiarity data in any of our
representational spaces.
To improve the model, we introduce two parameters to allow the prior (kernel function
height) and standard deviation (kernel function width) to vary with the distinctiveness of the
studied exemplar. This modification has two intuitive motivations. First, when humans are
asked which of two parent faces a 50% morph is most similar to, if one parent is distinctive
and the other parent is typical, subjects tend to choose the more distinctive parent (Tanaka et
aI., submitted). Second, we hypothesize that when a human is asked to study and remember
a set of faces for a recognition test, faces with few neighbors will likely have more relaxed
(wider) discrimination boundaries than faces with many nearby neighbors.
Thus in each representation space, for each studied face x, we computed d(x), the theoretical distinctiveness of each face, as the Z-scored average distance to the five nearest studied
faces. We then allowed the height and width of each kernel function to vary with d(x):
predy =
L
xEX
0(1
+ cod(x?e
_
I!x_yl!2
2("(l+c .. d(x?2
As was the case for GCM and SimSample, we report the results of using a weighted Euclidean distance between y and x in MDS space only.
28
M. N Dailey. G. W. Cottrell and T. A. Busey
Model
" MDS space
GCM
0.1633
SimS ample
0.1521
MMOM
0.1601
I MDS + weights I PC projections I Gabor jets I
0.1417
0.1404
0.1528
0.1745
0.1756
0.1992
0.1624
0.1704
0.1668
Table 1: RMSE for the three models and three representations. Quality of fit for models
with adaptive attentional weights are only reported for the low-dimensional representation
("MDS + weights"). The baseline RMSE, achievable with a constant prediction, is 0.2044.
2.4 Parameter fitting and model evaluation
For each of the twelve combinations of models with face representations, we searched
parameter space by simple hill climbing for the parameter settings that minimized the mean
squared error between the model's predicted P(old) and the actual human P(old) data.
We rate each model's effectiveness with two criteria. First, we measure the models' global
fit with RMSE over all test set points. A model's RMSE can be compared to the baseline
performance of the "dumbest" model, which simply predicts the mean human P(old) of
0.5395, and achieves an RMSE of 0.2044. Second, we evaluate the extent to which a model
predicts the mean human response for each of the six categories of test set stimuli: 1) nonparent targets, 2) non-morph distractors, 3) similar parents, 4) dissimilar parents, 5) similar
morphs, and 6) dissimilar morphs. If a model correctly predicts the rank ordering of these
category means, it obviously accounts for the similar morph familiarity inversion pattern in
the human data. As long as models do an adequate job of fitting the human data overall, as
measured by RMSE, we prefer models that predict the morph familiarity inversion effect
as a natural consequence of minimizing RMSE.
3 Results
Table 1 shows the global fit of each model/representation pair. The SimSample model in
MDS space provides the best quantitative fit. GeM generally outperforms MMOM, indicating that for a tight quantitative fit, having parameters for a linear transformation built
into the model is more important than allowing the kernel function to vary with distinctiveness. Also of note is that the PC projection representation is consistently outperformed by
both the Gabor jet representation and the MDS space representation.
But for our purposes, the degree to which a model predicts the mean human responses for
each of the six categories of stimuli is more important, given that it is doing a reasonably
good job globally. Figure 2 takes a more detailed look at how well each model predicts
the human category means. Even though SimSample in MDS space has the best global
fit to the human familiarity ratings, it does not predict the familiarity inversion for similar
morphs. Only the mixture model in weighted MDS space correctly predicts the morph
familiarity effect. All of the other models underpredict the human responses to the similar
morphs.
4 Discussion
The results for the mixture model are consistent with the hypothesis that facial memory is
a kernel density estimation task, with the caveat that distinctive exemplars require larger
kernels. Whereas true density estimation would tend to deemphasize outliers in sparse
areas of the face space, the human data show that the priors and kernel function widths for
outliers should actually be increased. Two potentially significant problems with the work
presented here are first, we experimented with several models before finding that MMOM
was able to predict the morph familiarity inversion effect, and second, we are fitting a single
Facial Memory Is Kernel Density Estimation (Almost)
GCMlMDS
SimSamplelMDS
0.6
i
~
f
~
i
0."
iI:"
0.2
f
~
T
SP OM
OA
iE"
0.4
0.2
f
0.2
0
0 .4
iE"
0.2
f
~
T
SP OM
0."
0 .2
f
0 .2
~
T
SP OM
OP SM
0
SimSampleIPC
0 .6
0."
iil:"0A
iill::
0."
0.2
f
~
f
0.2
SP OM
~
0.2
GCWGabor
T
SP
OM
OP SM
0
0.6
0.6
0."
iill:: 0."
iit'"
0."
0.2
~
~
t
0.2
DP SM
T
SP OM
0
D
T
SP
OM
0
0.6
0.2
0.0
0.0
0 .0
OM
MMOWGabor
SimSample/Gabor
f
SP
0.0
DP SM
0
T
0.6
0.0
T
0
MMOMIPC
0 .6
OP SM
OM
0.0
OP SM
0,0
SP
MMOMlMDS+wts
iill::
0
T
0.6
0."
GCMlPC
r
DP SM
0
0.0
OP SM
iE"
SP OM
0.6
0.0
r
0.0
T
SimSamplelMDS+wts
0.6
iit'"
~
OP SM
GCMlMDS+wts
~
0.6
0 .0
OP SM
f
MMOMlMDS
0.6
0 .0
iE"
29
OP SM
T
SP OM
0
OP SM
T
SP
OM
0
Actual
_Predicted
r::::=:I
Figure 2: Average actual/predicted responses to the faces in each category. Key: DP =
Dissimilar parents; SM Similar morphs; T Non-parent targets; SP Similar parents;
DM Dissimilar morphs; D = Distractors.
=
=
=
=
experiment. The model thus must be carefully tested against new data, and its predictions
empirically validated.
Since a theoretical distinctiveness measure based on the sparseness of face space around an
exemplar was sufficient to account for the similar morphs' familiarity inversion, we predict
that distinctiveness with respect to the study set is the critical factor influencing kernel size,
rather than context-free human distinctiveness judgments. We can easily test this prediction
by having subjects rate the distinctiveness of the stimuli without prior exposure and then
determine whether their distinctiveness ratings improve or degrade the model's fit.
A somewhat disappointing (though not particularly surprising) aspect of our results is that
the model requires a representation based on human similarity judgments. Ideally, we
would prefer to provide an information-processing account using image-based representations like eigenface projections or Gabor filter responses. Interestingly, the efficacy of the
image-based representations seems to depend on how similar they are to the MDS representations. The PC projection representation performed the worst, and distances between
pairs of PC representations had a correlation of 0.388 with the distances between pairs of
MDS representations. For the Gabor filter representation, which performed better, the correlation is 0.517. In future work, we plan to investigate how the MDS representation (or a
representation like it) might be derived directly from the face images.
30
M N. Dailey, G. W Cottrell and T A. Busey
Besides providing an infonnation-processing account of the human data, there are several
other avenues for future research. These include empirical testing of our distinctiveness
predictions, evaluating the applicability of the distinctiveness model in domains other than
face processing, and evaluating the ability of other modeling paradigms to account for this
data.
Acknowledgements
We thank Chris Vogt for comments on a previous draft, and other members of Gary's
Unbelievable Research Unit (GURU) for earlier comments on this work. This research was
supported in part by NIMH grant MH57075 to GWe.
References
Bishop, C. M. (1995). Neural networks for pattern recognition. Oxford University Press,
Oxford.
Busey, T. A. (1999). Where are morphed faces in multi-dimensional face space? Psychological Science. In press.
Busey, T. A. and Tunnicliff, J. (submitted). Accounts of blending, distinctiveness and
typicality in face recognition. Journal of Experimental Psychology: Learning, Memory,
and Cognition.
Dailey, M. N., Cottrell, G. W., and Busey, T. A. (1998). Eigenfaces for familiarity. In
Proceedings of the Twentieth Annual Conference of the Cognitive Science Society, pages
273-278, Mahwah, NJ. Erlbaum.
Gillund, G. and Shiffrin, R. (1984). A retrieval model for both recognition and recall.
Psychological Review, 93(4):411-428.
J. Buhmann, M. L. and von der Malsburg, C. (1990). Size and distortion invariant object
recognition by hierarchical graph matching. In Proceedings of the IJCNN International
Joint Conference on Neural Networks, volume II, pages 411-416.
Nosofsky, R. M. (1986). Attention, similarity, and the identification-categorization relationship. Journal of Experimental Psychology: General, 116(1):39-57.
Reinitz, M., Lammers, W., and Cochran, B. (1992). Memory-conjunction errors: Miscombination of stored stimulus features can produce illusions of memory. Memory &
Cognition, 20(1):1-11.
Solso, R. L. and McCarthy, J. E. (1981). Prototype formation offaces: A case of pseudomemory. British Journal of Psychology, 72(4):499-503.
Tanaka, J., Giles, M., Kremen, 5., and Simon, V. (submitted). Mapping attract or fields in
face space: The atypicality bias in face recognition.
Turk, M. and Pentland, A. (1991). Eigenfaces for recognition. The Journal of Cognitive
Neuroscience, 3:71-86.
Valentine, T. and Endo, M. (1992). Towards an exemplar model of face processing: The
effects of race and distinctiveness. The Quarterly Journal of Experimental Psychology,
44A(4):671-703.
| 1527 |@word version:3 inversion:9 achievable:1 seems:2 vogt:1 covariance:2 efficacy:1 interestingly:1 outperforms:1 surprising:1 yet:1 dx:1 must:3 cottrell:5 hypothesize:1 xex:3 discrimination:2 xk:1 caveat:2 provides:1 draft:1 contribute:1 location:2 gillund:3 positing:1 five:2 height:4 constructed:1 fitting:4 introduce:1 distractor:1 multi:1 globally:1 automatically:1 actual:3 underlying:2 finding:1 indiana:2 transformation:1 nj:1 remember:1 quantitative:2 multidimensional:4 scaled:2 hit:1 unit:1 grant:1 evoking:1 producing:1 before:1 engineering:1 influencing:2 dropped:1 limit:1 consequence:1 oxford:2 might:1 studied:10 examined:1 testing:2 illusion:1 area:1 empirical:1 gabor:12 projection:7 matching:1 get:1 onto:2 cannot:3 context:5 applying:1 exposure:1 attention:1 typicality:1 insight:1 notion:1 diego:1 target:4 us:2 hypothesis:3 recognition:12 particularly:1 located:1 predicts:7 database:1 observed:2 vlk:1 worst:1 oe:1 ordering:1 yk:1 nimh:1 asked:3 ideally:1 depend:1 tight:1 distinctive:5 completely:1 basis:3 translated:1 easily:2 joint:1 iit:2 various:2 effective:1 cod:1 formation:1 larger:3 distortion:1 ability:2 itself:1 noisy:2 obviously:1 blob:1 propose:1 aligned:1 shiffrin:3 representational:1 intuitive:1 parent:18 produce:1 categorization:1 rotated:1 object:1 wider:1 derive:1 measured:1 exemplar:18 nearest:1 op:9 job:2 c:1 predicted:5 involves:1 filter:7 centered:1 human:22 eigenface:2 require:3 blending:1 around:1 iil:1 cognition:2 predict:5 mapping:1 matthew:1 bald:2 vary:3 achieves:1 purpose:1 estimation:11 outperformed:1 infonnation:1 create:1 weighted:3 clearly:1 gaussian:4 rather:1 probabilistically:1 conjunction:1 derived:2 validated:1 consistently:1 rank:1 baseline:2 attract:1 typically:1 overall:1 orientation:2 plan:1 spatial:1 summed:3 fairly:1 field:1 shaped:1 having:2 sampling:3 look:1 future:2 minimized:1 report:3 stimulus:18 few:1 x_:1 widen:1 resulted:2 phase:1 attempt:1 investigate:1 evaluation:1 male:1 mixture:4 pc:7 behind:1 held:1 dailey:6 wts:3 facial:10 old:13 euclidean:3 theoretical:3 psychological:2 increased:1 modeling:2 earlier:1 giles:1 blended:1 applicability:1 deviation:2 uniform:1 erlbaum:1 reported:1 stored:1 guru:1 morph:11 morphs:21 combined:1 density:13 twelve:1 international:1 ie:4 probabilistic:1 together:1 nosofsky:2 von:3 squared:2 opposed:1 choose:1 cognitive:2 account:11 wk:2 explicitly:3 race:2 fonnulate:1 performed:6 doing:1 cochran:2 simon:1 rmse:7 om:12 il:1 square:1 offaces:1 responded:2 minimize:2 largely:1 judgment:3 climbing:1 identification:1 none:1 lighting:2 researcher:2 published:1 submitted:7 against:1 colleague:1 frequency:1 turk:2 dm:1 endo:3 associated:1 rational:1 bloomington:1 pilot:1 recall:2 distractors:2 dimensionality:2 carefully:1 actually:1 higher:1 response:10 though:2 correlation:3 gcm:8 quality:2 effect:6 normalized:3 true:1 remedy:1 lades:1 during:1 width:7 criterion:2 generalized:3 stress:1 hill:1 image:21 novel:2 recently:1 common:1 empirically:2 volume:1 alann:1 occurred:1 extend:1 unfamiliar:1 significant:2 morphed:1 ai:1 grid:2 had:3 similarity:12 surface:5 etc:1 mccarthy:2 perspective:1 optimizing:1 jolla:1 disappointing:1 store:1 der:3 additional:1 relaxed:1 somewhat:1 determine:1 paradigm:1 ii:2 jet:3 morpho:1 long:1 retrieval:1 prediction:7 basic:1 hair:1 essentially:1 kernel:22 represent:1 normalization:2 background:2 cropped:1 whereas:2 want:1 interval:1 comment:2 subject:20 tend:2 ample:2 member:1 effectiveness:1 call:1 near:1 affect:1 fit:9 psychology:5 reduce:1 idea:1 prototype:4 avenue:1 whether:1 six:2 reinitz:2 adequate:1 useful:1 generally:1 detailed:1 eigenvectors:2 category:5 neuroscience:1 correctly:2 key:1 four:1 threshold:1 graph:1 sum:2 convert:1 place:1 almost:4 reasonable:1 parsimonious:1 decision:1 prefer:2 scaling:4 fold:1 encountered:1 annual:1 ijcnn:1 encodes:1 nearby:1 aspect:1 speed:1 department:2 combination:2 poor:1 sam:2 modification:1 intuitively:1 outlier:2 invariant:1 taken:1 describing:1 mechanism:1 studying:1 unreasonable:1 probe:7 quarterly:1 hierarchical:1 thomas:1 original:2 responding:1 include:1 malsburg:3 society:1 blend:1 md:21 responds:2 dp:4 distance:6 attentional:2 thank:1 oa:1 degrade:1 chris:1 extent:3 toward:1 besides:1 relationship:1 providing:1 minimizing:1 potentially:2 negative:1 allowing:2 sm:13 pentland:2 variability:1 digitized:1 ucsd:1 varied:1 rating:3 introduced:1 pair:10 yll:1 unbelievable:1 narrow:1 tanaka:2 able:3 below:1 perception:1 pattern:2 built:2 including:1 memory:20 critical:2 overlap:1 natural:1 rely:1 predicting:2 buhmann:2 representing:1 improve:2 eye:2 created:1 x8:1 prior:7 understanding:1 acknowledgement:1 review:1 fully:2 men:1 limitation:1 proportional:2 degree:1 sufficient:1 consistent:5 principle:1 bank:1 supported:1 free:3 keeping:1 bias:1 allow:2 neighbor:2 eigenfaces:6 face:57 distinctiveness:17 template:1 sparse:1 boundary:2 evaluating:2 made:2 adaptive:5 san:1 projected:1 unstudied:1 ignore:1 implicitly:1 evoke:1 global:3 gem:1 grayscale:1 iill:3 table:2 reasonably:1 robust:1 ca:1 domain:1 sp:13 linearly:1 motivation:1 noise:2 alarm:1 scored:1 mahwah:1 child:2 allowed:1 solso:2 garrison:1 aid:1 position:3 explicit:1 third:1 wavelet:1 british:1 familiarity:13 bishop:3 r2:1 experimented:1 sims:2 evidence:1 false:3 sparseness:1 flavor:1 ilx:1 simply:3 likely:1 twentieth:1 visual:1 gary:2 towards:1 change:2 included:1 typical:2 determined:1 except:1 principal:3 tendency:1 la:1 experimental:4 indicating:1 support:1 searched:1 dissimilar:6 evaluate:1 tested:1 phenomenon:1 correlated:1 |
577 | 1,528 | Orientation, Scale, and Discontinuity as
Emergent Properties of Illusory Contour
Shape
Karvel K. Thornber
NEC Research Institute
4 Independence Way
Princeton, NJ 08540
Lance R. Williams
Dept. of Computer Science
University of New Mexico
Albuquerque, NM 87131
Abstract
A recent neural model of illusory contour formation is based on
a distribution of natural shapes traced by particles moving with
constant speed in directions given by Brownian motions. The input
to that model consists of pairs of position and direction constraints
and the output consists of the distribution of contours joining all
such pairs. In general, these contours will not be closed and their
distribution will not be scale-invariant. In this paper, we show
how to compute a scale-invariant distribution of closed contours
given position constraints alone and use this result to explain a
well known illusory contour effect.
1
INTRODUCTION
It has been proposed by Mumford[3] that the distribution of illusory contour shapes
can be modeled by particles travelling with constant speed in directions given by
Brownian motions. More recently, Williams and Jacobs[7, 8] introduced the notion
of a stochastic completion field, the distribution of particle trajectories joining pairs
of position and direction constraints, and showed how it could be computed in a
local parallel network. They argued that the mode, magnitude and variance of
the completion field are related to the observed shape, salience, and sharpness of
illusory contours.
Unfortunately, the Williams and Jacobs model, as described, has some shortcomings. Recent psychophysics suggests that contour salience is greatly enhanced by
closure[2]. Yet, in general, the distribution computed by the Williams and Jacobs
model does not consist of closed contours. Nor i:; it scale-invariant-doubling the
distances between the constraints does not produce a comparable completion field of
K. K. Thornber and L. R. Williams
832
double the size without a corresponding doubling of the particle's speeds. However,
the Williams and Jacobs model contains no intrinsic mechanism for speed selection. The speeds (like the directions) must be specified a priori. In this paper, we
show how to compute a scale-invariant distribution of closed contours given position
constraints alone.
2
2.1
TECHNICAL DETAILS
SHAPE DISTRIBUTION
Consistent with our earlier work[5, 6], in this paper we do not use the same distribution described by Mumford[3] but instead assume a distribution of completion
shapes consisting of straight-line base-trajectories modified by random impulses
drawn from a mixture of two limiting distributions. The first distribution consists
of weak but frequently acting impulses (we call this the Gaussian-limit). The distribution of these weak impulses has zero mean and variance equal to (7~. The weak
impulses act at Poisson times with rate R g . The second distribution consists of
strong but infrequently acting impulses (we call this the Poisson-limit). Here, the
magnitude of the random impulses is Gaussian distributed with zero mean. However, the variance is equal to (72 (where (7~ ? (7;). The strong impulses act at
Poisson times with rate Rp < <
Particles decay with half-life equal to a parameter T. The effect is that particles tend to travel in smooth, short paths punctuated
by occasional orientation discontinuities. See [5, 6].
kg.
2.2
EIGENSOURCES
Let i and j be position and velocity constraints, (xi,id and (xj,Xj). Then P(jl i)
is the conditional probability that a particle beginning at i will reach j. Note that
these transition probabilities are not symmetric, i.e., P(j 1 i) 1= P(i 1 j). However, by
time-reversal symmetry, P(j 1 i) = P(I 1 J) where I = (Xi, -Xi) and J = (Xj, -Xj).
Given only the matrix of transition probabilities, P, we would like to compute the
relative number of closed contours satisfying a given position and velocity constraint.
We begin by noting that, due to their randomness, only increasingly smaller and
smaller fractions of contours are likely to satisfy increasing numbers of constraints.
Suppose we let s~l) contours start at Xi with Xi. Then
(2) _ ~
Sj
- ui
P( J'I '/.)
(1)
, Si
is the relative number of contours through Xj with Xj, i.e., which satisfy two constraints. In general,
(n+1) _ ~ P( 'I .) (n)
Sj
- ui
J '/, Si
Now suppose we compute the eigenvector,
with largest, real positive eigenvalue, and take s~1) = Si. Then clearly si n + 1 ) = AnSi.
This implies that as the number of constraints satisfied increases by one, the number
of contours remaining in the sample of interest decreases by A. However, the ratios
of the Si remain invariant. Letting n pass to infinity, we see that the Si are just
the relative number of contours through i. To summarize, having started with all
possible contours, we are now left with only those bridging pairs of constraints at
all past-times. By solving AS = Ps for s we know their relative numbers. We refer
to the components of s as the eigensources of the stochastic completion field.
Emergent Properties of Illusory Contour Shape
2.3
833
STOCHASTIC COMPLETION FIELDS
Note that the eigensources alone do not represent a distribution of closed contours.
In fact, the majority of contours contributing to s will not satisfy a single additional
constraint. However , the following recurrence equation gives the number of contours
which begin at constraint i and end at constraint j and satisfy n - 1 intermediate
constraints
p(n+1) (j I i) = Lk P(j I k)p(n) (k Ii)
where p( 1) (j I i) = P(j Ii). Given the above recurrence equation , we can define an
expression for the relative number of contours of any length which begin and end
at constraint i:
Ci = lim n -+ oo p(n)(i I i)/ Lj p(n)(j I))
Using a result from the theory of positive matrices[l}, it is possible to show that
the above expression is simply
Ci = Si 8 d Lj Sj8j
where sand s are the right and left eigenvectors of P with largest positive real
eigenvalue, i.e., AS = Ps and AS = pTs. Because of the time-reversal symmetry
of P, the right and left eigenvectors are related by a permutation which exchanges
opposite directions, i.e. , 8i = St.
Finally, given sand s, it is possible to compute the relative number of closed
contours through an arbitrary position and velocity in the plane, i.e., to compute
the stochastic completion field. If", = (x, x) is an arbitrary position and velocity
in the plane, then
C(",) = >.s~s Li P(", I i)Si . Lj P(j I",)8j
gives the relative probability that a closed contour will pass through ",. Note, that
this is a natural generalization of the Williams and Jacobs[7] factorization of the
completion field into the product of source and sink fields.
2.4
SCALE-INVARIANCE
Under the restriction that particles have constant speed, the transition probability
matrix, P, becomes block-diagonal. Each block corresponds to a different possible
speed, 'Y- Since the components of any given eigenvector will be confined to a single
block , we can consider P to be a function of, and solve:
A(r) s(r) = P(r)s(r)
Let Amax (r) be the largest positive real eigenvalue of P(r) and let ,max be the speed
where Amax (r) is maximized. Then Sma x (rma x), i.e., the eigenvector of P (rmax)
associated with Amax (rma x), is the limiting distribution over all spatial scales.
3
3.1
EXPERIMENTS
EIGHT POINT CIRCLE
Given eight points spaced uniformly around the perimeter of a circle of diameter,
d = 16, we would like to find the distribution of directions through each point and
the corresponding completion field (Figure 1 (left)). Neither the order of traversal ,
directions, i.e., xdlxil, or speed , i.e. , , = IXil. are specified a priori. In all of
our experiments, we sample direction at 5? intervals. Consequently, there are 72
discrete directions and 576 position-direction pairs, i.e., P(r) is of size 576 x 576. 1
lThe parameters defining the distribution of completion shapes are T = Rga~ = 0.0005
and 'T = 9.5. For simplicity, we assume the pure Gaussian-limit case described in [6] .
K. K. Thornber and L. R. Williams
834
?
? ? ?
? ? ?
(two
sizes)
o
b
.-
w
I
/
"
Circle
,0
?
"
I
Point
w~
I .
a
/
Eight
,..,,,o
?a
I
i
Il.
'0
xo
og~~~~~~~~~~~~~~~
:l~
o
0
"
,0
'"
30
20
c
d
Figure 1: Left: (a) The eight position constraints. Neither the order of traversal, directions, or speed are specified a priori. (b) The eigenvector, Smax (,max) represents the limiting distribution over all spatial scales. (c) The product of smaxC!max) and smaxC!max).
Orientations tangent to the circle dominate the distribution of closed contours. (d) The
stochastic completion field, C, due to smaxC!max). Right: Plot of magnitude of maximum
positive real eigenvalue, >'max, vs. logl.l (1/,) for eight point circle with d = 16.0 (solid)
and d = 32.0 (dashed).
~
==
~
11
~
==::J
U
Figure 2: Observers report that as the width of the arms increases, the shape of the
illusory contour changes from a circle to a square[4].
First, we evaluated Amax b) over the velocity interval [1.1- 1 , 1.1- 3o J using standard
numerical routines and plotted the magnitude of the largest, real positive eigenvalue,
Amax vs. logl.l(l/,). The function reaches its maximum value at '"'(max:::::: 1.1- 2
Consequently, the eigenvector, Smax (1.1 - 2 ?) represents the limiting distribution over
all spatial scales (Figure 1 (right)).
?.
Next, we scaled the test Figure by a factor of two, i.e., d' = 32.0 and plotted
A~axb) over the same interval (Figure 1 (right)). We observe that A~ax(1.1-x+7)
:::::: Amax (1.1- X ), i.e., when plotted using a logarithmic x-axis, the functions are
identical except for a translation. It follows that '"'(~ax :::::: logl.1 7 x '"'(max:::::: 2.0 x '"'(max'
This confirms the scale-invariance of the system-doubling the size of the Figure
results in a doubling of the selected speed.
3.2
KOFFKA CROSS
The Koffka Cross stimulus (Figure 2) has two basic degrees of freedom which we call
diameter (i.e. , d) and arm width (i .e., w) (Figure 3 (a)). We are interested in how
Emergent Properties of Illusory Contour Shape
(a)
(b)
r---~
o
(-0 5w . O.5d)
835
( O.5w ,O.Sd )
(--O.Sd , 05w)
(e)
(OSd , 05w)
r-----
.- -......,
d
U
(--OSd, --05w)
( 0 Sd ,-O.5w )
(--05w. -O.5d)
(O.5w. --OSd )
n
u
(d)
------,
Figure 3: (a) Koffka Cross showing diameter, d, and width , w. (b) Orientation and
position constraints in terms of d and w. The normal orientation at each endpoint
is indicated by the solid lines while the dashed lines represent plus or minus one
standard deviation (i.e. , 12 .8?) of the Gaussian weighting function. (c) Typically
perceived as square. (d) Typically perceived as circle. The positions of the line
endpoints is the same.
the stochastic completion field changes as these parameters are varied. Observers
report that as the width of the arms increases, the shape of the illusory contour
changes from a circle to a square[4]. The endpoints of the lines comprising the
Koftka Cross can be used to define a set of position and orientation constraints
(Figure 3 (b)). The position constraints are specified in terms of the parameters, d
and w. The orientation constraints take the form of a Gaussian weighting function
which assigns higher probabilities to contours passing through the endpoints with
orientations normal to the lines. 2 The prior probabilities assigned to each positiondirection pair by the Gaussian weighting function form a diagonal matrix, D:
where P(r) is the transition probability matrix for the random process at scale
" A(r) is an eigenvalue of Q(,), and s(r) is the corresponding eigenvector. Let
Amax(r) be the largest positive real eigenvalue of Q(r) and let ,max be the scale
where Amax(r) is maximized. Then smax(rmax), i.e., the eigenvector of Q(rmax)
associated with Amax (rma x), is the limiting distribution over all spatial scales.
First, we used a Koffka Cross where d = 2.0 and w = 0.5 and evaluated Amax (r) over
the velocity interval [8.0 x 1.1- 1 , 8.0 x 1.1- 8 ?] using standard numerical routines. 3
The function reaches its maximum value at ,max::::; 8.0 X 1.1- 62 (Figure 4 (left)).
Observe that the completion field due to the eigenvector, smax(8.0 x 1.1- 62 ), is
dominated by contours of a predominantly circular shape (Figure 4 (right)). We
then uniformly scaled the Koffka Cross Figure by a factor of two, i.e., d' = 4.0 and
20bserve that Figure 3 (c) is perceived as a square while Figure 3 (d) is perceived as a
circle. Yet the positions of the line endpoints is the same. It follows that the orientations
of the lines affect the percept. We have chosen to model this dependence through the use
of a Gaussian weighting function which favors contours passing through the endpoints of
the lines in the normal direction. It is possible to motivate this based on the statistics of
natural scenes. The distribution of relative orientations at contour crossings is maximum
at 90? and drops to nearly zero at 0 0 and 180 0 ?
3The parameters defining the distribution of completion shapes were: T = RgO'~ =
0.0005, T = 9.5, ?p = O'~/T = 100.0 and Rp = 1.0 X 10- 8 . As an anti-aliasing measure, the
transition probabilities, P(j I i) , were averaged over initial conditions modeled as Gaussians
of variance
= = 0.00024 and O'J = 0.0019. See [6].
0'; 0';
K. K. Thornber and L. R. Williams
836
Koffke
...
Crosses
(TWO
sizes)
'"'~
LIl~
_ -_=0>==
,
I a
o
-o
>
c
00
-.
CI~
Wo
-o
,
Ira
...a
~o
a
D..
;
~/--
X0
/
i
/
og-HTnTrnTnTrnTnTrnTnTnTrnTn~~
~ci
a
20
'0
60
.. 0
x
Figure 4: Left: Plot of magnitude of maximum positive real eigenvalue, >'max, vs.
logl.l (1h) for Koffka Crosses with d = 2.0 and w = 0.5 (solid) and d = 4.0 and w = 1.0
(dashed). Right: The completion field due to the eigenvector, sma x (8 .0 x 1.1- 62 ) .
w' = 1.0 and plotted Anax (,) over the same interval (Figur~ 4 (left)) . Observe
that A~ax (8.0 X 1.1- x+ ) :::::: Amax(8.0 x 1.1- X). As before, thls confirms the scaleinvariance of the system.
Next, we studied how the relative magnitudes of the local maxima of Amax (,)
change as the parameter w is varied. We begin with a Koffka Cross where d = 2.0
and w = 0.5 and observe that Amax(r) has two local maxima (Figure 5 (left)).
We refer to the larger of these maxima as ,circle . As previously noted, this maximum is located at approximately 8.0 x 1.1- 62 . The second maximum is located
at approximately 8.0 x 1.1 -32. When the completion field due to the eigenvector,
smax(8.0 x 1.1- 32 ), is rendered, we observe that the distribution is dominated by
contours of predominantly square shape (Figure 5(a)). For this reason , we refer
to this local maximum as ,square. Now consider a Koffka Cross where the widths
of the arms are doubled but the diameter remains the same, i.e., d' = 2.0 and
w' = 1.0. We observe that A~ax (r) still has two local maxima, one at approximately 8.0 x 1.1- 63 and a second at approximately 8.0 x l.1- 29 (Figure 5 (left)).
When we render the completion fields due to the eigenvectors, s~ax(8.0x 1.1- 63 ) and
s~ax(8.0 x 1.1- 29 ), we find that the completion fields have the same general character as before-the contours associated with the smaller spatial scale (i.e., lower
speed) are approximately circular and those associated with the larger spatial scale
(Le., higher speed) are approximately square (Figure 5 (d) and (c)). Accordingly,
we refer to the locations of the respective local maxima as '~ircle and ,~quar e ' However, what is most interesting is that the relative magnitudes of the local maxima
have reversed. Whereas we previously observed that Amax(,circle) > Amax(rsquare),
we now observe that A~ax(r~quare) > A~ax(r~ircle)' Therefore, the completion field
due to the eigenvector, s~ax(r~quar e ) [not s~ax(r~ircle)!l represents the limiting
distribution over all spatial scales. This is consistent with the transition from circle
to square reported by human observers when the widths of the arms of the Koffka
Cross are increased.
837
Emergent Properties of Illusory Contour Shape
Koffke
Crosses
(two
widths)
-
b
-, ~:...-
a
c
o
>
(
Gl o
.-wo.
OI~
d
a
b
c
d
-o
??
(l:o
o
~o
o
Q.
20
40
60
80
X
Figure 5: Plot of magnitude of maximum positive real eigenvalue, Ama x, vs. log 1.1 (1/"'()
for Koffka Crosses with d = 2.0 and w = 0.5 (solid) and d = 2.0 and w = 1.0 (dashed) .
Stochastic completion fields for Koffka Cross due to (a) Sm ax ("'(.quar e ) is a local optimum
for w = 0.5 (b) Sma x ("'(ci rcl e ) is the global optimum for w = 0.5 (c) s~ax("'(~quar e ) is the
global optimum for w = 1.0 (d) s~a x ("'(~quar e ) is a local optimum for w = 1.0. These
results are consistent with the circle-to-square transition perceived by human subjects
when the width of the arms of the Koffka Cross are increased .
4
CONCLUSION
We have improved upon a previous model of illusory contour formation by showing how to compute a scale-invariant distribution of closed contours given position
constraints alone. We also used our model to explain a previously unexplained
perceptual effect.
References
[1] Horn, R.A ., and C.R. Johnson, Matrix Analysis, Cambridge Univ. Press , p. 500 ,
1985.
[2] Kovacs, I. and B. Julesz, A Closed Curve is Much More than an Incomplete One:
Effect of Closure in Figure-Ground Segmentation, Pmc. Natl. Acad. Sci. USA, 90,
pp. 7495-7497, 1993.
[3] Mumford, D., Elastica and Computer Vision, Algebraic Geometry and Its Applications, Chandrajit Bajaj (ed.) , Springer-Verlag, New York, 1994.
[4) Sambin , M., Angular Margins without Gradients, Italian Journal of Psychology 1,
pp. 355-361, 1974.
[5] Thornber, KK and L.R. Williams, Analytic Solution of Stochastic Completion
Fields, Biological Cybernetics 75 , pp. 141-151, 1996.
[6] Thornber , KK and L.R. Williams, Characterizing the Distribution of Completion
Shapes with Corners Using a Mixture of Random Processes, Intl. Workshop on
Energy Minimization Methods in Computer Vision, Venice, Italy, 1997.
[7] Williams, L.R. and D.W . Jacobs, Stochastic Completion Fields: A Neural Model of
Illusory Contour Shape and Salience, Neural Computation 9(4) , pp. 837-858, 1997.
[8) Williams, L.R. and D.W. Jacobs, Local Parallel Computation of Stochastic Completion Fields, Neural Computation 9(4), pp. 859-881 , 1997.
| 1528 |@word closure:2 confirms:2 jacob:7 minus:1 solid:4 initial:1 contains:1 past:1 si:8 yet:2 must:1 ixil:1 numerical:2 shape:17 analytic:1 plot:3 drop:1 v:4 alone:4 half:1 selected:1 accordingly:1 plane:2 beginning:1 short:1 location:1 consists:4 x0:1 nor:1 frequently:1 aliasing:1 increasing:1 becomes:1 begin:4 what:1 kg:1 rmax:3 eigenvector:11 nj:1 act:2 scaled:2 positive:9 before:2 local:10 sd:3 limit:3 acad:1 joining:2 id:1 path:1 approximately:6 plus:1 studied:1 suggests:1 factorization:1 averaged:1 horn:1 block:3 doubled:1 selection:1 restriction:1 williams:13 punctuated:1 sharpness:1 simplicity:1 assigns:1 pure:1 amax:15 dominate:1 notion:1 limiting:6 enhanced:1 suppose:2 pt:1 crossing:1 velocity:6 satisfying:1 infrequently:1 located:2 observed:2 decrease:1 ui:2 traversal:2 motivate:1 solving:1 upon:1 sink:1 emergent:4 univ:1 shortcoming:1 formation:2 larger:2 solve:1 favor:1 statistic:1 eigenvalue:9 product:2 elastica:1 rma:3 ama:1 double:1 p:2 optimum:4 smax:5 produce:1 intl:1 oo:1 completion:24 figur:1 quare:1 strong:2 implies:1 direction:13 stochastic:10 human:2 sand:2 argued:1 exchange:1 generalization:1 biological:1 around:1 ground:1 normal:3 sma:3 perceived:5 travel:1 unexplained:1 largest:5 minimization:1 clearly:1 gaussian:7 modified:1 og:2 ira:1 ax:12 greatly:1 lj:3 typically:2 italian:1 interested:1 comprising:1 orientation:10 priori:3 ircle:3 spatial:7 psychophysics:1 field:21 equal:3 having:1 identical:1 represents:3 nearly:1 report:2 stimulus:1 koffka:12 geometry:1 consisting:1 freedom:1 interest:1 circular:2 mixture:2 natl:1 perimeter:1 respective:1 incomplete:1 circle:13 plotted:4 increased:2 earlier:1 deviation:1 johnson:1 reported:1 st:1 nm:1 satisfied:1 corner:1 li:1 rcl:1 satisfy:4 observer:3 closed:11 start:1 parallel:2 il:1 square:9 oi:1 variance:4 percept:1 maximized:2 spaced:1 weak:3 albuquerque:1 trajectory:2 cybernetics:1 straight:1 randomness:1 explain:2 reach:3 ed:1 energy:1 pp:5 associated:4 illusory:12 lim:1 segmentation:1 routine:2 higher:2 improved:1 evaluated:2 just:1 angular:1 logl:4 mode:1 indicated:1 impulse:7 usa:1 effect:4 assigned:1 symmetric:1 width:8 recurrence:2 noted:1 motion:2 recently:1 predominantly:2 endpoint:6 jl:1 refer:4 cambridge:1 particle:8 moving:1 rgo:1 base:1 brownian:2 recent:2 showed:1 italy:1 verlag:1 life:1 additional:1 dashed:4 ii:2 smooth:1 technical:1 cross:15 basic:1 vision:2 poisson:3 represent:2 confined:1 thornber:6 whereas:1 interval:5 source:1 subject:1 tend:1 call:3 noting:1 intermediate:1 scaleinvariance:1 independence:1 xj:6 affect:1 psychology:1 opposite:1 expression:2 bridging:1 wo:2 render:1 algebraic:1 passing:2 york:1 chandrajit:1 eigenvectors:3 julesz:1 diameter:4 discrete:1 traced:1 drawn:1 neither:2 fraction:1 ansi:1 venice:1 comparable:1 constraint:22 infinity:1 scene:1 dominated:2 speed:13 rendered:1 smaller:3 remain:1 increasingly:1 character:1 invariant:6 xo:1 equation:2 previously:3 remains:1 mechanism:1 know:1 letting:1 reversal:2 end:2 travelling:1 gaussians:1 eight:5 observe:7 occasional:1 rp:2 remaining:1 mumford:3 dependence:1 diagonal:2 gradient:1 distance:1 reversed:1 sci:1 majority:1 lthe:1 reason:1 length:1 modeled:2 kk:2 ratio:1 mexico:1 pmc:1 unfortunately:1 lil:1 sm:1 anti:1 defining:2 varied:2 arbitrary:2 kovacs:1 introduced:1 pair:6 specified:4 discontinuity:2 summarize:1 max:12 lance:1 natural:3 arm:6 lk:1 started:1 axis:1 prior:1 tangent:1 contributing:1 relative:10 permutation:1 interesting:1 degree:1 consistent:3 translation:1 koftka:1 gl:1 salience:3 institute:1 characterizing:1 distributed:1 curve:1 transition:7 contour:39 osd:3 sj:2 global:2 xi:5 symmetry:2 bajaj:1 position:16 perceptual:1 weighting:4 showing:2 decay:1 consist:1 intrinsic:1 workshop:1 ci:5 nec:1 magnitude:8 margin:1 logarithmic:1 simply:1 likely:1 doubling:4 springer:1 corresponds:1 conditional:1 consequently:2 axb:1 change:4 except:1 uniformly:2 acting:2 pas:2 invariance:2 dept:1 princeton:1 |
578 | 1,529 | Stationarity and Stability of
Autoregressive Neural Network Processes
Friedrich Leisch\ Adrian Trapletti 2 & Kurt Hornik l
1 Institut fur Statistik
Technische UniversiUit Wien
Wiedner Hauptstrafie 8-10 / 1071
A-1040 Wien, Austria
firstname.lastname@ci.tuwien.ac.at
2
Institut fiir Unternehmensfiihrung
Wirtschaftsuniversi tat Wien
Augasse 2-6
A-lOgO Wien, Austria
adrian. trapletti@wu-wien.ac.at
Abstract
We analyze the asymptotic behavior of autoregressive neural network (AR-NN) processes using techniques from Markov chains and
non-linear time series analysis. It is shown that standard AR-NNs
without shortcut connections are asymptotically stationary. If linear shortcut connections are allowed, only the shortcut weights
determine whether the overall system is stationary, hence standard
conditions for linear AR processes can be used.
1
Introduction
In this paper we consider the popular class of nonlinear autoregressive processes
driven by additive noise, which are defined by stochastic difference equations of
form
(1)
where ft is an iid. noise process. If g( . .. , (J) is a feedforward neural network with
parameter ("weight") vector (J, we call Equation 1 an autoregressive neural network
process of order p, short AR-NN(p) in the following.
AR-NNs are a natural generalization of the classic linear autoregressive AR(p) process
(2)
See, e.g., Brockwell & Davis (1987) for a comprehensive introduction into AR and
ARMA (autoregressive moving average) models.
F. Leisch, A. Trapletti and K. Hornik
268
One of the most central questions in linear time series theory is the stationarity of
the model, i.e., whether the probabilistic structure of the series is constant over time
or at least asymptotically constant (when not started in equilibrium). Surprisingly,
this question has not gained much interest in the NN literature, especially there
are-up to our knowledge-no results giving conditions for the stationarity of ARNN models. There are results on the stationarity of Hopfield nets (Wang & Sheng,
1996), but these nets cannot be used to estimate conditional expectations for time
series prediction.
The rest of this paper is organized as follows: In Section 2 we recall some results
from time series analysis and Markov chain theory defining the relationship between
a time series and its associated Markov chain. In Section 3 we use these results to
establish that standard AR-NN models without shortcut connections are stationary.
We also give conditions for AR-NN models with shortcut connections to be stationary. Section 4 examines the NN modeling of an important class of non-stationary
time series, namely integrated series. All proofs are deferred to the appendix .
2
2.1
Some Time Series and Markov Chain Theory
Stationarity
Let ~t denote a time series generated by a (possibly nonlinear) autoregressive process as defined in (1). If lEft = 0, then 9 equals the conditional expectation
1E(~t I~t-l' ... , ~t-p) and g(~t-l' ... , ~t-p) is the best prediction for ~t in the mean
square sense.
If we are interested in the long term properties of the series, we may ask whether
certain features such as mean or variance change over time or remain constant.
The time series is called weakly stationary if lE~t = Jl and cov(~t,~t+h) = ,h, 'it,
i.e., mean and covariances do not depend on the time t. A stronger criterion is
that the whole distribution (and not only mean and covariance) of the process does
not depend on the time, in this case the series is called strictly stationary. Strong
stationarity implies weak stationarity if the second moments of the series exist. For
details see standard time series textbooks such as Brockwell & Davis (1987).
=
If ~t is strictly stationary, then IP (~t E A)
rr( A), 'it and rrO is called the stationary
distribution of the series. Obviously the series can only be stationary from the
beginning if it is started with the stationary distribution such that ~o '" rr. If
it is not started with rr, e.g., because ~o is a constant , then we call the series
asymptotically stationary if it converges to its stationary distribution:
lim IP(~t E A)
t-HX)
2.2
= rr(A)
Time Series as Markov Chains
Using the notation
Xt-l
(~t-l"" ,~t_p)'
(3)
(g(Xt-d,~t-l"" , ~t-p+d
(ft,O, ... ,O)'
(4)
(5)
we can write scalar autoregressive models of order p such as (1) or (2) as a first
order vector model
(6)
269
Stationarity and Stability ofAutoregressive Neural Network Processes
with Xt, et E lR P (e.g., Chan & Tong, 1985). If we write
=
pn(x,A)
p( x, A)
=
IP{Xt+n E Alxt
pl (x, A)
= x}
for the probability of going from point x to set A E B in n steps, then {xd with
p(x , A) forms a Markov chain with state space (lRP , B,>'), where B are the Borel
sets on lR P and>' is the usual Lebesgue measure.
The Markov chain {xd is called cp-irreducible, if for some IT-finite measure cp on
(lR P , B, >.)
00
n=l
whenever cp(A) > O. This me~ns essentially, that all parts of the state space can be
reached by the Markov chain irrespective of the starting point. Another important
property of Markov chains is aperiodicity, which loosely speaking means that there
are no (infinitely often repeated) cycles. See, e.g., Tong (1990) for details .
The Markov chain {Xt} is called geometrically ergodic, if there exists a probability
measure 1I"(A) on (lR P , B, >.) and a p > 1 such that
Vx E lR P
:
lim pnllpn(x,.) - 11"(?)11 = 0
n-+oo
where II . II denotes the total variation. Then 11" satisfies the invariance equation
1I"(A)
=
!
p(x, A) 1I"(dx) ,
VA E B
There is a close relationship between a time series and its associated Markov chain.
If the Markov chain is geometrically ergodic, then its distribution will converge to
11" and the time series is asymptotically stationary. If the time series is started with
distribution 11", i.e., Xo "" 11", then the series {~d is strictly stationary.
3
Stationarity of AR-NN Models
We now apply the concepts defined in Section 2 to the case where 9 is defined by
a neural network. Let x denote a p-dimensional input vector, then we consider the
following standard network architectures:
Single hidden layer perceptrons:
g(x)
= 'Yo + L,8ilT(ai + a~x)
(7)
where ai, ,8i and 'Yo are scalar weights, aj are p-dimensional weight vectors,
and IT(') is a bounded sigmoid function such as tanh(?).
Single hidden layer perceptrons with shortcut connections:
(8)
where c is an additional weight vector for shortcut connections between
inputs and output. In this case we define the characteristic polynomial c(z)
associated with the linear shortcuts as
c(z)
=1-
ClZ - C2z2 - . .. -
cP zP,
ZE
C.
270
F. Leisch, A. TrapleUi and K. Hornik
Radial basis function networks:
(9)
where mj are center vectors and ?( ...) is one of the usual bounded radial
basis functions such as ?(x) = exp( _x 2 ).
Lemma 1 Let {xtl be defined by (6), let IEjt:tl < 00 and let the PDF of f:t be
positive everywhere in JR. Then if 9 is defined by any of (7), (8) or (9), the Markov
chain {Xt} is ?-irreducible and aperiodic.
Lemma 1 basically says that the state space of the Markov chain, i.e., the points that
can be reached, cannot be reduced depending on the starting point. An example
for a reducible Markov chain would be a series that is always positive if only Xo >
(and negative otherwise). This cannot happen in the AR-NN(p) case due to the
unbounded additive noise term.
?
Theorem 1 Let {~tl be defined by (1), {xtl by (6), further let IEktl
PDF of f:t be positive everywhere in JR. Then
< 00
and the
1. If 9 is a network without linear shortcuts as defined in (7) and (9), then
{ x tl is geometrically ergodic and {~tl is asymptotically stationary.
2. If 9 is a network with linear shortcuts as defined in (8) and additionally
c(z) f 0, Vz E C : Izl ~ 1, then {xtl is geometrically ergodic and {~tl is
asymptotically stationary.
The time series {~t} remains stationary if we allow for more than one hidden layer
(-+ multi layer perceptron, MLP) or non-linear output units, as long as the overall
mapping has bounded range. An MLP with shortcut connections combines a (possibly non-stationary) linear AR(p) process with a non-linear stationary NN part.
Thus, the NN part can be used to model non-linear fluctuations around a linear
process like a random walk.
The only part of the network that controls whether the overall process is stationary
are the linear shortcut connections (if present). If there are no shortcuts, then the
process is always stationary. With shortcuts, the usual test for stability of a linear
system applies.
4
Integrated Models
An important method in classic time series analysis is to. first transform a nonstationary series into a stationary one and then model the remainder by a stationary
process. The probably most popular models of this kind are autoregressive integrated moving average (ARIMA) models, which can be transformed into stationary
ARMA processes by simple differencing.
Let I::!..k denote the k-th order difference operator
et -
~t-l
I::!..(~t - ~t-d
(10)
= ~t -
2~t-l
+ ~t-2
(11)
(12)
271
Stationarity and Stability ofAutoregressive Neural Network Processes
with ~ 1 = ~. E.g., a standard random walk ~t = ~t-l +ft is non-stationary because
of the growing variance, but can be transformed into the iid (and hence stationary)
noise process ft by taking first differences.
If a time series is non-stationary, but can be transformed into a stationary series
by taking k-th differences, we call the series integrated of order k. Standard MLPs
or RBFs without shortcuts are asymptotically stationary. It is therefore important
to take care that these networks are only used to model stationary processes. Of
course the network can be trained to mimic a non-stationary process on a finite time
interval, but the out-of-sample or prediction performance will be poor, because the
network inherently cannot capture some important features of the process. One way
to overcome this problem is to first transform the process into a stationary series
(e.g., by differencing an integrated series) and train the network on the transformed
series (Chng et al., 1996).
As differencing is a linear operation, this transformation can also be easily incorporated into the network by choosing the shortcut connections and weights from
input to hidden units accordingly. Assume we want to model an integrated series
of integration order k, such that
~k~t = g(~k~t_l' . .. ' ~k~t_p)
+ ft
where ~k~t is stationary. By (12) this is equivalent to
~t
k
~(-lt-l (~)~t-n + g(~k~t_l' ... ' ~k~t_p) +
ft
k
~(-lt-l (~)~t-n + g(~t-l' ... ,~t-p-k) +
ft
which (for p > k) can be modeled by an MLP with shortcut connections as defined
by (8) where the shortcut weight vector c is fixed to
(~)
:= 0 for n
>k
and 9 is such that g(~t-l' ... ,~t-p-k) = g(~kXt_d. This is always possible and
can basically be obtained by adding c to all weights between input and first hidden
layer of g.
An AR-NN(p) can model integrated series up to integration order p. If the order
of integration is known , the shortcut weights can either be fixed, or the differenced
series is used as input. If the order is unknown, we can also train the complete
network including the shortcut connections and implicitly estimate the order of
integration. After training the final model can be checked for stationarity by looking
at the characteristic roots of the polynomial defined by the shortcut connections.
4.1
Fractional Integration
Up to now we have only considered integrated series with positive integer order of
integration, i.e., kEN. In the last years models with fractional integration order
became very popular (again). Series with integration order of 0.5 < k < 1 can
be shown to exhibit self-similar or fractal behavior, and have long memory. These
type of processes were introduced by Mandelbrot in a series of paper modeling river
flows, e.g., see Mandelbrot & Ness (1968). More recently, self-similar processes were
used to model Ethernet traffic by Leland et al. (1994). Also some financial time
series such as foreign exchange data series exhibit long memory and self-similarity.
FLeisch. A. Trapletti and K. Hornik
272
The fractional differencing operator ~ k , k E [-1, 1] is defined by the series expansion
k
~
f(-k+n)
~ ~t = ~ r(-k)f(n + 1)~t-n
(13)
which is obtained from the Taylor series of (1 - z)k. For k > 1 we first use Equation (12) and then the above series for the fractional remainder. For practical
computation, the series (13) is of course truncated at some term n = N. An ARNN(p) model with shortcut connections can approximate the series up to the first
p terms.
5
Summary
We have shown that AR-NN models using standard NN architectures without shortcuts are asymptotically stationary. If linear shortcuts between inputs and outputs
are included-which many popular software packages have already implementedthen only the weights of the shortcut connections determine if the overall system
is stationary. It is also possible to model many integrated time series by this kind
of networks . The asymptotic behavior of AR-NNs is especially important for parameter estimation, predictions over larger intervals of time, or when using the
network to generate artificial time series. Limiting (normal) distributions of parameter estimates are only guaranteed for stationary series. We therefore always
recommend to transform a non-stationary series to a stationary series if possible
(e.g ., by differencing) before training a network on it.
Another important aspect of stationarity is that a single trajectory displays the
complete probability law of the process. If we have observed one long enough trajectory of the process we can (in theory) estimate all interesting quantities of the
process by averaging over time. This need not be true for non-stationary processes
in general, where some quantities may only be estimated by averaging over several
independent trajectories. E.g., one might train the network on an available sample and then use the trained network afterwards-driven by artificial noise from a
random number generator-to generate new data with similar properties than the
training sample. The asymptotic stationarity guarantees that the AR-NN model
cannot show "explosive" behavior or growing variance with time.
We currently are working on extensions of this paper in several directions. AR-NN
processes can be shown to be strong mixing (the memory of the process vanishes
exponentially fast) and have autocorrelations going to zero at an exponential rate .
Another question is a thorough analysis of the properties of parameter estimates
(weights) and tests for the order of integration. Finally we want to extend the univariate results to the multivariate case with a special interest towards cointegrated
processes.
Acknowledgement
This piece of research was supported by the Austrian Science Foundation (FWF) under
grant SFB#OlO ('Adaptive Information Systems and Modeling in Economics and Management Science') .
Stationarity and Stability ofAutoregressive Neural Network Processes
273
Appendix: Mathematical Proofs
Proof of Lemma 1
It can easily be shown that {xe} is <p-irreducible if the support of the probability density
function (PDF) of ?t is the whole real line, i.e., the PDF is positive everywhere in IR (Chan
& Tong, 1985). In this case every non-null p-dimensional hypercube is reached in p steps
with positive probability (and hence every non-null Borel set A).
A necessary and sufficient condition for {Xt} to be aperiodic is that there exists a set A
and positive integer n such that pn(x, A) > 0 and pn+l (x, A) > 0 for all x E A (Tong,
1990, p. 455). In our case this is true for all n due to the unbounded additive noise.
Proof of Theorem 1
We use the following result from nonlinear time series theory:
Theorem 2 (Chan & Tong 1985) Let {Xt} be defined by (1), (6) and let G be compact,
i.e. preserve compact sets. IfG can be decomposedasG = Gh+Gd andGd(-) is of bounded
range, G h(-) is continuous and homogeneous, i.e., Gh(ax) = aGh(x), the origin is a fixed
point of G h and Gh is uniform asymptotically stable, IEI?tl < 00 and the PDF of ?t is
positive everywhere in IR, then {Xt} is geometrically ergodic.
The noise process ?t fulfills the conditions by assumption. Clearly all networks are continuous compact functions. Standard MLPs without shortcut connections and RBFs have
a bounded range , hence G h == 0 and G == G d , and the series {ee} is asymptotically stationary. If we allow for linear shortcut connections between the input and the outputs,
we get G h = c'x and G d = 70
l.:i (3i(T(ai aix) i.e., G h is the linear shortcut part
of the network, and Gd is a standard MLP without shortcut connections. Clearly, G h is
continuous, homogeneous and has the origin as a fixed point. Hence, the series {eel is
asymptotically stationary if G h is asymptotically stable, i.e., when all characteristic roots
of Gh have a magnitude less than unity. Obviously the same is true for RBFs with shortcut
connections. Note that the model reduces to a standard linear AR(p) model if Gd == O.
+
+
References
Brockwell, P. J. & Davis, R. A. (1987). Time Series: Theory and Methods. Springer Series
in Statistics. New York, USA: Springer Verlag.
Chan, K. S. & Tong, H. (1985). On the use of the deterministic Lyapunov function for
the ergodicity of stochastic difference equations. Advances in Applied Probability, 17,
666-678.
Chng, E . S., Chen, S., & Mulgrew, B. (1996). Gradient radial basis function networks
for nonlinear and nonstationary time series prediction. IEEE Transactions on Neural
Networks, 7(1), 190- 194.
Husmeier, D. & Taylor, J. G . (1997). Predicting conditional probability densities of stationary stochastic time series. Neural Networks, 10(3),479-497.
Jones, D. A. (1978). Nonlinear autoregressive processes. Proceedings of the Royal Society
London A, 360, 71- 95.
Leland, W. E., Taqqu, M. S., Willinger, W., & Wilson, D. V. (1994) . On the self-similar
nature of ethernet traffic (extended version). IEEE/ACM Transactions on Networking,
2(1), 1- 15.
Mandelbrot, B. B. & Ness, J . W. V. (1968). Fractional brownian motions, fractional noises
and applications. SIAM Review, 10(4), 422-437.
Tong, H. (1990). Non-linear time series: A dynamical system approach. New York, USA:
Oxford University Press.
Wang, T. & Sheng, Z. (1996). Asymptotic stationarity of discrete-time stochastic neural
networks. Neural Networks, 9(6) , 957-963.
| 1529 |@word version:1 polynomial:2 stronger:1 adrian:2 tat:1 covariance:2 moment:1 series:59 kurt:1 dx:1 willinger:1 additive:3 happen:1 stationary:43 accordingly:1 beginning:1 short:1 lr:5 unbounded:2 mathematical:1 mandelbrot:3 xtl:3 combine:1 leland:2 behavior:4 growing:2 multi:1 tuwien:1 notation:1 bounded:5 null:2 kind:2 textbook:1 transformation:1 guarantee:1 thorough:1 every:2 xd:2 control:1 unit:2 grant:1 positive:8 before:1 oxford:1 fluctuation:1 logo:1 might:1 range:3 practical:1 radial:3 get:1 cannot:5 close:1 operator:2 equivalent:1 deterministic:1 center:1 economics:1 starting:2 ergodic:5 agh:1 examines:1 financial:1 stability:5 classic:2 variation:1 limiting:1 homogeneous:2 origin:2 ze:1 observed:1 ft:7 reducible:1 wang:2 capture:1 cycle:1 vanishes:1 taqqu:1 trained:2 weakly:1 depend:2 basis:3 easily:2 hopfield:1 train:3 fast:1 london:1 universiuit:1 artificial:2 choosing:1 larger:1 say:1 otherwise:1 cov:1 statistic:1 transform:3 ip:3 final:1 obviously:2 rr:4 net:2 remainder:2 alxt:1 brockwell:3 mixing:1 zp:1 converges:1 oo:1 depending:1 ac:2 strong:2 implies:1 ethernet:2 lyapunov:1 direction:1 aperiodic:2 stochastic:4 vx:1 exchange:1 hx:1 generalization:1 strictly:3 pl:1 extension:1 around:1 considered:1 normal:1 exp:1 equilibrium:1 mapping:1 estimation:1 tanh:1 currently:1 vz:1 lrp:1 clearly:2 always:4 pn:3 wilson:1 ax:1 yo:2 fur:1 sense:1 nn:15 foreign:1 integrated:9 hidden:5 going:2 transformed:4 interested:1 overall:4 integration:9 ness:2 special:1 equal:1 jones:1 mimic:1 recommend:1 t_l:2 irreducible:3 preserve:1 comprehensive:1 lebesgue:1 explosive:1 stationarity:15 interest:2 mlp:4 deferred:1 chain:15 necessary:1 institut:2 loosely:1 taylor:2 walk:2 arma:2 wiedner:1 modeling:3 ar:18 technische:1 uniform:1 nns:3 gd:3 density:2 river:1 siam:1 probabilistic:1 eel:1 aix:1 again:1 central:1 management:1 possibly:2 iei:1 wien:5 piece:1 root:2 analyze:1 traffic:2 reached:3 rbfs:3 mlps:2 square:1 ir:2 aperiodicity:1 became:1 variance:3 characteristic:3 weak:1 iid:2 basically:2 trajectory:3 networking:1 whenever:1 checked:1 associated:3 proof:4 popular:4 ask:1 austria:2 knowledge:1 recall:1 lim:2 fractional:6 organized:1 ergodicity:1 sheng:2 working:1 nonlinear:5 autocorrelations:1 leisch:3 aj:1 usa:2 concept:1 true:3 hence:5 self:4 lastname:1 davis:3 criterion:1 pdf:5 complete:2 cp:4 gh:4 motion:1 recently:1 sigmoid:1 exponentially:1 jl:1 extend:1 ai:3 moving:2 stable:2 similarity:1 multivariate:1 brownian:1 chan:4 driven:2 certain:1 verlag:1 xe:1 additional:1 care:1 determine:2 converge:1 husmeier:1 ii:2 afterwards:1 reduces:1 long:5 va:1 prediction:5 austrian:1 essentially:1 expectation:2 izl:1 want:2 interval:2 rest:1 probably:1 flow:1 call:3 nonstationary:2 integer:2 fwf:1 ee:1 feedforward:1 enough:1 architecture:2 whether:4 sfb:1 speaking:1 york:2 fractal:1 ken:1 reduced:1 generate:2 exist:1 estimated:1 chng:2 write:2 discrete:1 arima:1 asymptotically:12 geometrically:5 year:1 package:1 everywhere:4 ilt:1 wu:1 appendix:2 layer:5 guaranteed:1 display:1 software:1 statistik:1 aspect:1 poor:1 jr:2 remain:1 unity:1 xo:2 equation:5 remains:1 available:1 operation:1 apply:1 denotes:1 giving:1 especially:2 establish:1 society:1 hypercube:1 question:3 already:1 quantity:2 usual:3 exhibit:2 gradient:1 me:1 modeled:1 relationship:2 differencing:5 negative:1 unknown:1 markov:15 finite:2 truncated:1 defining:1 cointegrated:1 incorporated:1 looking:1 extended:1 introduced:1 namely:1 friedrich:1 connection:18 t_p:3 differenced:1 dynamical:1 firstname:1 including:1 memory:3 royal:1 natural:1 predicting:1 started:4 irrespective:1 review:1 literature:1 acknowledgement:1 asymptotic:4 law:1 interesting:1 generator:1 foundation:1 sufficient:1 course:2 summary:1 surprisingly:1 last:1 supported:1 allow:2 perceptron:1 taking:2 overcome:1 autoregressive:10 adaptive:1 transaction:2 approximate:1 compact:3 implicitly:1 fiir:1 continuous:3 additionally:1 mj:1 nature:1 inherently:1 hornik:4 expansion:1 whole:2 noise:8 allowed:1 repeated:1 tl:6 borel:2 tong:7 n:1 exponential:1 ifg:1 theorem:3 xt:9 exists:2 adding:1 gained:1 ci:1 magnitude:1 chen:1 lt:2 univariate:1 infinitely:1 scalar:2 applies:1 springer:2 satisfies:1 acm:1 conditional:3 towards:1 shortcut:30 change:1 included:1 averaging:2 lemma:3 called:5 total:1 invariance:1 perceptrons:2 support:1 fulfills:1 olo:1 |
579 | 153 | 594
Range Image Restoration
using Mean Field Annealing
Wesley E. Snyder
Griff L. Bilbro
Center for Communications and Signal Processing
North Carolina State University
Raleigh, NC
Abstract
A new optimization strategy, Mean Field Annealing, is presented.
Its application to MAP restoration of noisy range images is derived
and experimentally verified.
1
Introduction
The application which motivates this paper is image analysis; specifically the analysis of range images. We [BS86] [GS87] and others [YA85][BJ88] have found that
surface curvature has the potential for providing an excellent, view-invariant feature with which to segment range images. Unfortunately, computation of curvature
requires, in turn, computation of second derivatives of noisy data.
We cast this task as a restoration problem: Given a measurement g(z, y), we assume
that g(z, y) resulted from the addition of noise to some "ideal" image fez, y) which
we must estimate from three things:
1. The measurement g(z, y).
2. The statistics of the noise, here assumed to be zero mean with variance (1'2.
3. Some a priori knowledge of the smoothness of the underlying surface(s).
We will turn this restoration problem into a minimization, and solve that minimization using a strategy called Mean Field A nnealing. A neural net appears to be
the ideal architecture for the reSUlting algorithm, and some work in this area has
already been reported [CZVJ88].
2
Simulated Annealing and Mean Field Anneal?
Ing
The strategy of SSA may be summarized as follows:
Let H(f) be the objective function whose minimum we se~k, wher~ /is somt' parameter vector.
A parameter T controls the algorithm. The SSA algorithm begins at a relatively
high value of T which is gradually reduced. Under certain conditions, SSA will
converge to a global optimum, [GGB4] [RS87]
H (f) = min{ H (fie)} Vfie
(1)
Range Image Restoration Using Mean Field Annealing
even though local minima may occur. However, SSA suffers from two drawbacks:
? It is slow, and
? there is no way to directly estimate [MMP87] a continuously-valued
derivatives.
I
or its
The algorithm presented in section 2.1 perturbs (typically) a single element of fat
each iteration. In Mean Field Annealing, we perturb the entire vector f at each
iteration by making a deterministic calculation which lowers a certain average of
H, < H(f) >, at the current temperature. We thus perform a rather conventional
non-linear minimization (e.g. gradient descent), until a minimum is found at that
temperature. We will refer to the minimization condition at a given T as the
equilibrium for that T. Then, T is reduced, and the previous equilibrium is used as
the initial condition for another minimization.
MFA thus converts a hard optimization problem into a sequence of easier problems.
In the next section, we justify this approach by relating it to SSA.
2.1
Stochastic Simulated Annealing
The problem to be solved is to find j where
minimization with the following strategy:
j
minimizes H(f). SSA solves this
1. Define PT ex e- H / T .
2. Find the equilibrium conditions on PT, at the current temperature, T. By equilibrium, we mean that any statistic ofpT(f) is constant. These statistics could
be derived from the Markov chain which SSA constructs: jO, p, ... , IN, ... , although in fact such statistical analysis is never done in normal running of an
SSA algorithm.
3. Reduce T gradually.
4. As T --+ 0, PT(f) becomes sharply peaked at j, the minimum.
2.2
Mean Field Annealing
In Mean Field Annealing, we provide an analytic mechanism for approximating the
equilibrium at arbitrary T. In MFA, we define an error function,
-H
fe--ordl
EMF(Z, T) = Tln--=-H- f eTdl
+
fe
-Hfl
T
(H-Ho)dl
- / j ---
f e- TdJ
- --.
(2)
which follows from Peierl's inequality [BGZ76]:
F
-H
~
Fo+ < H - Ho >
-Hg
(3)
where F = -Tlnf e---r-dl and Fo = -Tlnf e T dl . The significance of EMF is as
follows: the minimum of EMF determines the best approximation given the form
595
596
Bilbro and Snyder
of Ho to the equilibrium statistics of the SSA-generated MRF at T. We will then
anneal on T. In the next section, we choose a special form for Ho to simplify this
process even further.
1. Define some Ho(f, z) which will be used to estimate H(f).
2. At temperature T, minimize EMF(Z) where EMF is a functional of Ho and
H which characterizes the difference between Ho and H. The process of
minimizing EMF will result in a value of the parameter z, which we will
denote as ZT.
3. Define HT(f) = Ho(f, ZT) and for(f) ex e- iiT / T .
3
Image Restoration Using MFA
We choose a Hamiltonian which represents both the noise in the image, and our a
priori knowledge of the local shape of the image data.
" -2
1 2 (Ii - gil 2
HN = "
L.J
?
(1'
,
(4)
(5)
where 18( represents [Bes86] the set of values of pixels neighboring pixel i (e.g. the
value of I at i along with the I values at the four nearest neighbors of i); A is some
scalar valued function of that set of pixels (e.g. the 5 pixel approximation to the
Laplacian or the 9 pixel approximation to the quadratic variation); and
(6)
The noise term simply says that the image should be similar to the data, given noise
of variance (1'2. The prior term drives toward solutions which are locally planar. Recently, a simpler V(z) = z2 and a similar A were successfully used to design a neural
net [CZVJ88] which restores images consisting of discrete, but 256-valued pixels.
Our formulation of the prior term emphasizes the importance of "point processes,"
as defined [WP85] by Wolberg and Pavlidis. While we accept the eventual necessity
of incorporating line processes [MMP87] [Mar85] [GG84] [Gem87] into restoration,
our emphasis in this paper is to provide a rigorous relationship between a point
process, the prior model, and the more usual mathematical properties of surfaces.
Using range imagery in this problem makes these relationships direct. By adopting
this philosophy, we can exploit the results of Grimson [Gri83] as well as those of
Brady and Horn [BH83] to improve on the Laplacian.
The Gaussian functional form of V is chosen because it is mathematically convenient for Boltzmann statistics and beca.use it reflects the following shape properties
recommended for grey level images in the literature and is especially important if
Range Image Restoration Using Mean Field Annealing
line processes are to be omitted: Besag [Bes86] notes that lito encourage smooth
variation", V(A) "should be strictly increasing" in the absolute value of its argument and if "occasional abrupt changes" are expected, it should "quickly reach a
maximum" .
Rational functions with shapes similar to our V have been used in recent stochastic
approaches to image processing [GM85]. In Eq. 6, T is a "soft threshold" which
represents our prior knowledge of the probability of various values of \7 2 f (the
Laplacian of the undegraded image). For T large, we imply that high values of
the Laplacian are common - f is highly textured; for small values of T, we imply
that f is generally smooth. We note that for high values of T, the prior term is
insignificant, and the best estimate of the image is simply the data.
We choose the Mean Field Hamiltonian to be
(7)
and find that the optimal ZT approximately minimizes
(8)
both at very high and very low T . We have found experimentally that this approximation to ZT does anneal to a satisfactory restoration. At each temperature, we
use gradient descent to find ZT with the following approximation to the gradient of
<H>:
(9)
and
-b
V(r?) -
, - y'2;(T+T)
e-
.. ?
2( ..
h)
(lO)
.
Differentiating Eq. 8 with this new notation, we find
(11)
Since 6'+11,; is non-zero only when i
8
< H > _8 :J! .
)
+ v = i,
:J!j (1'
2
gj
we have
+L
L
-II
IT'(.
')+11
)
II
and this derivative can be used to find the equilibrium condition.
Algorithm
(12)
597
598
Bilbro and Snyder
1. Initially, we use the high temperature assumption, which eliminates the prior
term entirely, and results in
Z;
T
=g;; for
= 00.
(13)
This will provide the initial estimate of z. Any other estimate quickly converges to g.
2. Given an image z;, form the image ri
(L ? z);, where the ? indicates
convolution.
=
3. Create the image V. = V' (r?)
P
,
..?
= - -----l=
_-.!'L e - :II(T~
~T+T)T+T
.. ) ?
4. Using 12, perform ordinary non-linear minimization of < H > starting from
the current z. The particular strategy followed is not critical. We have
successfully used steepest descent and more sophisticated conjugate gradient [PFTV88] methods. The simpler methods seem adequate fot Gaussian
noise.
5. Update z to the minimizing z found in step 4.
6. Reduce T and go to 2. When T is sufficiently close to 0, the algorithm is
complete.
In step 6 above, T essentially defines the appropriate low-temperature stopping
point. In section 5, we will elaborate on the determination of T and other such
constants.
4
Performance
In this section, we describe the performance of the algorithm as it is applied to
several range images. We will use range images, in which the data is of the form
z = z(z, y).
4.1
(14)
Images With High Levels of Noise
Figure 1 illustrates a range image consisting of three objects, a wedge (upper left),
a cylinder with rounded end and hole (right), and a trapezoidal block viewed from
the top. The noise in this region is measured at (1' = 3units out of a total range of
about 100 units. Unsophisticated smoothing will not estimate second derivatives of
such data without blurring. Following the surface interpolation literature, [Gri83]
[BB83] we use the quadratic variation as the argument of the penalty function for
the prior term to
(15)
and performing the derivative in a manner analogous to Eq. 11 and 12. The
Laplacian of the restoration is shown in Figure 2. Figure 3 shows a cross-section
taken as indicated by the red line on Figure 2.
Fig. 1 Original rallge image
I
n
J~~ l\~lll
Fig. 2 Laplacian of the restored image
4.2
Fig. 3 Cross section
Through Laplacian along
Red Line
Comparison With Existing Techniques
Accurate computation of surface derivatives requires extremely good smoothing of
surface noise, while segmentation requires preservat.ion of edges. One suc.h adapt.ive
smoothing technique,[Ter87] iterative Gaussian smoothing (IGS) has been successfully applied to range imagery. [PB87] Following this strategy, step edges are first
detected, and smoothing is then applied using a small center-weighted kernel. At
edges, an even smaller kernel, called a "molecule", is used to smooth right up to
the edge without blurring the edge. The smoothing is then iterated.
600
Bilbro and Snyder
The results, restoration and Laplacian, of IGS are not nearly as sharp as those
shown in Figure 2.
Determining the Parameters
5
Although the optimization strategy described in section 3 has no hard thresholds,
several parameters exist either explicitly in Eq. 8 or implicitly in the iteration.
Good estimates of these parameters will result in improved performance, faster
convergence, or both. The parameters are:
(1' the standard deviation of the noise
b the relative magnitude of the prior term
11 = T + T the initial temperature and
T the final temperature.
The decrement in T which defines the annealing schedule could also be considered
a parameter. However, we have observed that 10% or less per step is always good
enough.
We find that for depth images of polyhedral scenes, T = 0 so that only one parameter
is problem dependent: (1'. For the more realistic case of images which also contain
curved surfaces, however, see our technical report [BS88], which also describes the
MFA derivation in much more detail.
The standard deviation of the noise must be determined independently for each
problem class. It is straightforward to estimate (1' to within 50%, and we have
observed experimentally that performance of the algorithm is not sensitive to this
order of error.
We can analytically show that annealing occurs in the region T:::::: IL12(1'2 and choose
TJ
2ILI2(1'2. Here, ILI2 is the squared norm of the operator Land ILI2
20 for
the usual Laplacian and ILI2 12.5 for the quadratic variation.
Further analysis shows that b .J2;ILI(1' is a good choice for the coefficient of the
prior term.
=
=
=
=
References
[Bes86]
[BGZ76]
[BH83]
[BJ88]
J. Besag. On the statistical analysis of dirty pictures. Journal of the
Royal Stati6ticCJl Society, B 48(3), 1986.
E. Brezin, J. C. Le Guillon, and J. Zinn-Justin. Field theoretical approach to critical phenomena. In C. Domb and M.S. Green, editors,
PhCJ6e Tran6ition6 and Critical Phenomena, chapter 3, Academic Press,
New York, 1976.
M. Brady and B.K.P Horn. Symm~tric operators for surface interpolation. CVGIP, 22, 1983.
P.J. Besl and R.C. Jain. Segmentation through variable-order surface
fitting. IEEE PAMI, 10(2), 1988.
Range Image Restoration Using Mean Field Annealing
G. Bilbro and W. Snyder. A Linear Time Theory for Recognizing Surface6 in 3-D. Technical Report CCSP-NCSU TR-86/8, Center for Communications and Signal Processing, North Carolina State University,
1986.
[BS88]
G. L. Bilbro and W. E. Snyder. Range Image Re6toration U6ing Mean
Field Annealing. Technical Report NETR 88-19, Available from the
Center for Communications and Signal Processing, North Carolina State
University, 1988.
[CZVJ88] R. Chellappa, Y.-T. Zhou, A. Vaid, and B. K. Jenkins. Image restoration
using a neural network. IEEE Tran6action6 on ASSP, 36(7):1141-1151,
July 1988.
[Gem87] D. Geman. Stochastic model for boundary detection. Vi6ion and Image
Computing, 5(2):61-65, 1987.
[GG84]
D. Geman and S. Geman. Stochastic relaxation, Gibbs Distributions,
and the Bayesian restoration of images. IEEE Tran6action6 on PAMI,
PAMI-6(6):721-741, November 1984.
[GM85]
S. Geman and D.E. McClure. Bayesian image analysis: an application
to single photon emission tomography. Proceeding6 of the American
Stati6tical A uociation, Stati6tical Computing Section, 12-18, 1985.
rGri83]
W.E.L. Grimson. An implementation of computational theory of visual
surface interpolation. CVGIP, 22, 1983.
[GS87]
B. R. Groshong and W. E. Snyder. Range image segmentation for object
recognition. In 18th Pitt6burgh Conference on Modeling and Simulation,
Pittsburgh, PA, April 1987.
[Mar85]
J .L. Marroquin. Probabili6tic Solution to Inver6e Problem6. PhD thesis,
M.LT, Cambridge, MA, September 1985.
[MMP87] J. Marroquin, S. Mitter, and T. Poggio. Probabilistic solution of illposed problems in computational vision. Journal of American Stati6tical
Auociation, 82(397):76-89, March 1987.
T. Ponce and M. Brady. Toward a surface primal sketch. In T. Kanade,
[PB87]
editor, Three Dimen6ional Machine Vi6ion, Kluwer Press, 1987.
[PFTV88] W. H. Press, B. P. Flannery, S. A. Teukolsky, and W. T. Vetterling.
Numerical Recipe6 in C. Cambridge University Press, 1988.
[RS87]
F. Romeo and A. Sangiovanni-Vencentelli. Probabalistic hill climbing
algorithms: properties and applications. In Chapel Hill Conference on
VLSI, Computer Science Press, Chapel Hill, NC, 1987.
[Ter87]
D. Terzopoulos. The role of constraints and discontiuities in visiblesurface reconstruction. In Proc. of 7th International Conf. on AI,
pages 1073-1077, 1987.
[WP85]
G. Wolberg and T. Pavlidis. Restoration of binary images using stochastic relaxation with annealing. Pattern Recognition Letter6, 3(6):375-388,
December 1985.
M. Brady A. Yiulle and H. Asada. Describing surfaces. CVGIP, August
[YA85]
1985.
[BS86]
601
| 153 |@word norm:1 grey:1 simulation:1 carolina:3 tr:1 necessity:1 initial:3 existing:1 current:3 z2:1 must:2 numerical:1 realistic:1 shape:3 analytic:1 update:1 steepest:1 hamiltonian:2 simpler:2 mathematical:1 along:2 direct:1 fitting:1 polyhedral:1 manner:1 expected:1 fez:1 lll:1 increasing:1 becomes:1 begin:1 underlying:1 notation:1 proceeding6:1 minimizes:2 brady:4 fat:1 control:1 unit:2 local:2 interpolation:3 approximately:1 pami:3 emphasis:1 range:15 pavlidis:2 bilbro:6 horn:2 fot:1 unsophisticated:1 block:1 illposed:1 area:1 convenient:1 beca:1 close:1 operator:2 conventional:1 map:1 deterministic:1 center:4 go:1 straightforward:1 starting:1 independently:1 abrupt:1 chapel:2 variation:4 analogous:1 pt:3 pa:1 element:1 recognition:2 trapezoidal:1 geman:4 observed:2 role:1 solved:1 region:2 sangiovanni:1 wher:1 grimson:2 segment:1 blurring:2 textured:1 iit:1 various:1 chapter:1 derivation:1 jain:1 describe:1 chellappa:1 detected:1 whose:1 solve:1 valued:3 say:1 ive:1 besl:1 statistic:5 noisy:2 final:1 sequence:1 net:2 reconstruction:1 neighboring:1 j2:1 convergence:1 optimum:1 converges:1 object:2 measured:1 nearest:1 eq:4 solves:1 wedge:1 drawback:1 stochastic:5 tln:1 mathematically:1 strictly:1 sufficiently:1 considered:1 normal:1 equilibrium:7 omitted:1 proc:1 sensitive:1 create:1 successfully:3 reflects:1 weighted:1 minimization:7 gaussian:3 always:1 rather:1 zhou:1 derived:2 emission:1 romeo:1 ponce:1 indicates:1 besag:2 rigorous:1 dependent:1 stopping:1 vetterling:1 typically:1 entire:1 accept:1 initially:1 vlsi:1 pixel:6 priori:2 restores:1 special:1 smoothing:6 field:13 construct:1 never:1 represents:3 nearly:1 peaked:1 others:1 report:3 simplify:1 resulted:1 consisting:2 cylinder:1 detection:1 highly:1 primal:1 tj:1 hg:1 chain:1 accurate:1 edge:5 encourage:1 poggio:1 theoretical:1 soft:1 modeling:1 restoration:15 ordinary:1 deviation:2 recognizing:1 reported:1 international:1 probabilistic:1 rounded:1 continuously:1 quickly:2 jo:1 imagery:2 squared:1 thesis:1 choose:4 hn:1 conf:1 american:2 derivative:6 potential:1 photon:1 domb:1 summarized:1 north:3 coefficient:1 explicitly:1 view:1 characterizes:1 red:2 minimize:1 variance:2 climbing:1 bayesian:2 iterated:1 emphasizes:1 drive:1 fo:2 suffers:1 reach:1 rational:1 gg84:2 knowledge:3 segmentation:3 schedule:1 sophisticated:1 marroquin:2 appears:1 wesley:1 planar:1 improved:1 april:1 formulation:1 done:1 though:1 until:1 sketch:1 defines:2 indicated:1 contain:1 analytically:1 satisfactory:1 hill:3 complete:1 temperature:9 image:36 recently:1 common:1 functional:2 relating:1 kluwer:1 measurement:2 refer:1 cambridge:2 gibbs:1 ai:1 smoothness:1 surface:12 gj:1 curvature:2 recent:1 certain:2 inequality:1 binary:1 minimum:5 converge:1 recommended:1 signal:3 ii:4 july:1 ing:1 technical:3 smooth:3 academic:1 fie:1 determination:1 calculation:1 cross:2 mcclure:1 adapt:1 faster:1 laplacian:9 mrf:1 emf:6 essentially:1 vision:1 symm:1 iteration:3 kernel:2 adopting:1 ion:1 hfl:1 addition:1 annealing:14 eliminates:1 thing:1 december:1 seem:1 ideal:2 enough:1 architecture:1 reduce:2 penalty:1 york:1 wolberg:2 adequate:1 generally:1 se:1 locally:1 tomography:1 reduced:2 exist:1 gil:1 per:1 discrete:1 snyder:7 four:1 threshold:2 verified:1 ht:1 relaxation:2 convert:1 zinn:1 nnealing:1 entirely:1 followed:1 quadratic:3 occur:1 constraint:1 sharply:1 ri:1 scene:1 ncsu:1 argument:2 min:1 extremely:1 performing:1 relatively:1 march:1 conjugate:1 smaller:1 describes:1 making:1 invariant:1 gradually:2 taken:1 turn:2 describing:1 mechanism:1 end:1 available:1 jenkins:1 occasional:1 appropriate:1 netr:1 ho:8 original:1 top:1 running:1 dirty:1 exploit:1 cvgip:3 perturb:1 especially:1 approximating:1 society:1 objective:1 already:1 occurs:1 restored:1 strategy:7 usual:2 september:1 gradient:4 perturbs:1 simulated:2 toward:2 relationship:2 providing:1 minimizing:2 nc:2 unfortunately:1 fe:2 design:1 implementation:1 motivates:1 zt:5 boltzmann:1 perform:2 upper:1 convolution:1 markov:1 descent:3 curved:1 november:1 communication:3 assp:1 arbitrary:1 sharp:1 august:1 cast:1 justin:1 pattern:1 royal:1 green:1 mfa:4 critical:3 improve:1 imply:2 picture:1 prior:9 literature:2 determining:1 relative:1 editor:2 land:1 lo:1 raleigh:1 terzopoulos:1 neighbor:1 differentiating:1 absolute:1 boundary:1 depth:1 ig:2 implicitly:1 global:1 pittsburgh:1 assumed:1 iterative:1 kanade:1 molecule:1 probabalistic:1 excellent:1 anneal:3 suc:1 significance:1 decrement:1 noise:11 fig:3 mitter:1 elaborate:1 slow:1 insignificant:1 ssa:9 dl:3 incorporating:1 importance:1 phd:1 magnitude:1 illustrates:1 hole:1 easier:1 flannery:1 lt:1 simply:2 visual:1 lito:1 scalar:1 determines:1 teukolsky:1 ma:1 viewed:1 eventual:1 experimentally:3 hard:2 change:1 specifically:1 determined:1 asada:1 justify:1 called:2 total:1 ili:1 u6ing:1 philosophy:1 griff:1 phenomenon:2 ex:2 |
580 | 1,530 | A Precise Characterization of the Class of
Languages Recognized by Neural Nets under
Gaussian and other Common Noise Distributions
Wolfgang Maass*
Inst. for Theoretical Computer Science,
Technische Universitat Graz
Klosterwiesgasse 3212,
A-80lO Graz, Austria
email: maass@igi.tu-graz.ac.at
Eduardo D. Sontag
Oep. of Mathematics
Rutgers University
New Brunswick, NJ 08903, USA
email: sontag@hilbert.rutgers.edu
Abstract
We consider recurrent analog neural nets where each gate is subject to
Gaussian noise, or any other common noise distribution whose probability density function is nonzero on a large set. We show that many regular
languages cannot be recognized by networks of this type, for example
the language {w E {O, I} * I w begins with O}, and we give a precise
characterization of those languages which can be recognized. This result
implies severe constraints on possibilities for constructing recurrent analog neural nets that are robust against realistic types of analog noise. On
the other hand we present a method for constructing feedforward analog
neural nets that are robust with regard to analog noise of this type.
1 Introduction
A fairly large literature (see [Omlin, Giles, 1996] and the references therein) is devoted
to the construction of analog neural nets that recognize regular languages. Any physical
realization of the analog computational units of an analog neural net in technological or
biological systems is bound to encounter some form of "imprecision" or analog noise at
its analog computational units. We show in this article that this effect has serious consequences for the computational power of recurrent analog neural nets. We show that any
analog neural net whose computational units are subject to Gaussian or other common
noise distributions cannot recognize arbitrary regular languages. For example, such analog
neural net cannot recognize the regular language {w E {O, I} * I w begins with O}.
? Partially supported by the Fonds zur F6rderung der wissenschaftlichen Forschnung (FWF), Austria, project P12153.
W Maass and E. D. Sontag
282
A precise characterization of those regular languages which can be recognized by such
analog neural nets is given in Theorem 1.1. In section 3 we introduce a simple technique
for making feedforward neural nets robust with regard to the same types of analog noise.
This method is employed to prove the positive part of Theorem 1.1. The main difficulty in
proving Theorem 1.1 is its negative part, for which adequate theoretical tools are introduced
in section 2.
Before we can give the exact statement of Theorem 1.1 and discuss related preceding work
we have to give a precise definition of computations in noisy neural networks. From the
conceptual point of view this definition is basically the same as for computations in noisy
boolean circuits (see [Pippenger, 1985] and [Pippenger, 1990]). However it is technically
more involved since we have to deal here with an infinite state space.
We will first illustrate this definition for a concrete case, a recurrent sigmoidal neural net
with Gaussian noise, and then indicate the full generality of our result, which makes it
applicable to a very large class of other types of analog computational systems with analog
noise. Consider a recurrent sigmoidal neural net N consisting of n units, that receives
at each time step t an input Ut from some finite alphabet U (for example U = {O, I}).
The internal state of N at the end of step t is described by a vector Xt E [-1, l]n, which
consists of the outputs of the n sigmoidal units at the end of step t. A computation step of
the network N is described by
Xt+1
=
a(Wxt
+ h + UtC + Vi)
where W E IRnxn and c, h E IRn represent weight matrix and vectors, a is a sigmoidal
activation function (e.g., a(y) = 1/(1 + e- Y applied to each vector component, and
VI, V2 , ? ?? is a sequence of n-vectors drawn independently from some Gaussian distribution. In analogy to the case of noisy boolean circuits [Pippenger, 1990] one says that this
network N recognizes a language L ~ U* with reliability c (where c E (O,~] is some
given constant) if immediately after reading an arbitrary word w E U* the network N is
with probability 2: ~ + c in an accepting state in case that w E L, and with probability
c in an accepting state in case that w rf. LI.
?
: :; ! -
We will show in this article that even if the parameters of the Gaussian noise distribution for
each sigmoidal unit can be determined by the designer of the neural net, it is impossible to
find a size n, weight matrix W, vectors h, C and a reliability c E (0, so that the resulting
recurrent sigmoidal neural net with Gaussian noise accepts the simple regular language
{w E {0,1}*1 w begins with O} with reliability c. This result exhibits a fundamental
limitation for making a recurrent analog neural net noise robust, even in a case where the
noise distribution is known and of a rather benign type. This quite startling negative result
should be contrasted with the large number of known techniques for making a feedforward
boolean circuit robust against noise, see [Pippenger, 1990].
!]
Our negative result turns out to be of a very general nature, that holds for virtually all related
definitions of noisy analog neural nets and also for completely different models for analog
computation in the presence of Gaussian or similar noise. Instead of the state set [-1, l]n
one can take any compact set n ~ IRn , and instead of the map (x, u) t-+ W x + h + uc one
can consider an arbitrary map I : n x U ~ 0 for a compact set 0 ~ IRn where f (', u) is
Borel measurable for each fixed U E U. Instead of a sigmoidal activation function a and a
Gaussian distributed noise vector V it suffices to assume that a : IRn ~ n is some arbitrary
Borel measurable function and V is some IRn -valued random variable with a density ?(.)
that has a wide support2 ? In order to define a computation in such system we consider for
1 According to this definition a network N that is after reading some w E U? in an accepting state
with probability strictly between
c and + c does not recognize any language L ~ U?.
2More precisely: We assume that there exists a subset no of n and some constant Co > 0 such
t-
t
283
Analog Neural Nets with Gaussian Noise
each U E U the stochastic kernel Ku defined by Ku(x, A) := Prob [a(f(x, u) + V) E A]
for x E n and A S;;; n. For each (signed, Borel) measure /-l on n, and each U E U, we
let lKu/-l be the (signed, Borel) measure defined on n by (lKu/-l)(A) := Ku(x , A)d/-l(x) .
Note that lKu /-l is a probability measure whenever /-l is. For any sequence of inputs W =
U1 , .?. ,U r , we consider the composition of the evolution operators lKu. :
J
(1)
If the probability distribution of states at any given instant is given by the measure /-l, then
the distribution of states after a single computation step on input U E U is given by lKu /-l,
and after r computation steps on inputs W = U1,"" U r , the new distribution is IKw /-l,
where we are using the notation (1). In particular, if the system starts at a particular initial
state ~, then the distribution of states after r computation steps on W is IKwbe, where be is
the probability measure concentrated on {O. That is to say, for each measurable subset
Fen
Prob [Xr+1 E F
I Xl
=~, input =
w] = (lKwbe)(F).
We fix an initial state ~ E n, a set of "accepting" or "final" states F, and a "reliability"
level E > 0, and say that the resulting noisy analog computational system M recognizes
the language L S;;; U* if for all w E U* :
1
wEL ~ (lKwbe)(F) ;::: 2 + E
w(j.L
(lKwbe)(F) :::;
1
2-
E .
In general a neural network that simulates a DFA will carry out not just one, but a fixed
number k of computation steps (=state transitions) of the form x' = a(W x + h +
uc + V) for each input symbol U E U that it reads (see the constructions described in
[Omlin, Giles, 1996], and in section 3 of this article). This can easily be reflected in our
model by formally replacing any input sequence w = UI, U2 , . . . , U r from U* by a padded
sequence W = UI , bk - I , U2 , bk - I , ... ,U r , bk - I from (U U {b})*, where b is a blank symbol not in U, and bk - I denotes a sequence of k - 1 copies of b (for some arbitrarily fixed
k ;::: 1). This completes our definition of language recognition by a noisy analog computational system M with discrete time. This definition essentially agrees with that given in
[Maass, Orponen, 1997].
We employ the following common notations from formal language theory: We write WI w2
for the concatenation of two strings WI and W2,
for the set of all concatenations of r
strings from U, U* for the set of all concatenations of any finite number of strings from U,
and UV for the set of all strings WI W2 with WI E U and W2 E V . The main result of this
article is the following:
ur
Theorem 1.1 Assume that U is some arbitrary finite alphabet. A language L S;;; U* can
be recognized by a noisy analog computational system of the previously specified type if
and only if L E1 UU* E2 for two finite subsets E1 and E2 of U* .
=
A corresponding version of Theorem 1.1 for discrete computational systems was previously
shown in [Rabin, 1963]. More precisely, Rabin had shown that probabilistic automata with
strictly positive matrices can recognize exactly the same class of languages L that occur
in our Theorem 1.1. Rabin referred to these languages as definite languages. Language
recognition by analog computational systems with analog noise has previously been investigated in [Casey, 1996] for the special case of bounded noise and perfect reliability
n
that the following two properties hold: <jJ(v) :::: Co for all v E Q := a-I (no) (that is, Q is the
set consisting of all possible differences z - y, with a(z) E no and yEn) and a-I (no) has finite
and nonzero Lebesgue measuremo =). (a- 1 (no)).
W Maass and E. D. Sontag
284
(i.e. ~ l vll:S;71 ?>(v)dv = 1 for some small TJ > 0 and c = 1/2 in our terminology), and in
[Maass, Orponen, 1997] for the general case. It was shown in [Maass, Orponen, 1997] that
any such system can only recognize regular languages. Furthermore it was shown there
that if ~lvll:S;71 ?>(v)dv = 1 for some small TJ > 0 then all regular languages can be recognized by such systems. In the present paper we focus on the complementary case where the
condition "~lvll:S;71 ?>(v)dv = 1 for some small '" > 0" is not satisfied, i.e. analog noise may
move states over larger distances in the state space. We show that even if the probability of
such event is arbitrarily small, the neural net will no longer be able to recognize arbitrary
regular languages.
2 A Constraint on Language Recognition
We prove in this section the following result for arbitrary noisy computational systems M
as specified at the end of section 1:
Theorem 2.1 Assume that U is some arbitrary alphabet. If a language L ~ U* is recognized by M, then there are subsets E1 and E2 of u:S;r, for some integer r, such that
L = E1 U U* E 2. In other words: whether a string w E U* belongs to the language L
can be decided by just inspecting the first r and the last r symbols of w.
2.1
A General Fact about Stochastic Kernels
Let (5, S) be a measure space, and let K be a stochastic kernel 3 . As in the special case of
the Ku's above, for each (signed) measure f-t on (5, S), we let II<?t be the (signed) measure
defined on S by (II<?t)(A) := J K(x, A)df-t(x) . Observe that II<?t is a probability measure
whenever f-t is. Let c > 0 be arbitrary. We say that K satisfies Doeblin's condition (with
constant c) if there is some probability measure p on (5, S) so that
K(x, A) ~ cp(A) for all x E 5, A E S.
(2)
(Necessarily c ::; 1, as is seen by considering the special case A = 5.) This condition is
due to [Doeblin, 1937].
We denote by Ilf-til the total variation of the (signed) measure f-t. Recall that Ilf-til is defined as follows. One may decompose 5 into a disjoint union of two sets A and B, in
such a manner that f-t is nonnegative on A and nonpositive on B. Letting the restrictions
of f-t to A and B be "f-t+" and "-f-t-" respectively (and zero on B and A respectively),
we may decompose f-t as a difference of nonnegative measures with disjoint supports,
f-t = f-t+ - f-t- . Then, Ilf-til = f-t+ (A) + f-t- (B). The following Lemma is a "folk"
fact ([Papinicolaou, 1978]).
Lemma 2.2 Assume that K satisfies Doeblin's condition with constant c. Let f-t be any
(signed) measure such that f-t(5) = o. Then 111I<?t11 ::; (1 - c) 1If-t1l.
?
2.2
Proof of Theorem 2.1
Lemma 2.3 There is a constant c
constantc,foreveryu E U.
> 0 such
that Ku satisfies Doeblin's condition with
Proof Let no, co, and 0 < rno < 1 be as in the second footnote, and introduce the
following (Borel) probability measure on no:
Ao(A) :=
~A (0'-1 (A))
rno
.
3That is to say, K(x,?) is a probability distribution for each x, and K(-, A) is a measurable
function for each Borel measurable set A.
285
Analog Neural Nets with Gaussian Noise
Pick any measurable A ~ no and any yEn. Then,
Z(y, A)
Prob [O"(y
=
+ V)
(?(v) dv
JAy
~
E A] = Prob [y
+V
E 0"-1 (A)]
coA(Ay) = CoA (0"-1 (A?) = comoAo(A) ,
where Ay := 0"-1 (A) - {y} ~ Q. We conclude that Z(y, A) ~ cAo(A) for all y, A,
where c = como. Finally, we extend the measure AO to all of n by assigning zero measure
to the complement of no, that is, p(A) := Ao(A no) for all measurable subsets A of n .
Pick u E U; we will show that Ku satisfies Doeblin's condition with the above constant
c (and using p as the "comparison" measure in the definition). Consider any x E nand
measurable A ~ n. Then,
n
Ku(x, A)
= Z(f(x, u), A)
~ Z(f(x, u), A
n
no) ~ cAo(A
n
no)
= cp(A) ,
?
as required.
For every two probability measures 1-'1,1-'2 on n, applying Lemma 2.2 to I-' := 1-'1 -1-'2, we
know that 111Ku1-'1 -1Ku1-'211 ::; (1 - c) 111-'1 - 1-'211 for each u E U . Recursively, then, we
conclude:
(3)
IllKwl-'l -lKw1-'211 ::; (1 111-'1 - 1-'211::; 2(1 -
ct
for all words w of length
~
ct
r.
Now pick any integer r such that (1 -
ct < 2c. From Equation (3), we have that
for all w of length ~ r and any two probability measures 1-'1,1-'2. In particular, this means
that, for each measurable set A,
(4)
for all such w. (Because, for any two probability measures
set A, 2Ivl(A) - v2(A)1 ::; Ilvl - v211 ?)
VI
and
V2,
and any measurable
Lemma 2.4 Pick any v E U* and wE ur. Then
w E L {:::::::} vw E L .
Proof Assume that w E L, that is, (lKw t5e)(F) ~ ~+E. Applying inequality (4) to the measures 1-'1 := t5e and 1-'2 := lKvt5e and A = F, we have that l(lKwt5e)(F) - (lKvw t5e)(F) I <
2E,andthisimpliesthat(lKvwt5e)(F) > ~-E,i.e. , vw E L. (Since ~-E < (lKvwt5e)(F) <
~ + E is ruled out.) If w ~ L, the argument is similar.
?
We have proved that
So,
n
where El := L u~r and E2 := L
proof of Theorem 2.1.
nur are both included in u~r. This completes the
?
286
3
W. Maass and E. D. Sontag
Construction of Noise Robust Analog Neural Nets
In this section we exhibit a method for making feedforward analog neural nets robust with
regard to arbitrary analog noise of the type considered in the preceding sections. This
method will be used to prove in Corollary 3.2 the missing positive part of the claim of the
main result (Theorem 1.1) of this article.
Theorem 3.1 Let C be any (noiseless) feedfOlward threshold circuit, and let u : ~ -+
[-1, 1] be some arbitrary function with u( u) -+ 1 for u -+ 00 and u( u) -+ -1 for
u -+ -00. Furthermore assume that 8, p E (0, 1) are some arbitrary given parameters.
Then one can transform for any given analog noise of the type considered in section 1 the
noiseless threshold circuit C into an analog neural net Nc with the same number of gates,
whose gates employ the given function u as activation function, so that for any circuit input
~ E {-I, l}m the output of the noisy analog neural net Nc differs with probability ~ 1- 8
by at most p from the output ofC.
Idea of the proof Let k be the maximal fan-in of a gate in C, and let w be the maximal
absolute value of a weight in C. We choose R > so large that the density function ?>(.) of
the noise vector V satisfies for each gate with n inputs in C
?
{
JIVil~R
?>(v)dv:c:;
?
28 for i= 1, ... ,n.
n
Furthermore we choose Uo > so large that u(u) ~ 1 - p/(wk) for u ~ Uo and u(u) :c:;
-1 + p/(wk) for u :c:; -Uo . Finally we choose a factor "/ > so large that ,,/(1- p) - R ~
Uo. LetNc be the analog neural net that results from C through multiplication of all weights
and thresholds with "/ and through replacement of the Heaviside activation functions of the
gates in C by the given activation function u.
?
?
The following Corollary provides the proof of the positive part of our main result Theorem
1.1. It holds for any u considered in Theorem 3.1.
Corollary 3.2 Assume that U is some arbitrary finite alphabet, and language L ~ U* is
of the form L = El U U* E2 for two arbitrary finite subsets El and E2 of U*. Then the
language L can be recognized by a noisy analog neural net N with any desired reliability
E E (0, ~), in spite of arbitrary analog noise of the type considered in section 1.
Proof. We first construct a feed forward threshold circuit C for recognizing L, that receives
each input symbol from U in the form of a bitstring u E {a, 1}' (for some fixed I ~
log2 \U\), that is encoded as the binary states of l input units of the boolean circuit C. Via
a tapped delay line of fixed length d (which can easily be implemented in a feedforward
threshold circuit by d layers, each consisting of l gates that compute the identity function on
a single binary input from the preceding layer) one can achieve that the feed forward circuit
C computes any given boolean function of the last d sequences from {O, 1}1 that were
presented to the circuit. On the other hand for any language of the form L = El U U* E2
with E 1 , E2 finite there exists some dEN such that for each w E U* one can decide
whether w E L by just inspecting the last d characters of w. Therefore a feedforward
threshold circuit C with a tapped delay line of the type described above can decide whether
wE L.
=
t).
We apply Theorem 3.1 to this circuit C for 8
p = min(~ - E,
We define the set F
of accepting states for the resulting analog neural net Nc as the set of those states where
the computation is completed and the output gate of Nc assumes a value ~ 3/4. Then
according to Theorem 3.1 the analog neural net Nc recognizes L with reliability E. To be
formally precise, one has to apply Theorem 3.1 to a threshold circuit C that receives its
Analog Neural Nets with Gaussian NOise
287
input not in a single batch, but through a sequence of d batches. The proof of Theorem 3.1
readily extends to this case.
?
4
Conclusions
We have exhibited a fundamental limitation of analog neural nets with Gaussian or other
common noise distributions whose probability density function is nonzero on a large set:
They cannot accept the very simple regular language {w E {O, 1 }*I w begins with O}.
This holds even if the designer of the neural net is allowed to choose the parameters of
the Gaussian noise distribution and the architecture and parameters of the neural net. The
proof of this result introduces new mathematical arguments into the investigation of neural
computation, which can also be applied to other stochastic analog computational systems.
We also have presented a method for makingfeedfOlward analog neural nets robust against
the same type of noise. This implies that certain regular languages, such as for example
{w E {O, 1 }*I wends with O} can be recognized by a recurrent analog neural net with
Gaussian noise. In combination with our negative result this yields a precise characterization of all regular languages that can be recognized by recurrent analog neural nets with
Gaussian noise, or with any other noise distribution that has a large support.
References
[Casey, 1996] Casey, M., "The dynamics of discrete-time computation, with application to
recurrent neural networks and finite state machine extraction", Neural Computation
8,1135-1178,1996.
[Doeblin, 1937] Doeblin, W., "Sur Ie proprietes asymtotiques de mouvement regis par
certain types de chaInes simples", Bull. Math. Soc. Roumaine Sci. 39(1): 57-115; (2)
3-61,1937.
[Maass, Orponen, 1997] Maass, W., and Orponen, P. "On the effect of analog noise on
discrete-time analog computations", Advances in Neural Information Processing Systems 9, 1997, 218-224; journal version: Neural Computation 10(5), 1071-1095,
1998.
[amlin, Giles, 1996] amlin, C. W., Giles, C. L. "Constructing deterministic finite-state
automata in recurrent neural networks", J. Assoc. Comput. Mach. 43 (1996), 937972.
[Papinicolaou, 1978] Papinicolaou, G., "Asymptotic Analysis of Stochastic Equations", in
Studies in Probability Theory, MAA Studies in Mathematics, vol. 18, 111-179, edited
by M. Rosenblatt, Math. Assoc. of America, 1978.
[Pippenger, 1985] Pippenger, N., "On networks of noisy gates", IEEE Sympos. on Foundations of Computer Science, vol. 26, IEEE Press, New York, 30-38, 1985.
[Pippenger, 1989] Pippenger, N., ':Invariance of complexity measures for networks with
unreliable gates", J. of the ACM, vol. 36, 531-539,1989.
[Pippenger, 1990] Pippenger, N., "Developments in 'The Synthesis of Reliable Organisms
from Unreliable Components' ", Proc. of Symposia in Pure Mathematics, vol. 50,
311-324,1990.
[Rabin, 1963] Rabin, M., "Probabilistic automata", Information and Control, vol. 6, 230245, 1963.
| 1530 |@word version:2 rno:2 pick:4 recursively:1 doeblin:7 carry:1 initial:2 irnxn:1 orponen:5 blank:1 activation:5 assigning:1 readily:1 realistic:1 benign:1 nur:1 accepting:5 v211:1 characterization:4 provides:1 math:2 sigmoidal:7 mathematical:1 lku:5 symposium:1 prove:3 consists:1 manner:1 introduce:2 coa:2 utc:1 considering:1 begin:4 project:1 notation:2 bounded:1 circuit:14 string:5 nj:1 eduardo:1 every:1 exactly:1 assoc:2 control:1 unit:7 uo:4 positive:4 before:1 consequence:1 mach:1 signed:6 therein:1 co:3 decided:1 union:1 definite:1 differs:1 xr:1 word:3 regular:12 spite:1 cannot:4 operator:1 impossible:1 applying:2 ilf:3 restriction:1 measurable:10 deterministic:1 map:2 missing:1 independently:1 automaton:3 immediately:1 pure:1 proving:1 variation:1 vll:1 construction:3 exact:1 tapped:2 recognition:3 graz:3 technological:1 edited:1 ui:2 complexity:1 dynamic:1 support2:1 technically:1 klosterwiesgasse:1 completely:1 easily:2 america:1 alphabet:4 sympos:1 whose:4 quite:1 larger:1 valued:1 encoded:1 say:5 transform:1 noisy:11 final:1 sequence:7 net:35 maximal:2 tu:1 cao:2 realization:1 achieve:1 perfect:1 illustrate:1 recurrent:11 ac:1 soc:1 implemented:1 implies:2 indicate:1 uu:1 stochastic:5 suffices:1 fix:1 ao:3 decompose:2 investigation:1 biological:1 inspecting:2 strictly:2 hold:4 considered:4 claim:1 proc:1 applicable:1 bitstring:1 agrees:1 ofc:1 tool:1 gaussian:16 rather:1 corollary:3 focus:1 casey:3 ivl:1 regis:1 inst:1 el:4 nand:1 accept:1 irn:5 development:1 special:3 fairly:1 uc:2 construct:1 extraction:1 serious:1 employ:2 recognize:7 consisting:3 lebesgue:1 replacement:1 possibility:1 severe:1 introduces:1 tj:2 devoted:1 folk:1 ruled:1 desired:1 theoretical:2 rabin:5 giles:4 boolean:5 bull:1 technische:1 subset:6 recognizing:1 delay:2 universitat:1 density:4 fundamental:2 ie:1 probabilistic:2 synthesis:1 concrete:1 satisfied:1 choose:4 til:3 li:1 ku1:2 de:2 wk:2 igi:1 vi:3 view:1 wolfgang:1 start:1 yen:2 yield:1 basically:1 startling:1 footnote:1 whenever:2 email:2 definition:8 against:3 involved:1 e2:8 proof:9 nonpositive:1 proved:1 austria:2 recall:1 ut:1 hilbert:1 feed:2 reflected:1 generality:1 furthermore:3 just:3 hand:2 receives:3 replacing:1 usa:1 effect:2 evolution:1 imprecision:1 read:1 nonzero:3 maass:10 deal:1 ay:2 cp:2 common:5 physical:1 analog:47 extend:1 organism:1 composition:1 uv:1 mathematics:3 language:31 had:1 reliability:7 longer:1 wxt:1 belongs:1 certain:2 inequality:1 binary:2 arbitrarily:2 der:1 fen:1 seen:1 preceding:3 employed:1 recognized:10 ii:3 full:1 e1:4 essentially:1 noiseless:2 rutgers:2 df:1 represent:1 kernel:3 zur:1 completes:2 w2:4 exhibited:1 subject:2 virtually:1 simulates:1 fwf:1 integer:2 vw:2 presence:1 feedforward:6 architecture:1 idea:1 whether:3 sontag:5 york:1 wissenschaftlichen:1 jj:1 adequate:1 dfa:1 concentrated:1 designer:2 disjoint:2 rosenblatt:1 discrete:4 write:1 vol:5 terminology:1 threshold:7 amlin:2 drawn:1 padded:1 prob:4 extends:1 decide:2 bound:1 wel:1 ct:3 layer:2 fan:1 nonnegative:2 occur:1 constraint:2 precisely:2 u1:2 lkw:1 argument:2 min:1 according:2 combination:1 character:1 ur:2 wi:4 making:4 dv:5 den:1 equation:2 previously:3 discus:1 turn:1 know:1 letting:1 end:3 apply:2 observe:1 v2:3 batch:2 encounter:1 gate:10 denotes:1 assumes:1 t11:1 recognizes:3 completed:1 log2:1 instant:1 ikw:1 move:1 exhibit:2 distance:1 t1l:1 sci:1 concatenation:3 length:3 sur:1 nc:5 statement:1 negative:4 finite:10 precise:6 arbitrary:15 introduced:1 bk:4 complement:1 required:1 specified:2 accepts:1 able:1 reading:2 rf:1 reliable:1 power:1 omlin:2 event:1 difficulty:1 literature:1 multiplication:1 asymptotic:1 par:1 limitation:2 analogy:1 foundation:1 article:5 lo:1 supported:1 last:3 copy:1 formal:1 wide:1 absolute:1 distributed:1 regard:3 transition:1 computes:1 forward:2 compact:2 unreliable:2 conceptual:1 conclude:2 nature:1 ku:7 robust:8 investigated:1 necessarily:1 constructing:3 main:4 noise:35 allowed:1 complementary:1 referred:1 borel:6 xl:1 comput:1 jay:1 theorem:18 xt:2 symbol:4 exists:2 fonds:1 partially:1 u2:2 maa:1 satisfies:5 acm:1 identity:1 pippenger:10 included:1 infinite:1 determined:1 contrasted:1 lemma:5 total:1 invariance:1 formally:2 internal:1 support:2 brunswick:1 heaviside:1 |
581 | 1,531 | Finite-Sample Convergence Rates for
Q-Learning and Indirect Algorithms
Michael Kearns and Satinder Singh
AT&T Labs
180 Park Avenue
Florham Park , NJ 07932
{mkearns,baveja}@research.att.com
Abstract
In this paper, we address two issues of long-standing interest in the reinforcement learning literature. First, what kinds of performance guarantees can be made for Q-learning after only a finite number of actions?
Second, what quantitative comparisons can be made between Q-learning
and model-based (indirect) approaches, which use experience to estimate
next-state distributions for off-line value iteration?
We first show that both Q-learning and the indirect approach enjoy
rather rapid convergence to the optimal policy as a function of the number of state transitions observed. In particular, on the order of only
(Nlog(1/c)/c 2 )(log(N) + loglog(l/c)) transitions are sufficient for both
algorithms to come within c of the optimal policy, in an idealized model
that assumes the observed transitions are "well-mixed" throughout an
N-state MDP. Thus, the two approaches have roughly the same sample
complexity. Perhaps surprisingly, this sample complexity is far less than
what is required for the model-based approach to actually construct a good
approximation to the next-state distribution. The result also shows that
the amount of memory required by the model-based approach is closer to
N than to N 2 ?
For either approach, to remove the assumption that the observed transitions are well-mixed, we consider a model in which the transitions are
determined by a fixed, arbitrary exploration policy. Bounds on the number
of transitions required in order to achieve a desired level of performance
are then related to the stationary distribution and mixing time of this
policy.
1
Introduction
There are at least two different approaches to learning in Markov decision processes:
indirect approaches, which use control experience (observed transitions and payoffs)
to estimate a model, and then apply dynamic programming to compute policies from
the estimated model; and direct approaches such as Q-Iearning [2], which use control
Convergence Rates for Q-Leaming and Indirect Algorithms
997
experience to directly learn policies (through value functions) without ever explicitly
estimating a model. Both are known to converge asymptotically to the optimal policy [1, 3] . However, little is known about the performance of these two approaches
after only a finite amount of experience .
A common argument offered by proponents of direct methods is that it may require
much more experience to learn an accurate model than to simply learn a good policy.
This argument is predicated on the seemingly reasonable assumption that an indirect
method must first learn an accurate model in order to compute a good policy. On
the other hand, proponents of indirect methods argue that such methods can do
unlimited off-line computation on the estimated model, which may give an advantage
over direct methods, at least if the model is accurate. Learning a good model may
also be useful across tasks, permitting the computation of good policies for multiple
reward functions [4]. To date, these arguments have lacked a formal framework for
analysis and verification.
In this paper, we provide such a framework, and use it to derive the first finite-time
convergence rates (sample size bounds) for both Q-learning and the standard indirect
algorithm. An important aspect of our analysis is that we separate the quality of the
policy generating experience from the quality of the two learning algorithms. In
addition to demonstrating that both methods enjoy rather rapid convergence to the
optimal policy as a function of the amount of control experience, the convergence rates
have a number of specific and perhaps surprising implications for the hypothetical
differences between the two approaches outlined above. Some of these implications,
as well as the rates of convergence we derive, were briefly mentioned in the abstract;
in the interests of brevity, we will not repeat them here, but instead proceed directly
into the technical material.
2
MDP Basics
Let M be an unknown N-state MDP with A actions . We use PM(ij) to denote the
probability of going to state j, given that we are in state i and execute action a;
and RM(i) to denote the reward received for executing a from i (which we assume is
fixed and bounded between 0 and 1 without loss of generality). A policy 1r assigns
an action to each state. The value of state i under policy 1r, VM(i), is the expected
discounted sum of rewards received upon starting in state i and executing 1r forever :
VM(i) = E7r[rl + ,r2 + ,2r3 + ...], where rt is the reward received at time step t
under a random walk governed by 1r from start state i, and 0 ~ , < 1 is the discount
factor . It is also convenient to define values for state-action pairs (i, a): QM(i, a) =
RM (i)
Lj PM (ij) VM(j) . The goal of learning is to approximate the optimal policy
1r* that maximizes the value at every state; the optimal value function is denoted QM.
Given QM' we can compute the optimal policy as 1r*(i) = argmaxa{QM(i,a)}.
+,
If M is given , value iteration can be used to compute a good approximation to the
optimal value function. Setting our initial guess as Qo(i, a) = 0 for all (i, a), we
iterate as follows:
RM(i)
+, 2)PM(ij)Ve(j)]
(1)
j
where we define \Il(j) = maxv{Qe(j, b)}. It can be shown that after I! iterations,
max(i,aj{IQe(i, a) - QM(i , a)1} ~
Given any approximation Q to QM we can compute the greedy approximation 1r to the optimal policy 1r* as 1r(i) = argmaxa{Q(i, a)}.
,e.
998
3
M Kearns and S. Singh
The Parallel Sampling Model
In reinforcement learning, the transition probabilities PM(ij) are not given, and a
good policy must be learn ed on the basis of observed experience (transitions) in M .
Classical convergence results for algorithms such as Q-Iearning [1] implicitly assume
that the observed experience is generated by an arbitrary "exploration policy" 7r, and
then proceed to prove convergence to the optimal policy if 7r meets certain minimal conditions - namely, 7r must try every state-action pair infinitely often, with
probability 1. This approach conflates two distinct issues : the quality of the exploration policy 7r, and the quality ofreinforcement learning algorithms using experience
generated by 7r. In contrast, we choose to separate these issues. If the exploration
policy never or only very rarely visits some state-action pair, we would like to have
this reflected as a factor in our bounds that depends only on 7r; a separate factor
depending only on the learning algorithm will in turn reflect how efficiently a particular learning algorithm uses the experience generated by 7r . Thus, for a fixed 7r, all
learning algorithms are placed on equal footing, and can be directly compared.
There are probably various ways in which this separation can be accomplished; we
now introduce one that is particularly clean and simple. We would like a model of
the ideal exploration policy - one that produces experiences that are "well-mixed",
in the sense that every state-action pair is tried with equal frequency. Thus, let us
define a parallel sampling subroutine PS(M) that behaves as follows: a single call to
PS( M) returns, for every state-action pair (i, a), a random next state j distributed
according to PM (ij). Thus, every state-action pair is executed simultaneously, and
the resulting N x A next states are reported. A single call to PS(M) is therefore really
simulating N x A transitions in M, and we must be careful to multiply the number
of calls to PS(M) by this factor if we wish to count the total number of transitions
witnessed.
What is PS(M) modeling? It is modeling the idealized exploration policy that manages to visit every state-action pair in succession, without duplication, and without
fail. It should be intuitively obvious that such an exploration policy would be optimal,
from the viewpoint of gathering experience everywhere as rapidly as possible.
We shall first provide an analysis, in Section 5, of both direct and indirect reinforcement learning algorithms, in a setting in which the observed experience is generated
by calls to PS(M). Of course, in any given MDP M , there may not be any exploration
policy that meets the ideal captured by PS(M) - for instance, there may simply be
some states that are very difficult for any policy to reach, and thus the experience
generated by any policy will certainly not be equally mixed around the entire MDP.
(Indeed, a call to PS(M) will typically return a set of transitions that does not even
correspond to a trajectory in M.) Furthermore, even if PS(M) could be simulated
by some exploration policy, we would like to provide more general results that express the amount of experience required for reinforcement learning algorithms under
any exploration policy (where the amount of experience will , of course, depend on
properties of the exploration policy).
Thus, in Section 6, we sketch how one can bound the amount of experience required
under any 7r in order to simulate calls to PS(M) . (More detail will be provided in a
longer version of this paper.) The bound depends on natural properties of 7r, such as
its stationary distribution and mixing time. Combined with the results of Section 5,
we get the desired two-factor bounds discussed above: for both the direct and indirect
approaches, a bound on the total number of transitions required, consisting of one
factor that depends only on the algorithm, and another factor that depends only on
the exploration policy.
Convergence Rates for Q-Learning and Indirect Algorithms
4
999
The Learning Algorithms
We now explicitly state the two reinforcement learning algorithms we shall analyze
and compare. In keeping with the separation between algorithms and exploration
policies already discussed, we will phrase these algorithms in the parallel sampling
framework, and Section 6 indicates how they generalize to the case of arbitrary exploration policies. We begin with the direct approach.
Rather than directly studying standard Q-Iearning, we will here instead examine a
variant that is slightly easier to analyze, and is called phased Q-Iearning. However, we
emphasize that all of our resuits can be generalized to apply to standard Q-learning
(with learning rate a(i, a) = t(i~a)' where t(i, a) is the number oftrials of (i, a) so far) .
Basically, rather than updating the value function with every observed transition from
(i , a), phased Q-Iearning estimates the expected value of the next state from (i, a)
on the basis of many transitions, and only then makes an update. The memory
requirements for phased Q-learning are essentially the same as those for standard
Q-Iearning.
Direct Algorithm - Phased Q-Learning: As suggested by the name , the algorithm operates in phases. In each phase, the algorithm will make mD calls to PS(M)
(where mD will be determined by the analysis), thus gathering mD trials of every
state-action pair (i, a) . At the fth phase, the algorithm updates the estimated value
function as follows: for every (i , a),
Ql+d i , a)
= RM(i) + ,_1_ ~ Oeu?)
(2)
mD k=l
where jf, ... , j~ are the m D next states observed from (i, a) on the m D calls to
PS(M) during t~e fth phase. The policy computed by the algorithm is then the
greedy policy determined by the final value function. Note that phased Q-learning
is quite like standard Q-Iearning, except that we gather statistics (the summation in
Equation (2)) before making an update.
We now proceed to describe the standard indirect approach .
Indirect Algorithm: The algorithm first makes m[ calls to PS(M) to obtain m[
next state samples for each (i, a) . It then builds an empirical model of the transition
probabilities as follows: PM(ij) = #(~aj) , where #(i -+a j) is the number of times
state j was reached on the m[ trials of (i, a). The algorithm then does value iteration
(as described in Section 2) on the fixed model PM(ij) for f[ phases. Again , the policy
computed by the algorithm is the greedy policy dictated by the final value function .
Thus , in phased Q-Iearning, the algorithm runs for some number fD phases, and each
phase requires mD calls to PS(M), for a total number of transitions fD x mD x N x A .
The direct algorithm first makes m j calls to PS(M) , and then runs f[ phases of
value iteration (which requires no additional data) , for a total number of transitions
m[ x N x A. The question we now address is: how large must mD, m[, fD' f[ be
so that, with probability at least 1 - 6, the resulting policies have expected return
within f. of the optimal policy in M? The answers we give yield perhaps surprisingly
similar bounds on the total number of transitions required for the two approaches in
the parallel sampling model.
5
Bounds on the Number of Transitions
We now state our main result.
M Kearns and S. Singh
1000
Theorem 1 For any MDP M:
? For an appropriate choice of the parameters mJ and and fJ, the total number
of calls to PS(M) required by the indirect algorithm in order to ensure that,
with probability at least 1 - 6, the expected return of the resulting policy will
be within f of the optimal policy, is
O((I/f 2 )(log(N/6)
+ loglog(l/f)).
(3)
? For an appropriate choice of the parameters mD and fD, the total number of
calls to PS(M) required by phased Q-learning in order to ensure that, with
probability at least 1 - 6, the expected return of the resulting policy will be
within f of the optimal policy, is
O((log(1/f)/f 2 )(log(N/6)
+ log log(l/f)).
(4)
The bound for phased Q-learning is thus only O(log(l/f)) larger than that for the
indirect algorithm. Bounds on the total number of transitions witnessed in either
case are obtained by multiplying the given bounds by N x A .
Before sketching some of the ideas behind the proof of this result, we first discuss
some of its implications for the debate on direct versus indirect approaches. First of
all, for both approaches, convergence is rather fast: with a total number of transitions
only on the order of N log(N) (fixing f and 6 for simplicity), near-optimal policies
are obtained. This represents a considerable advance over the classical asymptotic
results: instead of saying that an infinite number of visits to every state-action pair
are required to converge to the optimal policy, we are claiming that a rather small
number of visits are required to get close to the optimal policy. Second, by our
analysis, the two approaches have similar complexities, with the number of transitions
required differing by only a log(l/f) factor in favor of the indirect algorithm. Third
- and perhaps surprisingly - note that since only O(log(N)) calls are being made
to PS(M) (again fixing f and 6), and since the number of trials per state-action pair
is exactly the number of calls to PS(M), the total number of non-zero entries in the
model PM (ij) built by the indirect approach is in fact only O(log( N)). In other
words , PM (ij) will be extremely sparse - and thus, a terrible approximation to the
true transition probabilities - yet still good enough to derive a near-optimal policy!
Clever representation of PM(ij) will thus result in total memory requirements that
are only O(N log(N)) rather than O(N2). Fourth, although we do not have space
to provide any details, if instead of a single reward function, we are provided with L
reward functions (where the L reward functions are given in aqvance of observing any
experience), then for both algorithms, the number of transitions required to compute
near-optimal policies for all L reward functions simultaneously is only a factor of
O(log(L)) greater than the bounds given above.
Our own view of the result and its implications is:
? Both algorithms enjoy rapid convergence to the optimal policy as a function
of the amount of experience.
? In general, neither approach enjoys a significant advantage in convergence
rate, memory requirements, or handling multiple reward functions. Both are
quite efficient on all counts.
We do not have space to provide a detailed proof of Theorem 1, but instead provide
some highlights of the main ideas. The proofs for both the indirect algorithm and
phased Q-Iearning are actually quite similar, and have at their heart two slightly
/001
Convergence Rates for Q-Learning and Indirect Algorithms
different uniform convergence lemmas. For phased Q-Iearning, it is possible to show
that, for any bound fD on the number of phases to be executed, and for any T > 0,
we can choose mD so that
mD
(l/mD)LVtU?)- LPijVtU) < T
(5)
j
k=l
will hold simultaneously for every (i, a) and for every phase f = 1, . . . , fD. In other
words, at the end of every phase, the empirical estimate of the expected next-state
value for every (i, a) will be close to the true expectation, where here the expectation
is with respect to the current estimated value function Vt.
For the indirect algorithm, a slightly more subtle uniform convergence argument is
required. Here we show that it is possible to choose, for~any bound fI on the number
of iterations of value iteration to be executed on the PM(ij), and for any T > 0, a
value mI such that
(6)
j
j
for every (i,a) and every phase f = 1, . . . ,fI, where the VtU) are the value functions
resulting from performing true value iteration (that is, on the PM (ij)). Equation (6)
essentially says that expectations of the true value functions are quite similar under
either the true or estimated model, even though the indirect algorithm never has
access to the true value functions .
In either case, the uniform convergence results allow us to argue that the corresponding algorithms still achieve successive contractions, as in the classical proof
of value iteration. For instance, in the case of phased Q-Iearning, if we define
b..l
max(i ,a){IQe(i, a) - Ql(i , a)l}, we can derive a recurrence relation for b..l+ 1
as follows :
=
m
,(l/m) L VtU?) -, L Pij VtU)
j
k=l
"E'I',~x,} {
<
7
<
,T + ,b..l .
(7)
(y
P;j v,(j)
+") -
y
P;j V, (j)
}S)
(9)
~
Here we have made use of Equation (5). Since b.. o = 0 (Qo = Qo) , this recurrence
gives b..l :::; Tb/(l--,)) for any f. From this it is not hard to show that for any (i,a)
IQdi , a) - Q*(i, a)1 :::; Tb/(l -,))
+ ,l .
(10)
From this it can be shown that the regret in expected return suffered by the policy
computed by phased Q-Learning after f phases is at most (T, /(1-,) +,l )(2/(1-,)).
The proof proceeds by setting this regret smaller than the desired f, solving for f and
T, and obtaining the resulting bound on m D. The derivation of bounds for the indirect
algorithm is similar.
6
Handling General Exploration Policies
As promised, we conclude our technical results by briefly sketching how we can translate the bounds obtained in Section 5 under the idealized parallel sampling model into
1002
M Kearns and S. Singh
bounds applicable when any fixed policy 1r is guiding the exploration. Such bounds
must, of course, depend on properties of 1r. Due to space limitations, we can only
outline the main ideas; the formal statements and proofs are deferred to a longer
version of the paper.
Let us assume for simplicity that 1r (which may be a stochastic policy) defines an
ergodic Markov process in the MDP M. Thus, 1r induces a unique stationary distribution PM,1[(i, a) over state-action pairs - intuitively, PM ,1[(i, a) is the frequency of
executing action a from state i during an infinite random walk in M according to
1r. Furthermore, we can introduce the standard notion of the mixing time of 1r to
its stationary distribution - informally, this is the number T1[ of steps required such
that the distribution induced on state-action pairs by T1[-step walks according to 1r
will be "very close" to PM,1[ 1. Finally, let us define P1[ = min(i,a){PM,1[(i,
an.
Armed with these notions, it is not difficult to show that the number of steps we must
take under 1r in order to simulate, with high probability, a call to the oracle PS(M) ,
is polynomial in the quantity T1[ / P1[. The intuition is straightforward: at most every
T1[ steps, we obtain an "almost independent" draw from PM,1[(i, a); and with each
independent draw, we have at least probability p of drawing any particular (i, a)
pair. Once we have sampled every (i, a) pair, we have simulated a call to PS(M).
The formalization of these intuitions leads to a version of Theorem 1 applicable to
any 1r, in which the bound is multiplied by a factor polynomial in T1[ / P1[, as desired.
However, a better result is possible . In cases where P1[ may be small or even 0 (which
would occur when 1r simply does not ever execute some action from some state), the
factor T1[ / P1[ is large or infinite and our bounds become weak or vacuous. In such
cases, it is better to define the sub-MDP M1[(O'), which is obtained from M by simply
deleting any (i, a) for which PM,1[(i, a) < a, where a> 0 is a parameter of our choosing . In M1[ (a), P1[ > a by construction, and we may now obtain convergence rates
to the optimal policy in M1[ (a) for both Q-Iearning and the indirect approach like
those given in Theorem 1, multiplied by a factor polynomial in T1[/O'. (Technically,
we must slightly alter the algorithms to have an initial phase that detects and eliminates small-probability state-action pairs, but this is a minor detail.) By allowing
a to become smaller as the amount of experience we receive from 1r grows, we can
obtain an "anytime" result, since the sub-MDP M1[(O') approaches the full MDP M
as 0'-+0.
References
[1] Jaakkola, T., Jordan, M. I., Singh, S. On the convergence of stochastic iterative dynamic programming algorithms. Neural Computation, 6(6), 1185-1201, 1994.
[2] C. J. C. H. Watkins. Learning from Delayed Rewards. Ph.D. thesis, Cambridge University, 1989.
[3] R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. MIT Press,
1998.
[4] S. Mahadevan. Enhancing Transfer in Reinforcement Learning by Building Stochastic
Models of Robot Actions. In Machine Learning: Proceedings of the Ninth International
Conference, 1992.
1 Formally, the degree of closeness is measured by the distance between the transient and
stationary distributions. For brevity here we will simply assume this parameter is set to a
very small, constant value.
| 1531 |@word trial:3 version:3 briefly:2 polynomial:3 tried:1 contraction:1 initial:2 mkearns:1 att:1 current:1 com:1 surprising:1 yet:1 must:8 remove:1 update:3 maxv:1 stationary:5 greedy:3 guess:1 footing:1 successive:1 direct:9 become:2 prove:1 fth:2 introduce:2 indeed:1 expected:7 rapid:3 p1:6 examine:1 roughly:1 discounted:1 detects:1 little:1 armed:1 provided:2 estimating:1 bounded:1 begin:1 maximizes:1 what:4 kind:1 differing:1 nj:1 guarantee:1 quantitative:1 every:18 hypothetical:1 iearning:12 exactly:1 rm:4 qm:6 control:3 enjoy:3 before:2 t1:7 sutton:1 meet:2 phased:12 unique:1 regret:2 empirical:2 convenient:1 word:2 argmaxa:2 get:2 close:3 clever:1 straightforward:1 starting:1 ergodic:1 simplicity:2 assigns:1 notion:2 construction:1 programming:2 us:1 particularly:1 updating:1 observed:9 mentioned:1 intuition:2 complexity:3 reward:10 dynamic:2 singh:5 depend:2 solving:1 technically:1 upon:1 basis:2 indirect:24 various:1 derivation:1 distinct:1 fast:1 describe:1 choosing:1 quite:4 larger:1 say:1 drawing:1 florham:1 statistic:1 favor:1 final:2 seemingly:1 advantage:2 nlog:1 date:1 rapidly:1 mixing:3 translate:1 achieve:2 convergence:19 p:21 requirement:3 produce:1 generating:1 executing:3 derive:4 depending:1 fixing:2 measured:1 ij:12 minor:1 received:3 come:1 stochastic:3 exploration:16 transient:1 material:1 require:1 really:1 summation:1 ofreinforcement:1 hold:1 around:1 proponent:2 applicable:2 mit:1 rather:7 barto:1 jaakkola:1 indicates:1 contrast:1 e7r:1 sense:1 lj:1 entire:1 typically:1 relation:1 going:1 subroutine:1 issue:3 denoted:1 equal:2 construct:1 never:2 once:1 sampling:5 represents:1 park:2 alter:1 simultaneously:3 ve:1 delayed:1 phase:14 consisting:1 vtu:3 interest:2 fd:6 multiply:1 certainly:1 deferred:1 behind:1 implication:4 accurate:3 closer:1 experience:21 walk:3 desired:4 minimal:1 witnessed:2 instance:2 modeling:2 phrase:1 entry:1 uniform:3 reported:1 answer:1 combined:1 international:1 standing:1 off:2 vm:3 michael:1 sketching:2 again:2 reflect:1 thesis:1 choose:3 return:6 explicitly:2 idealized:3 depends:4 try:1 view:1 lab:1 analyze:2 observing:1 reached:1 start:1 parallel:5 il:1 efficiently:1 succession:1 correspond:1 yield:1 generalize:1 weak:1 manages:1 basically:1 trajectory:1 multiplying:1 reach:1 ed:1 frequency:2 obvious:1 proof:6 mi:1 sampled:1 anytime:1 subtle:1 actually:2 reflected:1 execute:2 though:1 generality:1 furthermore:2 predicated:1 hand:1 sketch:1 qo:3 defines:1 quality:4 perhaps:4 aj:2 mdp:10 grows:1 building:1 name:1 true:6 during:2 recurrence:2 qe:1 generalized:1 outline:1 fj:1 fi:2 common:1 behaves:1 rl:1 discussed:2 m1:4 significant:1 cambridge:1 outlined:1 pm:18 baveja:1 access:1 robot:1 longer:2 own:1 dictated:1 certain:1 vt:1 accomplished:1 captured:1 additional:1 greater:1 converge:2 multiple:2 full:1 technical:2 long:1 equally:1 visit:4 permitting:1 variant:1 basic:1 essentially:2 expectation:3 enhancing:1 iteration:9 receive:1 addition:1 suffered:1 eliminates:1 probably:1 duplication:1 induced:1 jordan:1 call:17 near:3 ideal:2 mahadevan:1 enough:1 iterate:1 idea:3 avenue:1 proceed:3 action:20 useful:1 detailed:1 informally:1 amount:8 discount:1 ph:1 induces:1 terrible:1 estimated:5 per:1 shall:2 express:1 demonstrating:1 promised:1 neither:1 clean:1 asymptotically:1 sum:1 run:2 everywhere:1 fourth:1 throughout:1 reasonable:1 saying:1 almost:1 separation:2 draw:2 decision:1 bound:22 oracle:1 occur:1 unlimited:1 aspect:1 simulate:2 argument:4 extremely:1 min:1 performing:1 according:3 across:1 slightly:4 smaller:2 making:1 intuitively:2 gathering:2 heart:1 equation:3 turn:1 r3:1 count:2 fail:1 discus:1 end:1 studying:1 lacked:1 multiplied:2 apply:2 appropriate:2 simulating:1 assumes:1 ensure:2 build:1 classical:3 already:1 question:1 quantity:1 rt:1 md:11 distance:1 separate:3 simulated:2 argue:2 difficult:2 executed:3 ql:2 statement:1 debate:1 claiming:1 policy:55 unknown:1 allowing:1 markov:2 finite:4 payoff:1 ever:2 ninth:1 arbitrary:3 conflates:1 vacuous:1 pair:15 required:15 namely:1 address:2 suggested:1 proceeds:1 tb:2 built:1 max:2 memory:4 deleting:1 natural:1 literature:1 asymptotic:1 loss:1 highlight:1 mixed:4 limitation:1 versus:1 degree:1 offered:1 sufficient:1 verification:1 gather:1 pij:1 viewpoint:1 course:3 surprisingly:3 repeat:1 placed:1 keeping:1 enjoys:1 formal:2 allow:1 sparse:1 distributed:1 transition:25 made:4 reinforcement:7 far:2 approximate:1 emphasize:1 forever:1 implicitly:1 satinder:1 conclude:1 iterative:1 learn:5 mj:1 transfer:1 obtaining:1 main:3 n2:1 formalization:1 sub:2 guiding:1 wish:1 governed:1 watkins:1 loglog:2 third:1 theorem:4 specific:1 r2:1 closeness:1 easier:1 simply:5 infinitely:1 goal:1 careful:1 leaming:1 jf:1 considerable:1 hard:1 determined:3 except:1 operates:1 infinite:3 kearns:4 lemma:1 total:11 called:1 rarely:1 formally:1 brevity:2 handling:2 |
582 | 1,532 | Mean field methods for classification with
Gaussian processes
Manfred Opper
Neural Computing Research Group
Division of Electronic Engineering and Computer Science
Aston University Birmingham B4 7ET, UK.
opperm~aston.ac.uk
Ole Winther
Theoretical Physics II, Lund University, S6lvegatan 14 A
S-223 62 Lund, Sweden
CONNECT, The Niels Bohr Institute, University of Copenhagen
Blegdamsvej 17, 2100 Copenhagen 0, Denmark
winther~thep.lu.se
Abstract
We discuss the application of TAP mean field methods known from
the Statistical Mechanics of disordered systems to Bayesian classification models with Gaussian processes. In contrast to previous approaches, no knowledge about the distribution of inputs is needed.
Simulation results for the Sonar data set are given.
1
Modeling with Gaussian Processes
Bayesian models which are based on Gaussian prior distributions on function spaces
are promising non-parametric statistical tools. They have been recently introduced
into the Neural Computation community (Neal 1996, Williams & Rasmussen 1996,
Mackay 1997). To give their basic definition, we assume that the likelihood of the
output or target variable T for a given input s E RN can be written in the form
p(Tlh(s)) where h : RN --+ R is a priori assumed to be a Gaussian random field.
If we assume fields with zero prior mean, the statistics of h is entirely defined by
the second order correlations C(s, S') == E[h(s)h(S')], where E denotes expectations
M Opper and 0. Winther
310
with respect to the prior. Interesting examples are
C(s, s')
(1)
C(s, s')
(2)
The choice (1) can be motivated as a limit of a two-layered neural network with
infinitely many hidden units with factorizable input-hidden weight priors (Williams
1997). Wi are hyperparameters determining the relevant prior lengthscales of h(s).
The simplest choice C(s, s') = 2::i WiSiS~ corresponds to a single layer percept ron
with independent Gaussian weight priors.
In this Bayesian framework, one can make predictions on a novel input s after
having received a set Dm of m training examples (TJ.1., sJ.1.), J.L = 1, ... , m by using
the posterior distribution of the field at the test point s which is given by
p(h(s)IDm)
=
J
p(h(s)l{hV}) p({hV}IDm)
II dhJ.1..
(3)
J.1.
p(h(s)1 {hV}) is a conditional Gaussian distribution and
p({hV}IDm) =
~P({hV}) II p(TJ.1.IhJ.1.).
(4)
J.1.
is the posterior distribution of the field variables at the training points. Z is a
normalizing partition function and
(5)
is the prior distribution of the fields at the training points. Here, we have introduced
the abbreviations hJ.1. = h(sJ.1.) and CJ.1.V == C(sJ.1., SV).
The major technical problem of this approach comes from the difficulty in performing the high dimensional integrations. Non-Gaussian likelihoods can be only
treated by approximations, where e.g. Monte Carlo sampling (Neal 1997), Laplace
integration (Barber & Williams 1997) or bounds on the likelihood (Gibbs & Mackay
1997) have been used so far. In this paper, we introduce a further approach, which
is based on a mean field method known in the Statistical Physics of disordered
systems (Mezard, Parisi & Virasoro 1987).
We specialize on the case of a binary classification problem, where a binary class
label T = ?1 must be predicted using a training set corrupted by i.i.d label noise.
The likelihood for this problem is taken as
where I\, is the probability that the true classification label is corrupted, i. e. flipped
and the step function, 0(x) is defined as 0(x) = 1 for x > 0 and 0 otherwise. For
such a case, we expect that (by the non-smoothness of the model), e.g. Laplace's
method and the bounds introduced in (Gibbs & Mackay 1997) are not directly
applicable.
31J
Mean Field Methods for Classification with Gaussian Processes
2
Exact posterior averages
In order to make a prediction on an input s, ideally the label with maximum posterior probability should be chosen, i.e. TBayes = argmax r p( TIDm), where the predictive probability is given by P(TIDm) = dhp(Tlh) p(hIDm). For the binary case the
Bayes classifier becomes TBayes = sign(signh(s)), where we throughout the paper let
brackets (... ) denote posterior averages. Here, we use a somewhat simpler approach
by using the prediction
T = sign((h(s))) .
J
This would reduce to the ideal prediction, when the posterior distribution of h(s)
is symmetric around its mean (h(s)). The goal of our mean field approach will be
to provide a set of equations for approximately determining (h( s)) . The starting
point of our analysis is the partition function
Z =
JII
dX;:ihJ.L
J.L
IIp(TJ.LlhJ.L)e~ LI' ,u cI'UxI'X U-
(6)
L I' hl'xl' ,
J.L
where the new auxiliary variables x/l (integrated along the imaginary axis) have
been introduced in order to get rid of C- l in (5).
It is not hard to show from (6) that the posterior averages of the fields at the m
training inputs and at a new test point s are given by
(7)
l/
l/
We have thus reduced our problem to the calculation of the "microscopic orderparameters" (x/l). 1 Averages in Statistical Physics can be calculated from derivatives
of -In Z with respect to small external fields, which are then set to zero, An
equivalent formulation uses the Legendre transform of -In Z as a function of the
expectations , which in our case is given by
G( {(XJ.L) , ((XJ.L)2)}) = -In Z(, /l, A)
+ L(xJ.L)'yJ.L + ~ L
J.L
with
Z( bJ.L, A/l})
=
JII
dX;:ihJ.L IIp(TJ.LlhJ.L)e~
/l
J.L
AJ.L((XJ.L)2) .
(8)
J.L
LI',JAI'Ol'u+Cl' u)x l'x u + LI' xl'(-yl' - hl').
(9)
The additional averages ((XJ.L)2) have been introduced , because the dynamical variables xJ.L (unlike Ising spins) do not have fixed length. The external fields ,J.L , AJ.L
must be eliminated from t~ =
= 0 and the true expectation values of xJ.L and
8G ) -- 0 ,
(x J.L)2 are th ose wh'ICh sat'IS fy 8 ? 8G
xl' )2) -- 8(xl'
t;
3
Naive mean field theory
So far, this description does not give anything new. Usually G cannot be calculated
exactly for the non-Gaussian likelihood models of interest. Nevertheless, based on
mean field theory (MFT) it is possible to guess an approximate form for G.
1 Although the integrations are over the imaginary axis, these expectations come out
positive. This is due to the fact that the integration "measure" is complex as well.
312
M Opper and 0. Winther
Mean field methods have found interesting applications in Neural Computing within
the framework of ensemble learning, where the the exact posterior distribution is
approximated by a simpler one using product distributions in a variational treatment. Such a "standard" mean field method for the posterior of the hf.L (for the
case of Gaussian process classification) is in preparation and will be discussed somewhere else. In this paper, we suggest a different route, which introduces nontrivial
corrections to a simple or "naive" MFT for the variables xl-'. Besides the variational
method (which would be purely formal because the distribution of the xf.L is complex
and does not define a probability), there are other ways to define the simple MFT.
E.g., by truncating a perturbation expansion with respect to the "interactions" Cf.LV
in G after the first order (Plefka 1982). These approaches yield the result
G
~ Gnaive
=
Go -
~ :LCI-'f.L((XI-')2) - ~
I-'
:L
CI-'v(xl-')(XV).
(10)
1-' , v, wl-f.L
Go is the contribution to G for a model without any interactions i.e. when CI-'v = 0
in (9), i.e. it is the Legendre transform of
- In Zo
= l: In [~+ (1 - 2~) <I>
(TI-' ; ; ) ] ,
I-'
where <I>(z) = J~oo .:}f;e- t2 / 2 is an error function. For simple models in Statistical
Physics, where all interactions CI-'V are positive and equal, it is easy to show that
Gnaive will become exact in the limit of an infinite number of variables xl-'. Hence,
for systems with a large number of nonzero interactions having the same orders of
magnitude, one may expect that the approximation is not too bad.
4
The TAP approach
Nevertheless, when the interactions Cf.LV can be both positive and negative (as one
would expect e.g. when inputs have zero mean), even in the thermodynamic limit
and for nice distributions of inputs, an additional contribution tlG must be added
to the "naive" mean field theory (10). Such a correction (often called an Onsager
reaction term) has been introduced for a spin glass model by (Thouless, Anderson
& Palmer 1977) (TAP). It was later applied to the statistical mechanics of single
layer perceptrons by (Mezard 1989) and then generalized to the Bayesian framework
by (Opper & Winther 1996, 1997). For an application to multilayer networks, see
(Wong 1995). In the thermodynamic limit of infinitely large dimension of the input
space, and for nice input distributions, the results can be shown coincide with the
results of the replica framework. The drawback of the previous derivations of the
TAP MFT for neural networks was the fact that special assumptions on the input
distribution had been made and certain fluctuating terms have been replaced by
their averages over the distribution of random data, which in practice would not be
available. In this paper, we will use the approach of (Parisi & Potters 1995), which
allows to circumvent this problem. They concluded (applied to the case of a spin
model with random interactions of a specific type), that the functional form of tlG
should not depend on the type of the "single particle" contribution Go. Hence, one
may use any model in Go, for which G can be calculated exactly (e.g. the Gaussian
regression model) and subtract the naive mean field contribution to obtain the
313
Mean Field Methods for Classification with Gaussian Processes
desired I:1G. For the sake of simplicity, we have chosen the even simpler model
p( TI-'l hi-') '"" 6 (hi-') without changing the final result. A lengthy but straightforward
calculation for this problem leads to the result
(11)
with RI-' == ((xl-')2) - (Xi-')2. The Ai-' must be eliminated using
to the equation
t j( = 0,
which leads
I'
(12)
Note, that with this choice, the TAP mean field theory becomes exact for Gaussian
likelihoods , i. e. for standard regression problems.
Finally, setting the derivatives of GT AP = Gnaive + I:1G with respect to the 4
variables (xl-'), ((xl-')2) ,rl-" AI-' equal to zero, we obtain the equat ions
(13)
v
where D( z ) = e- z 2 /2 /..,j2; is the Gaussian measure. These eqs . have to be solved
numerically together with (12). In contrast, for the naive MFT , the simpler result
AI-' = C 1-'1-' is found.
5
Simulations
Solving the nonlinear system of equations (12,13) by iteration turns out to be quite
straightforward. For some data sets to get convergence, one has to add a diagonal
term v to the covariance matrix C: Cij -+ Cij +6ijV. It may be shown that this term
corresponds to learning with Gaussian noise (with variance v ) added the Gaussian
random field.
Here, we present simulation results for a single data set, the Sonar - Mines versus
Rocks using the same training/test set split as in the original study by (Gorman &
Sejnowski 1988). The input data were pre-processed by linear rescaling such that
over the training set each input variable has zero mean and unit variance. In some
cases the mean field equations failed to converge using the raw data.
A further important feature of TAP MFT is the fact that the method also gives
an approximate leave-one-out estimator for the generalization error , C]oo expressed
in terms of the solution to the mean field equations (see (Opper & Winther 1996,
1997) for more details) . It is also possible to derive a leave-one-out estimator for
the naive MFT (Opper & Winther to be published).
Since we so far haven't dealt with the problem of automatically estimating the
hyperparameters, their number was drastically reduced by setting Wi = (TiN in the
covariances (1) and (2). The remaining hyperparameters, a 2 , K, and v were chosen
M. Opper and 0. Winther
314
Table 1: The result for the Sonar data.
Covariance Function
Algorithm
?test
TAP Mean Field
(1)
0.183
(2)
0.077
Naive Mean Field (1)
0.154
(2)
0.077
Simple Percept ron
0.269(?0.048)
Back-Prop
Best 21ayer - 12 Hidden 0.096(?0.018)
as to minimize ?Ioo . It turned out that the lowest
without noise: K, = v = O.
?Ioo
?exact
100
0.260
0.212
0.269
0.221
?Joo
0.260
0.212
0.269
0.221
was found from modeling
The simulation results are shown in table 1. The comparisons for back-propagation
is taken from (Gorman & Sejnowski 1988). The solution found by the algorithm
turned out to be unique, i.e. different order presentation of the examples and different initial values for the (XIL) converged to the same solution.
In table 1, we have also compared the estimate given by the algorithm with the
exact leave-one-out estimate ?i~~ct obtained by going through the training set and
keeping an example out for testing and running the mean field algorithm on the
rest. The estimate and exact value are in complete agreement. Comparing with
the test error we see that the training set is 'hard' and the test set is 'easy'. The
small difference for test error between the naive and full mean field algorithms also
indicate that the mean field scheme is quite robust with respect to choice of AIL '
6
Discussion
More work has to be done to make the TAP approach a practical tool for Bayesian
modeling. One has to find better methods for solving the equations. A conversion
into a direct minimization problem for a free energy maybe helpful. To achieve this,
one may probably work with the real field variables hJ.l. instead of the imaginary XIL .
A further problem is the determination of the hyperparameters of the covariance
functions. Two ways seem to be interesting here. One may use the approximate
free energy G, which is essentially the negative logarithm of the Bayesian evidence
to estimate the most probable values of the hyperparameters. However, an estimate
on the errors made in the TAP approach would be necessary. Second, one may use
the built-in leave-one-out estimate to estimate the generalization error. Again an
estimate on the validity of the approximation is necessary. It will further be interesting to apply our way of deriving the TAP equations to other models (Boltzmann
machines, belief nets, combinatorial optimization problems), for which standard
mean field theories have been applied successfully.
Acknowledgments
This research is supported by the Swedish Foundation for Strategic Research and
by the Danish Research Councils for the Natural and Technical Sciences through
the Danish Computational Neural Network Center (CONNECT).
Mean Field Methods for Classification with Gaussian Processes
315
References
D. Barber and C. K. I. Williams, Gaussian Processes for Bayesian Classification via
Hybrid Monte Carlo, in Neural Information Processing Systems 9, M . C. Mozer,
M. I. Jordan and T. Petsche, eds., 340-346. MIT Press (1997).
M. N. Gibbs and D. J. C . Mackay, Variational Gaussian Process Classifiers, Preprint
Cambridge University (1997).
R. P. Gorman and T. J. Sejnowski, Analysis of Hidden Units in a Layered Network
Trained to Classify Sonar Targets, Neural Networks 1, 75 (1988) .
D. J. C. Mackay, Gaussian Processes, A Replacement for Neural Networks , NIPS tutorial 1997, May be obtained from http://wol.ra.phy.cam.ac . uk/pub/mackay/.
M. Mezard, The Space of interactions in Neural Networks: Gardner's Computation
with the Cavity Method, J. Phys. A 22, 2181 (1989).
M. Mezard and G. Parisi and M. A. Virasoro, Spin Glass Theory and Beyond,
Lecture Notes in Physics, 9, World Scientific (1987).
R. Neal, Bayesian Learning for Neural Networks, Lecture Notes in Statistics,
Springer (1996).
R. M. Neal, Monte Carlo Implementation of Gaussian Process Models for Bayesian
Regression and Classification, Technical Report CRG-TR-97-2, Dept. of Computer
Science, University of Toronto (1997).
M. Opper and O. Winther, A Mean Field Approach to Bayes Learning in FeedForward Neural Networks, Phys. Rev. Lett. 76, 1964 (1996).
M. Opper and O. Winther, A Mean Field Algorithm for Bayes Learning in Large
Feed-Forward Neural Networks, in Neural Information Processing Systems 9, M. C.
Mozer, M. I. Jordan and T. Petsche, eds., 225-231. MIT Press (1997).
G. Parisi and M. Potters, Mean-Field Equations for Spin Models with Orthogonal
Interaction Matrices, J . Phys. A: Math. Gen . 28 5267 (1995).
T. Plefka, Convergence Condition of the TAP Equation for the Infinite-Range Ising
Spin Glass, J. Phys. A 15, 1971 (1982).
D. J. Thouless, P. W. Anderson and R. G. Palmer , Solution of a 'Solvable Model
of a Spin Glass ', Phil. Mag. 35, 593 (1977).
C. K. I. Williams, Computing with Infinite Networks, in Neural Information Processing Systems 9, M. C. Mozer, M. I. Jordan and T. Petsche, eds., 295-301. MIT
Press (1997).
C. K. I. Williams and C. E. Rasmussen, Gaussian Processes for Regression, in
Neural Information Processing Systems 8, D. S. Touretzky, M. C. Mozer and M. E.
Hasselmo eds., 514-520, MIT Press (1996).
K. Y. M. Wong, Microscopic Equations and Stability Conditions in Optimal Neural
Networks, Europhys. Lett. 30, 245 (1995).
| 1532 |@word simulation:4 covariance:4 tr:1 initial:1 phy:1 pub:1 mag:1 imaginary:3 reaction:1 comparing:1 dx:2 written:1 must:4 partition:2 guess:1 manfred:1 math:1 toronto:1 ron:2 simpler:4 along:1 direct:1 become:1 specialize:1 introduce:1 ra:1 mechanic:2 ol:1 automatically:1 becomes:2 estimating:1 lowest:1 ail:1 onsager:1 ti:2 exactly:2 classifier:2 uk:3 unit:3 positive:3 engineering:1 xv:1 limit:4 approximately:1 ap:1 palmer:2 range:1 unique:1 practical:1 acknowledgment:1 yj:1 testing:1 practice:1 pre:1 suggest:1 get:2 cannot:1 layered:2 wong:2 equivalent:1 center:1 phil:1 williams:6 go:4 starting:1 truncating:1 straightforward:2 simplicity:1 estimator:2 deriving:1 stability:1 laplace:2 target:2 exact:7 us:1 agreement:1 approximated:1 ising:2 preprint:1 solved:1 hv:5 equat:1 mozer:4 ideally:1 mine:1 cam:1 trained:1 depend:1 solving:2 predictive:1 purely:1 division:1 derivation:1 zo:1 monte:3 ole:1 sejnowski:3 lengthscales:1 europhys:1 quite:2 otherwise:1 statistic:2 transform:2 final:1 parisi:4 net:1 rock:1 interaction:8 product:1 j2:1 relevant:1 turned:2 gen:1 achieve:1 opperm:1 description:1 convergence:2 xil:2 leave:4 oo:2 derive:1 ac:2 received:1 eq:1 auxiliary:1 predicted:1 come:2 indicate:1 drawback:1 disordered:2 wol:1 ijv:1 generalization:2 probable:1 crg:1 correction:2 around:1 bj:1 major:1 niels:1 birmingham:1 applicable:1 label:4 combinatorial:1 council:1 hasselmo:1 wl:1 successfully:1 tool:2 minimization:1 mit:4 gaussian:23 hj:2 likelihood:6 contrast:2 glass:4 helpful:1 integrated:1 hidden:4 going:1 classification:10 priori:1 integration:4 mackay:6 special:1 field:35 equal:2 having:2 sampling:1 eliminated:2 flipped:1 t2:1 report:1 haven:1 thouless:2 replaced:1 argmax:1 replacement:1 interest:1 introduces:1 bracket:1 tj:4 ihj:3 bohr:1 necessary:2 sweden:1 orthogonal:1 logarithm:1 desired:1 theoretical:1 virasoro:2 classify:1 modeling:3 strategic:1 plefka:2 too:1 connect:2 corrupted:2 sv:1 winther:10 physic:5 yl:1 together:1 again:1 iip:2 external:2 derivative:2 rescaling:1 li:3 jii:2 later:1 bayes:3 hf:1 contribution:4 minimize:1 spin:7 variance:2 percept:2 ensemble:1 yield:1 dealt:1 bayesian:9 raw:1 lu:1 carlo:3 lci:1 published:1 converged:1 dhj:1 phys:4 touretzky:1 ed:4 lengthy:1 definition:1 danish:2 energy:2 dm:1 treatment:1 wh:1 knowledge:1 cj:1 back:2 feed:1 ayer:1 swedish:1 formulation:1 done:1 anderson:2 correlation:1 nonlinear:1 propagation:1 aj:2 scientific:1 validity:1 true:2 hence:2 symmetric:1 nonzero:1 neal:4 anything:1 generalized:1 complete:1 variational:3 novel:1 recently:1 functional:1 rl:1 b4:1 discussed:1 numerically:1 mft:7 cambridge:1 gibbs:3 ai:3 smoothness:1 particle:1 had:1 joo:1 tlg:2 gt:1 add:1 posterior:9 route:1 certain:1 binary:3 additional:2 somewhat:1 converge:1 ii:3 thermodynamic:2 full:1 technical:3 xf:1 determination:1 calculation:2 prediction:4 basic:1 regression:4 multilayer:1 essentially:1 expectation:4 iteration:1 ion:1 else:1 concluded:1 rest:1 unlike:1 probably:1 seem:1 jordan:3 ideal:1 feedforward:1 split:1 easy:2 xj:7 reduce:1 motivated:1 se:1 maybe:1 processed:1 simplest:1 reduced:2 http:1 tutorial:1 sign:2 group:1 nevertheless:2 changing:1 replica:1 tlh:2 ich:1 throughout:1 electronic:1 entirely:1 layer:2 bound:2 hi:2 ct:1 nontrivial:1 ri:1 sake:1 performing:1 legendre:2 wi:2 rev:1 hl:2 taken:2 equation:10 discus:1 turn:1 needed:1 available:1 apply:1 fluctuating:1 petsche:3 original:1 denotes:1 remaining:1 cf:2 running:1 somewhere:1 uxi:1 added:2 parametric:1 diagonal:1 microscopic:2 blegdamsvej:1 barber:2 fy:1 idm:3 denmark:1 potter:2 length:1 besides:1 dhp:1 cij:2 negative:2 implementation:1 boltzmann:1 conversion:1 rn:2 perturbation:1 community:1 introduced:6 copenhagen:2 tap:11 nip:1 beyond:1 dynamical:1 usually:1 lund:2 built:1 belief:1 difficulty:1 treated:1 circumvent:1 natural:1 hybrid:1 solvable:1 scheme:1 aston:2 gardner:1 axis:2 naive:8 jai:1 prior:7 nice:2 determining:2 expect:3 lecture:2 interesting:4 versus:1 lv:2 foundation:1 supported:1 rasmussen:2 keeping:1 free:2 drastically:1 formal:1 institute:1 opper:9 calculated:3 dimension:1 world:1 lett:2 forward:1 made:2 coincide:1 far:3 sj:3 approximate:3 cavity:1 rid:1 sat:1 assumed:1 xi:2 thep:1 sonar:4 table:3 promising:1 robust:1 ioo:2 expansion:1 cl:1 complex:2 factorizable:1 noise:3 hyperparameters:5 ose:1 mezard:4 xl:10 tin:1 bad:1 specific:1 normalizing:1 evidence:1 ci:4 magnitude:1 gorman:3 subtract:1 infinitely:2 failed:1 expressed:1 springer:1 corresponds:2 prop:1 conditional:1 abbreviation:1 goal:1 presentation:1 hard:2 infinite:3 called:1 perceptrons:1 preparation:1 dept:1 |
583 | 1,533 | Modeling Surround Suppression in VI Neurons
with a Statistically-Derived Normalization Model
Eero P. Simoncelli
Center for Neural Science, and
Courant Institute of Mathematical Sciences
New York University
eero.simoncelli@nyu.edu
Odelia Schwartz
Center for Neural Science
New York University
odelia@cns.nyu.edu
Abstract
We examine the statistics of natural monochromatic images decomposed
using a multi-scale wavelet basis. Although the coefficients of this representation are nearly decorrelated, they exhibit important higher-order
statistical dependencies that cannot be eliminated with purely linear proc~ssing. In particular, rectified coefficients corresponding to basis functions at neighboring spatial positions, orientations and scales are highly
correlated. A method of removing these dependencies is to divide each
coefficient by a weighted combination of its rectified neighbors. Several successful models of the steady -state behavior of neurons in primary
visual cortex are based on such "divisive normalization" computations,
and thus our analysis provides a theoretical justification for these models.
Perhaps more importantly, the statistical measurements explicitly specify
the weights that should be used in computing the normalization signal.
We demonstrate that this weighting is qualitatively consistent with recent physiological experiments that characterize the suppressive effect
of stimuli presented outside of the classical receptive field. Our observations thus provide evidence for the hypothesis that early visual neural
processing is well matched to these statistical properties of images.
An appealing hypothesis for neural processing states that sensory systems develop in response to the statistical properties of the signals to which they are exposed [e.g., 1, 2].
This has led many researchers to look for a means of deriving a model of cortical processing purely from a statistical characterization of sensory signals. In particular, many such
attempts are based on the notion that neural responses should be statistically independent.
The pixels of digitized natural images are highly redundant, but one can always find a
linear decomposition (i.e., principal component analysis) that eliminates second-order corResearch supported by an Alfred P. Sloan Fellowship to EPS, and by the Sloan Center for Theoretical
Neurobiology at NYU.
154
E. P Simoncelli and 0. Schwartz
relation. A number of researchers have used such concepts to derive linear receptive fields
similar to those determined from physiological measurements [e.g., 16,20]. The principal
components decomposition is, however, not unique. Because of this, these early attempts
required additional constraints, such as spatial locality and/or symmetry, in order to achieve
functions approximating cortical receptive fields.
More recently, a number of authors have shown that one may use higher-order statistical measurements to uniquely constrain the choice of linear decomposition [e.g., 7, 9].
This is commonly known as independent components analysis. Vision researchers have
demonstrated that the resulting basis functions are similar to cortical receptive fields, in
that they are localized in spatial position, orientation and scale [e.g., 17, 3]. The associated coefficients of such decompositions are (second-order) decorrelated, highly kurtotic,
and generally more independent than principal components.
But the response properties of neurons in primary visual cortex are not adequately described
by linear processes. Even if one chooses to describe only the mean firing rate of such
neurons, one must at a minimum include a rectifying, saturating nonlinearity. A number of
authors have shown that a gain control mechanism, known as divisive normalization, can
explain a wide variety of the nonlinear behaviors of these neurons [18, 4, II, 12,6]. In most
instantiations of normalization, the response of each linear basis function is rectified (and
typically squared) and then divided by a uniformly weighted sum of the rectified responses
of all other neurons. PhYSiologically, this is hypothesized to occur via feedback shunting
inhibitory mechanisms [e.g., 13, 5]. Ruderman and Bialek [19] have discussed divisive
normalization as a means of increasing entropy.
In this paper, we examine the joint statistics of coefficients of an orthonormal wavelet image decomposition that approximates the independent components of natural images. We
show that the coefficients are second-order decorrelated, but not independent. In particular, pairs of rectified responses are highly correlated. These pairwise dependencies may
be eliminated by dividing each coefficient by a weighted combination of the rectified responses of other neurons, with the weighting determined from image statistics. We show
that the resulting model, with all parameters determined from the statistics of a set of images, can account for recent physiological observations regarding suppression of cortical
responses by stimuli presented outside the classical receptive field. These concepts have
been previously presented in [21, 25].
1 Joint Statistics of Orthonormal Wavelet Coefficients
Multi-scale linear transforms such as wavelets have become popular for image representation. 'TYpically, the basis functions of these representations are localized in spatial position,
orientation, and spatial frequency (scale). The coefficients resulting from projection of
natural images onto these functions are essentially uncorrelated. In addition, a number
of authors have noted that wavelet coefficients have significantly non-Gaussian marginal
statistics [e.g., 10,14]. Because of these properties, we believe that wavelet bases provide
a close approximation to the independent components decomposition for natural images.
For the purposes of this paper, we utilize a typical separable decomposition, based on
symmetric quadrature mirror filters taken from [23]. The decomposition is constructed by
splitting an image into four subbands (lowpass, vertical, horizontal, diagonal), and then
recursively splitting the lowpass subband.
Despite the decorrelation properties of the wavelet decomposition, it is quite evident that
wavelet coefficients are not statistically independent [26, 22]. Large-magnitude coefficients
(either positive or negative) tend to lie along ridges with orientation matching that of the
subband. Large-magnitude coefficients also tend to occur at the same relative spatiallocations in subbands at adjacent scales, and orientations. To make these statistical relationships
| 1533 |@word dividing:1 effect:1 hypothesized:1 concept:2 approximating:1 classical:2 adequately:1 symmetric:1 filter:1 receptive:5 primary:2 decomposition:9 adjacent:1 diagonal:1 bialek:1 exhibit:1 uniquely:1 noted:1 steady:1 recursively:1 evident:1 ridge:1 demonstrate:1 image:11 relationship:1 recently:1 must:1 early:2 negative:1 purpose:1 discussed:1 proc:1 approximates:1 eps:1 measurement:3 vertical:1 neuron:7 surround:1 observation:2 weighted:3 provides:1 characterization:1 nonlinearity:1 always:1 gaussian:1 neurobiology:1 digitized:1 cortex:2 mathematical:1 along:1 constructed:1 base:1 become:1 derived:1 recent:2 pair:1 required:1 pairwise:1 suppression:2 behavior:2 examine:2 multi:2 minimum:1 additional:1 decomposed:1 typically:2 relation:1 increasing:1 redundant:1 signal:3 ii:1 matched:1 pixel:1 simoncelli:3 orientation:5 decorrelation:1 natural:5 spatial:5 marginal:1 field:5 divided:1 shunting:1 eliminated:2 look:1 nearly:1 vision:1 essentially:1 stimulus:2 schwartz:2 control:1 normalization:6 addition:1 positive:1 fellowship:1 relative:1 despite:1 cns:1 suppressive:1 eliminates:1 attempt:2 localized:2 firing:1 tend:2 highly:4 monochromatic:1 consistent:1 uncorrelated:1 statistically:3 variety:1 unique:1 supported:1 regarding:1 institute:1 neighbor:1 wide:1 divide:1 significantly:1 theoretical:2 projection:1 matching:1 feedback:1 cortical:4 modeling:1 sensory:2 author:3 kurtotic:1 qualitatively:1 cannot:1 onto:1 close:1 york:2 commonly:1 generally:1 demonstrated:1 center:3 successful:1 transforms:1 instantiation:1 characterize:1 dependency:3 eero:2 splitting:2 chooses:1 inhibitory:1 physiologically:1 importantly:1 deriving:1 orthonormal:2 alfred:1 notion:1 correlated:2 justification:1 symmetry:1 four:1 squared:1 hypothesis:2 utilize:1 account:1 sum:1 quadrature:1 coefficient:13 explicitly:1 sloan:2 vi:1 ssing:1 position:3 lie:1 weighting:2 wavelet:8 removing:1 rectifying:1 occur:2 exposed:1 constraint:1 purely:2 constrain:1 nyu:3 physiological:3 basis:5 evidence:1 joint:2 lowpass:2 mirror:1 separable:1 magnitude:2 rectified:6 researcher:3 describe:1 explain:1 combination:2 locality:1 entropy:1 outside:2 decorrelated:3 led:1 quite:1 visual:3 appealing:1 frequency:1 saturating:1 statistic:6 associated:1 gain:1 taken:1 popular:1 previously:1 mechanism:2 neighboring:1 higher:2 courant:1 determined:3 typical:1 uniformly:1 specify:1 achieve:1 response:8 subbands:2 principal:3 divisive:3 horizontal:1 ruderman:1 nonlinear:1 include:1 odelia:2 derive:1 develop:1 perhaps:1 believe:1 subband:2 |
584 | 1,534 | Analog VLSI Cellular Implementation of the
Boundary Contour System
Gert Cauwenberghs and James Waskiewicz
Department of Electrical and Computer Engineering
Johns Hopkins University
3400 North Charles Street
Baltimore, MD 21218-2686
E-mail: {gert, davros }@bach. ece. jhu. edu
Abstract
We present an analog VLSI cellular architecture implementing a simpli.fied version of the Boundary Contour System (BCS) for real-time image
processing. Inspired by neuromorphic models across several layers of
visual cortex, the design integrates in each pixel the functions of simple cells, complex cells, hyper-complex cells, and bipole cells, in three
orientations interconnected on a hexagonal grid. Analog current-mode
CMOS circuits are used throughout to perform edge detection, local inhibition, directionally selective long-range diffusive kernels, and renormalizing global gain control. Experimental results from a fabricated 12 x 10
pixel prototype in 1.2 J-tm CMOS technology demonstrate the robustness
of the architecture in selecting image contours in a cluttered and noisy
background.
1 Introduction
The Boundary Contour System (BCS) and Feature Contour System (FCS) combine models
for processes of image segmentation, feature filling, and surface reconstruction in biological vision systems [1 ],[2]. They provide a powerful technique to recognize patterns and
restore image quality under excessive fixed pattern noise, such as in SAR images [3]. A
related model with similar functional and structural properties is presented in [4].
The motivation for implementing a relatively complex model such as BCS and FCS on
the focal-plane is dual. First, as argued in [5], complex neuromorphic active pixel designs
become viable engineering solutions as the feature size of the VLSI technology shrinks
significantly below the optical diffraction limit, and more transistors can be stuffed in each
pixel. The pixel design that we present contains 88 transistors, likely the most complex
G. Cauwenberghs and J. Waskiewicz
658
Bipole Cells
(long-range orientational cooperaHon)
...... Diffusive Network
...... Local itlhibiticnt
Input Image
(locally normalized and contrast
enhanced; diffused)
BCS
Focal-Plane Receptors;
...... Ri1ndom-Access Inputs
FCS
Figure 1: Diagram of BCSIFCS model for image segmentation, feature filling, and surface
reconstruction. Three layers represent simple, complex and bipole cells.
active pixel imager ever put on silicon. Second, our motivation is to extend the functionality
of previous work on analog VLSI neuromorphic image processors for image boundary
segmentation, e.g. [6, 7, 5, 8,9] which are based on simplified physical models that do not
include directional selectivity and/or long-range signal aggregation for boundary fonnation
in the presence of significant noise and clutter. The analog VLSI implementation of BCS
reported here is a first step towards this goal, with the additional objectives of real-time,
low-power operation as required for demanding target recognition applications. As an
alternative to focal-plane optical input, the image can be loaded electronically through
random-access pixel addressing.
The BCS model encompasses visual processing at different levels, including several layers of cells interacting through shunting inhibition, long-range cooperative excitation, and
renonnalization. The implementation architecture, shown schematically in Figure 1, partitions the BCS model into three levels: simple cells, complex and hypercomplex cells, and
bipole cells.
Simple cells compute unidirectional gradients of nonnalized intensity obtained from the
photoreceptors. Complex (hyper-complex) cells perfonn spatial and directional competition (inhibition) for edge fonnation. Bipole cells perfonn long-range cooperation for
boundary contour enhancement, and exert positive feedback (excitation) onto the hypercomplex cells. Our present implementation does not include the FCS model, which completes and fills features through diffusive spatial filtering of the image blocked by the edges
fonned in BCS.
2
Modified BeS Algorithm and Implementation
We adopted the BCS algorithm for analog continuous-time implementation on a hexagonal
grid, extending in three directions u, v and w on the focal plane as indicated schematically
in Figure 2. For notational convenience, let subscript 0 denote the center pixel and ?u, ?v
and ?w its six neighbors. Components of each complex cell "vector" C i at grid location i,
along three directions of edge selectivity, are indicated with superscript indices u, v and w.
In the implemented circuit model, a pixel unit consists of a photosensor (or random-access
analog memory) sourcing a current indicating light intensity, gradient computation and
rectification circuits implementing simple cells in three directions, and one complex (hyper-
659
Analog VLSI Cellular Implementation of the Boundary Contour System
Figure 2: Hexagonal arrangement of Bes pixels, at the level of simple and complex cells,
extending in three directions u, v and w in the focal plane.
complex) cell and one bipole cell for each of the three directions.
The photosensors generate a current Ii that is proportional to intensity. Through current
mirrors, the currents Ii propagate in the three directions u, v, and w as noted in Figure
2. Rectified finite-difference gradient estimates of Ii are obtained for each of the three
hexagonal directions. These gradients excite the complex cells
cl.
Lateral inhibition among spatially (i) and directionally (j) adjacent complex cells implement the function of hypercomplex cells for edge enhancement and noise reduction. The
complex output (Cl) is inhibited by local complex cell outputs in the two competing directions of j. Co is additionally inhibited by the complex cells of the four nearest neighbors
in competing locations i with parallel orientation.
Bf,
A directionally selective interconnected diffusive network of bipole cells
interacting
provides long range cooperative feedback, and enhances smooth
with the complex cells
edge contours while reducing spurious edges due to image clutter.
is excited by bipole
on the line crossing i in the same direction j.
interaction received from the bipole cell
cl,
Bf
cl
The operation of the (hyper-)complex cells in the hexagonal arrangement is summarized in
the following equation, for one of the three directions u:
where:
1. 1~(Iv + Iw) - 101 represents the rectified gradient input as approximated on the
hexagonal grid;
+ Co) is the inhibition from locally opposing directions;
o:'(C;: + c::; + c~v + C~w) is inhibition from non-aligned neighbors in the same
2. 0:(C8
3.
direction; and
4.
f3B8 is the excitation through long-range cooperation from the bipole cell.
660
G. Cauwenberghs and J Waskiewicz
Figure 3: Network of bipole cells, implemented on a hexagonal resistive grid using orientationally tuned diffusors extending in three directions. glat! gvert determines the spatial
extent of the dipole, whereas glat! gcross sets the directional selectivity.
The bipole cell resistive grid (Figure 3) implements a three-fold cross-coupled, directionally polarized, long-range diffusive kernel, formulated as follows:
(2)
where K::, K::, and K~ represent spatial convolutional kernels implementing bipole fields
symmetrically polarized in the u, v and w directions. Diffusive kernels can be efficiently
implemented with a distributed representation using resistive diffusive elements [7, 10].
Three linear networks of diffusor elements are used, complemented with cross-links of
adjustable strength, to control the degree of direction selectivity and the spatial spread
of the kernel. Finally, the result (2) is locally normalized, before it is fed back onto the
complex cells.
3 Analog VLSI Implementation
The simplified circuit diagram of the BCS cell, including simple, complex and bipole cell
functions on a hexagonal grid, is shown in Figure 4.
The image is acquired either optically from phototransistors on the focal-plane, or in direct
electronic format through random-access pixel addressing, Figure 4 (a). The simple cell
portion in Figure 4 (b) combines the local intensity 10 with intensities Iv and Iw received
from neighboring cells to compute the rectified gradient in (l), using distributed current
mirrors and an absolute value circuit. A pMOS load converts the complex cell output into
a voltage representation C8 for distribution to neighboring nodes and complementary orientations: local inhibition for spatial and directional competition in Figure 4 (c), and longrange cooperation through the bipole layer in Figure 4 (d). The linear diffusive kernel is
implemented in current-mode using ladder structures of subthreshold MOS transistors [7],
three families extending in each direction with cross-links for directional dispersion as indicated in Figure 3.
Voltage biases control the spatial extent and directional selectivity of the interactions, as
661
Analog VLSI Cellular Implementation of the Boundary Contour System
Va
Vo
Cov
4
Va'
Cow
~
~_WU
C+~
CoU
PHOTO
II~YVin
T
(C)
Vnorm
Cou
(a)
Bou
Bov
Cou
V+v4
B+uu
BoW
Vve~
~IBoU
v~
(b)
(d)
VbM~1Mm
Vnorm
Vthresh4
(e)
Figure 4: Simplified circuit schematic of one BCS cell in the hexagonal array, showing
only one of three directions, the other directions being symmetrical in implementation. (a)
Photosensor and random-access input selection circuit. (b) Simple cell rectified gradient
calculation. (c) Complex cell spatial and orientational inhibition. (d) Bipole cell directionallong range cooperation. (e) Bipole global gain and threshold control.
well as the relative strength of inhibition and excitation, and the level of renormalization,
for the complex and bipole cells. The values for gvert. glat and gcross controlling the
bipole kernel are set externally by applying gate bias voltages Vvert. Vlat and Vcross, respectively. Likewise, the constants a, a' and /3 in (1) are set independently by the applied
source voltages Va, Va' and Vt1. Global normalization and thresholding of the bipo1e response for improved stability of edge formation is achieved through an additional diffusive
network that acts as a localized Gilbert-type current normalizer (only partially shown in
Figure 4 (e?.
4 Experimental Results
A prototype 12 x 10 pixel array has been fabricated and tested. The pixel unit, illustrated in
Figure 5 (a), has been designed for testability, and has not been optimized for density. The
pixel contains 88 transistors including a phototransistor, a large sample-and-hold capacitor,
and three networks of interconnections in each of the three directions, requiring a fanin/fan-out of 18 node voltages across the interface of each pixel unit. A micrograph of the
Tiny 2.2 x 2.2 sq. mm chip, fabricated through MOSIS in 1.2 J.Lm CMOS technology, is
shown in Figure 5 (b).
We have tested the BCS chip both under focal-plane optical inputs, and random-access
direct electronic inputs. Input currents from optical input under ambient room lighting
conditions are around 30 nA. The experimental results reported here are obtained by feeding test inputs electronically. The response of the BCS chip to two test images of interest
are shown in Figures 6 and 7.
G. Cauwenberghs and J Waskiewicz
662
(a)
(b)
Figure 5: BCS processor. (a) Pixel layout. (b) Chip micrograph.
. ,
~~~~~-ir\~"':";-4;?
-
.. - ..
-,,-
-\
'I \ I II-~ -->';-
:";- ':""'l){
' -\-~
X \ \
,
\
\
..
.
--
-\-*~ *-\-*+ , .:.,..-';-
.' \
\ - \ \ \ X\ \ / \
\
(b)
.
..
.
"
--:- ,***\ / \ \-"
\ '.
*- ~
, '.
-~---~ -\-+ ':"""T
... \ ){ + +~~ --.:.,.. ,
- \ 1 \ -:- -:-- ' \ + :..;* :..; .:,, \- ..,.+ , ...\,..
..... -*-\-+-r '\:
"
.
-- ~~~ -\- -\-,""""':"--+-4;-
.... ->.;-* :..;-.;- :..; -I-
(a)
-
\
\
\
l
\ . -'.;... -
'* * \ '........
-'o- ~ -'o- -\- -\-- -\-....,.* \-
\ \ \ Y \
\
\
\
j
j
\
.
\
\
(c)
Figure 6: Experimental response of the BCS chip to a curved edge. (a) Reconstructed
input image. (b) Complex field. (c) Bipolefield. The thickness of the bars on the grid
represent the measured components in the three directions.
Figure 6 illustrates the interpolating directional response to a curved edge in the input. varying in direction between two of the principal axes (u and w in the example). Interpolation
between quantized directions is important since implementing more axes on the grid incurs
a quadratic cost in complexity. The second example image contains a bar with two gaps of
different diameter. for the purpose of testing BCS's capacity to extend contour boundaries
across clutter. The response in Figure 7 illustrates a characteristic of bipole operation. in
which short-range discontinuities are bridged but large ones are preserved.
5
Conclusions
An analog VLSI cellular architecture implementing the Boundary Contour System (BCS)
on the focal plane has been presented. A diffusive kernel with distributed resistive networks
has been used to implement long-range interactions of bipole cells without the need of
excessive global interconnects across the array of pixels. The cellular model is fairly easy to
implement. and succeeds in selecting boundary contours in images with significant clutter.
Analog VLSI Cellular Implementation of the Boundary Contour System
-T~~+-\-~ ---:--...!.r-\-~
\.\ \
'''
\
~~ -\-~ ~ ___ -+---Io;-*"'\
\
\\\\
I \ I , \ \ I \ I X
---~---~--
\ ; \ 1" \ . X \
+-*~",-"'"-~?
----1 .. ----
\
-~
\
,\
\
\ \ \ I \
~
\
\
\
1 \
~
-T++ ~ + - +++*
~
.
'\
\
\
,v
\,
.
.
(b)
?
,
'.
?
----+~~--~
j x x Y
x \ X/
\
~
~
\,
\
\
\
-f--'-*-'I""-\-~+-'-*...:_
\ \ \ ~ \ 1 '. \ X \
'\ ,; '. :,' , 1. \ \ '/ \
\
\
. . ~+~...!.::-+ .. ~
-\- * - - - -I- - -\- + -\?
(a)
663
\\
.:\\
\
\\\\
"\ -.\\\\\
\~~
'~\
\\\\\\\
(c)
Figure 7: Experimental response of the BCS chip to a bar with two gaps of different size.
(a) Reconstructed input image. (b) Complex field. (c) Bipolefield.
Experimental results from a 12 x 10 pixel prototype demonstrate expected BCS operation
on simple examples. While this size is small for practical applications, the analog cellular
architecture is fully scalable towards higher resolutions. Based on the current design, a
10, OOO-pixel array in 0.5 J.tm CMOS technology would fit a 1 cm 2 die.
Acknowledgments
This research was supported by DARPA and ONR under MURI grant NOO0l4-95-1-0409.
Chip fabrication was provided through the MOSIS service.
References
[1] S. Grossberg, "Neural Networks for Visual Perception in Variable Illumination," Optics News,
pp. 5-10, August 1988.
[2] S. Grossberg, "A Solution of the Figure-Ground Problem for Biological Vision," Neural Networks, vol. 6, pp. 463-482, 1993.
[3] S. Grossberg, E. Mingolla, and J. Williamson, "Synthetic Aperture Radar Processing by a
Multiple Scale Neural System for Boundary and Surface Representation," Neural Networks,
vol. 9 (1), January 1996.
[4] Z.P. Li, "A Neural Model of Contour Integration in the Primary Visual Cortex," Neural Computation, vol. 10 (4), pp. 903-940, 1998.
[5] K.A. Boahen, "A Retinomorphic Vision System," IEEE Micro, vol. 16 (5), pp. 30-39, Oct.
1996.
[6] lG. Harris, C. Koch, and J. Luo, "A Two-Dimensional Analog VLSI Circuit for Detecting
Discontinuities in Early Vision," Science, vol. 248, pp. 1209-1211, June 1990.
[7] A.G. Andreou, K.A. Boahen, P.O. Pouliquen, A. Pavasovic, R.E. Jenkins, and K. Strohbehn,
"Current-Mode Subthreshold MOS Circuits for Analog VLSI Neural Systems," IEEE Transactions on Neural Networks, vol. 2 (2), pp 205-213, 1991.
[8] L. Dron McIlrath, "A CCD/CMOS Focal-Plane Array Edge Detection Processor Implementing
the Multiscale Veto Algorithm," IEEE 1. Solid State Circuits, vol. 31 (9), pp 1239-1248, 1996.
[9] P. Venier, A. Mortara. X. Arreguit and E.A. Vittoz, "An Integrated Cortical Layer for Orientation Enhancement," IEEE 1. Solid State Circuits, vol. 32 (2), pp 177-186, Febr. 1997.
[10] E. Fragniere, A. van Schaik and E. Vittoz, "Reactive Components for Pseudo-Resistive Networks," Electronic Letters, vol. 33 (23), pp 1913-1914, Nov. 1997.
| 1534 |@word version:1 bf:2 propagate:1 excited:1 incurs:1 solid:2 reduction:1 contains:3 selecting:2 optically:1 tuned:1 current:11 luo:1 john:1 partition:1 designed:1 plane:9 short:1 schaik:1 provides:1 quantized:1 node:2 location:2 detecting:1 along:1 direct:2 become:1 viable:1 consists:1 resistive:5 combine:2 acquired:1 expected:1 inspired:1 provided:1 circuit:11 cm:1 fabricated:3 perfonn:2 pseudo:1 act:1 control:4 imager:1 unit:3 grant:1 positive:1 before:1 engineering:2 local:5 service:1 limit:1 io:1 receptor:1 subscript:1 interpolation:1 exert:1 co:2 range:11 grossberg:3 practical:1 acknowledgment:1 testing:1 implement:4 sq:1 jhu:1 significantly:1 bipole:21 onto:2 convenience:1 selection:1 put:1 applying:1 phototransistor:1 gilbert:1 center:1 layout:1 cluttered:1 bou:1 independently:1 resolution:1 dipole:1 array:5 fill:1 stability:1 gert:2 sar:1 enhanced:1 target:1 controlling:1 crossing:1 element:2 recognition:1 approximated:1 muri:1 cooperative:2 electrical:1 news:1 boahen:2 complexity:1 testability:1 radar:1 darpa:1 chip:7 vt1:1 hyper:4 formation:1 pmos:1 interconnection:1 cov:1 noisy:1 superscript:1 directionally:4 transistor:4 reconstruction:2 interconnected:2 interaction:3 neighboring:2 aligned:1 bow:1 competition:2 enhancement:3 extending:4 renormalizing:1 cmos:5 measured:1 nearest:1 received:2 implemented:4 uu:1 vittoz:2 direction:22 functionality:1 implementing:7 argued:1 venier:1 feeding:1 biological:2 mm:2 hold:1 around:1 koch:1 ground:1 fragniere:1 mo:2 lm:1 early:1 purpose:1 integrates:1 iw:2 modified:1 varying:1 voltage:5 ax:2 june:1 notational:1 contrast:1 normalizer:1 integrated:1 spurious:1 vlsi:12 selective:2 pixel:19 dual:1 orientation:4 among:1 retinomorphic:1 spatial:8 integration:1 fairly:1 field:3 represents:1 filling:2 excessive:2 inhibited:2 micro:1 recognize:1 opposing:1 detection:2 interest:1 light:1 ambient:1 edge:11 mortara:1 iv:2 neuromorphic:3 cost:1 addressing:2 fabrication:1 reported:2 thickness:1 synthetic:1 density:1 v4:1 hopkins:1 na:1 li:1 summarized:1 north:1 cauwenberghs:4 portion:1 aggregation:1 parallel:1 unidirectional:1 ir:1 convolutional:1 loaded:1 characteristic:1 efficiently:1 likewise:1 subthreshold:2 directional:7 lighting:1 rectified:4 processor:3 pp:9 james:1 gain:2 bridged:1 segmentation:3 orientational:2 back:1 higher:1 response:6 improved:1 ooo:1 photosensors:1 shrink:1 multiscale:1 mode:3 quality:1 indicated:3 normalized:2 requiring:1 spatially:1 illustrated:1 adjacent:1 noted:1 excitation:4 die:1 demonstrate:2 vo:1 interface:1 image:18 charles:1 functional:1 physical:1 analog:15 extend:2 silicon:1 significant:2 blocked:1 grid:9 focal:9 arreguit:1 access:6 cortex:2 surface:3 inhibition:9 selectivity:5 onr:1 additional:2 simpli:1 signal:1 ii:5 multiple:1 bcs:19 fonnation:2 smooth:1 calculation:1 bach:1 long:9 cross:3 shunting:1 va:4 schematic:1 scalable:1 vision:4 pavasovic:1 kernel:8 represent:3 normalization:1 achieved:1 cell:43 diffusive:10 preserved:1 background:1 whereas:1 schematically:2 baltimore:1 diagram:2 completes:1 source:1 veto:1 photosensor:2 capacitor:1 structural:1 presence:1 symmetrically:1 easy:1 fit:1 architecture:5 competing:2 cow:1 prototype:3 tm:2 six:1 mingolla:1 clutter:4 locally:3 diameter:1 generate:1 vol:9 four:1 threshold:1 micrograph:2 stuffed:1 mosis:2 convert:1 letter:1 powerful:1 throughout:1 family:1 electronic:3 wu:1 polarized:2 diffraction:1 layer:5 fold:1 fan:1 quadratic:1 strength:2 optic:1 c8:2 hexagonal:9 optical:4 relatively:1 format:1 department:1 across:4 rectification:1 equation:1 pouliquen:1 fed:1 photo:1 adopted:1 operation:4 jenkins:1 alternative:1 robustness:1 gate:1 include:2 ccd:1 diffused:1 objective:1 arrangement:2 primary:1 md:1 enhances:1 gradient:7 link:2 lateral:1 capacity:1 street:1 mail:1 extent:2 cellular:8 index:1 lg:1 implementation:11 design:4 adjustable:1 perform:1 dispersion:1 finite:1 curved:2 january:1 ever:1 nonnalized:1 diffusor:2 interacting:2 august:1 vbm:1 intensity:5 required:1 optimized:1 andreou:1 discontinuity:2 renonnalization:1 bar:3 below:1 pattern:2 perception:1 encompasses:1 bov:1 including:3 memory:1 power:1 demanding:1 restore:1 interconnects:1 technology:4 ladder:1 coupled:1 relative:1 fully:1 filtering:1 proportional:1 localized:1 degree:1 thresholding:1 tiny:1 cooperation:4 sourcing:1 supported:1 electronically:2 bias:2 neighbor:3 absolute:1 distributed:3 van:1 boundary:13 feedback:2 cortical:1 contour:14 fonned:1 simplified:3 hypercomplex:3 transaction:1 reconstructed:2 nov:1 longrange:1 aperture:1 global:4 active:2 photoreceptors:1 symmetrical:1 excite:1 continuous:1 additionally:1 williamson:1 complex:27 cl:4 interpolating:1 spread:1 motivation:2 noise:3 complementary:1 fied:1 vve:1 renormalization:1 externally:1 load:1 showing:1 phototransistors:1 mirror:2 illumination:1 illustrates:2 gap:2 fanin:1 fcs:4 strohbehn:1 likely:1 visual:4 partially:1 determines:1 complemented:1 harris:1 oct:1 goal:1 formulated:1 towards:2 room:1 reducing:1 principal:1 ece:1 experimental:6 succeeds:1 indicating:1 reactive:1 tested:2 |
585 | 1,535 | DTs: Dynamic Trees
Christopher K. I. Williams
Nicholas J. Adams
Institute for Adaptive and Neural Computation
Division of Informatics, 5 Forrest Hill
Edinburgh, EHI 2QL, UK.
http://www.anc.ed.ac . uk/
ckiw~dai.ed.ac.uk
nicka~dai.ed.ac.uk
Abstract
In this paper we introduce a new class of image models, which we
call dynamic trees or DTs. A dynamic tree model specifies a prior
over a large number of trees, each one of which is a tree-structured
belief net (TSBN). Experiments show that DTs are capable of
generating images that are less blocky, and the models have better
translation invariance properties than a fixed, "balanced" TSBN.
We also show that Simulated Annealing is effective at finding trees
which have high posterior probability.
1
Introduction
In this paper we introduce a new class of image models, which we call dynamic
trees or DTs. A dynamic tree model specifies a prior over a large number of trees,
each one of which is a tree-structured belief net (TSBN) . Our aim is to retain
the advantages of tree-structured belief networks, namely the hierarchical structure
of the model and (in part) the efficient inference algorithms, while avoiding the
"blocky" artifacts that derive from a single, fixed TSBN structure. One use for
DTs is as prior models over labellings for image segmentation problems.
Section 2 of the paper gives the theory of DTs, and experiments are described in
section 3.
2
Theory
There are two essential components that make up a dynamic tree network (i) the
tree architecture and (ii) the nodes and conditional probability tables (CPTs) in
the given tree. We consider the architecture question first.
DTs: Dynamic Trees
635
o
o
o
o
o
o
0
0
o
0
0
o
000
10000000000000000
(a)
(c)
(d)
Figure 1: (a) "Naked" nodes, (b) the "balanced" tree architecture, (c) a sample
from the prior over Z, (d) data generated from the tree in (c).
Consider a number of nodes arranged into layers, as in Figure lea). We wish
to construct a tree structure so that any child node in a particular layer will be
connected to a parent in the layer above. We also allow there to be a null parent for
each layer, so that any child connected to it will become a new root. (Technically we
are constructing a forest rather than a tree.) An example of a structure generated
using this method is shown in Figure 1 (c).
There are a number of ways of specifying a prior over trees. If we denote by Zi the
indicator vector which shows to which parent node i belongs, then the tree structure
is specified by a matrix Z whose columns are the individual Zi vectors (one for each
node). The scheme that we have investigated so far is to set P(Z) = It P(Zi).
In our work we have specified P(Zi) as follows. Each child node is considered to
have a "natural" parent-its parent in the balanced structure shown in Figure l(b).
Each node in the parent layer is assigned an "affinity" for each child node, and
the "natural" parent has the highest affinity. Denote the affinity of node k in the
parent layer by ak. Then we choose P(Zi = ek) = e!3a/e / EjEPai e!3 a j , where (3 is
some positive constant and ek is the unit vector with a 1 in position k. Note that
the "null" parent is included in the sum, and has affinity anull associated with it,
which affects the relative probability of "orphans". We have named this prior the
"full-time-node-employment" prior as all the nodes participate in the creation of
the tree structure to some degree.
Having specified the prior over architectures, we now need to translate this into a
TSBN. The units in the tree are taken to be C-class multinomial random variables.
Each layer of the structure has associated with it a prior probability vector 7f1
and CPT MI. Given a particular Z matrix which specifies a forest structure, the
probability of a particular instantiation of all of the random variables is simply
the product of the probabilities of all of the trees, where the appropriate root
probabilities and CPTs are picked up from the 7fIS and MIS. A sample generated
from the tree structure in Figure l(c) is shown in Figure led).
C. K. I. Williams and N. 1. Adams
636
Our intuition as to why DTs may be useful image models is based on the idea that
most pixels in an image are derived from a single object. We think of an object as
being described by a root of a tree, with the scale of the object being determined
by the level in the tree at which the root occurs. In this interpretation the ePTs
will have most of their probability mass on the diagonal.
Given some data at the bottom layer of units, we can form a posterior over the tree
structures and node instantiations of the layers above. This is rather like obtaining
a set of parses for a number of sentences using a context-free grammar l .
In the DT model as described above different examples are explained by different
trees. This is an important difference with the usual priors over belief networks as
used, e.g. in Bayesian averaging over model structures. Also, in the usual case of
model averaging, there is normally no restriction to TSBN structures, or to tying
the parameters (1rIS and MIS) between different structures.
2.1
Inference in DTs
We now consider the problem of inference in DTs, i.e. obtaining the posterior
P(Z, XhlXv) where Z denotes the tree-structure, Xv the visible units (the image
clamped on the lowest level) and X h the hidden units. In fact, we shall concentrate on obtaining the posterior marginal P(ZIXv), as we can obtain samples from
P(XhIXv, Z) using standard techniques for TSBNs.
There are a very large number of possible structures; in fact for a set of nodes created from a balanced tree with branching factor b and depth D (with the top level
indexed by 1) there are IT~=2(b(d-2) + l)b(d-l) possible forest structures. Our objective will be to obtain the maximum a posteriori (MAP) state from the posterior
P(ZIXv) ex P(Z)P(XvIZ) using Simulated Annealing.2 This is possible because
two components P(Z) and P(XvIZ) are readily evaluated. P(XvIZ) can be computed from ITr (Exr A(X r )'7r(xr)), where A(Xr) and 7r(xr) are the Pearl-style vectors
of each root r of the forest.
An alternative to sampling from the posterior P(Z, XhlXv ) is to use approximate
inference. One possibility is to use a mean-field-type approximation to the posterior
of the form QZ(Z)Qh(Xh) (Zoubin Ghahramani, personal communication, 1998).
2.2
Comparing DTs to other image models
Fixed-structure TSBNs have been used by a number of authors as models of images
(Bouman and Shapiro, 1994), (Luettgen and Willsky, 1995). They have an attractive multi-scale structure, but suffer from problems due to the fixed tree structure,
which can lead to very "blocky" segmentations. Markov Random Field (MRF)
models are also popular image models; however, one of their main limitations is
that inference in a MRF is NP-hard. Also, they lack an hierarchical structure. On
the other hand, stationarity of the process they define can be easily ensured, which
lCFGs have a O(n 3 ) algorithm to infer the MAP parse; however, this algorithm depends
crucially on the one-dimensional ordering of the inputs. We believe that the possibility of
crossed links in the DT architecture means that this kind of algorithm is not applicable to
the DT case. Also, the DT model can be applied to 2-d images, where the O(n 3 ) algorithm
is not applicable.
2It is also possible to sample from the posterior using, e.g. Gibbs Sampling.
637
DTs: Dynamic Trees
is not the case for fixed-structure TSBNs. One strategy to overcome the fixed structure of TSBNs is to break away from the tree structure, and use belief networks
with cross connections e.g. (Dayan et ai., 1995). However, this means losing the
linear-time belief-propagation algorithms that can be used in trees (Pearl, 1988)
and using approximate algorithms. While it is true that inference over DTs is also
NP-hard, we do retain a"clean" semantics based on the fact that we expect that
each pixel should belong to one object, which may lead to useful approximation
schemes.
3
Experiments
In this section we describe two experiments conducted on the DT models. The first
has been designed to compare the translation performance of DTs with that of the
balanced TSBN structure and is described in section 3.1. In section 3.2 we generate
2-d images from the DT model, find the MAP Dynamic Tree for these images, and
contrast their performance in relative to the balanced TSBN.
3.1
Comparing DTs with the balanced TSBN
We consider a 5-1ayer binary tree with 16 leaf nodes, as shown in Figure 1. Each node
in the tree is a binary variable, taking on values of white/black. The 7r1'S, M,'s and
affinities were set to be equal in each layer. The values used were 7r = (0.75,0.25)
with 0.75 referring to white, and M had values 0.99 on the diagonal and 0.01 offdiagonal. The affinities 3 were set as 1 for the natural parent, 0 for the nearest
neighbour(s) of the natural parent, -00 for non-nearest neighbours and anull = 0,
with f3 = 1.25.
"
!\I~ ~i
. ..
,
., ..,
/"
I
r
\
1\
"
\,'
,
\
...
-
-~,~~~~.~.~,~.~~~~.
(a) 5 black nodes
I
\
r
\1
,
I
I'
~,/
"
\
I
r
\
\
.
?? "..
?" " '.
"
\,,:
-
.
.... .
:
~
\'
'/
\
I
I
I
\~
(b) 4 black nodes
Figure 2: Plots of the unnormalised log posterior vs position of the input pattern
for (a) the 5-black-nodes pattern and (b) 4-black-nodes pattern.
To illustrate the effects of translation, we have taken a stimulus made up of a bar
of five black pixels, and moved it across the image. The unnormalised log posterior
for a particular Z configuration is logP(Z) + logP(XvIZ). This is computed for
the balanced TSBN architecture, and compared to the highest value that can be
found by conducting a search over Z. These results are plotted in Figure 2(a).
The x-axis denotes the position of the left hand end of the bar (running from 1 to
3The affinities are defined up to the addition of an arbitrary constant.
638
C. K. I. Williams and N. 1. Adams
12), and the y-axis shows the posterior probability. Note that due to symmetries
there are in reality fewer than 12 distinct configurations. Figure 2(a) shows clearly
that the balanced TSBN is a poor model for this stimulus, and that much better
interpretations can be found using DTs, even though the "natural parent" idea
ensures that the logP(Z) is always larger for the balanced tree.
Notice also how the balanced TSBN displays greater sensitivity of the log posterior
with respect to position than the DT model. Figure 2 shows both the "optimal"
log posterior (found "by hand", using intuitions as to the best trees), and the those
of the MAP models discovered by Simulated Annealing. Annealing was conducted
from a starting temperature of 1.0 and exponentially decreased by a factor of 0.9.
At each temperature up to 2000 proposals could be made, although transition to
the next temperature would occur after 200 accepted steps. The run was deemed to
have converged after five successive temperature steps were made without accepting
a single step. We also show the log posterior of trees found by Gibbs sampling from
which we report the best configuration found from four separate runs (with different
random starting positions), each of which was run for 25,000 sweeps through all of
the nodes .
In Figure 2(b) we have shown the log posterior for a stimulus made up of four black
nodes 4 . In this case the balanced TSBN is even more sensitive to the stimulus
location, as the four black nodes fit exactly under one sub-tree when they are
in positions 1, 5, 9 or 13. By contrast, the dynamic tree is less sensitive to the
alignment, although it does retain a preference for the configuration most favoured
by the balanced TSBN. This is due to the concept of a "natural" parent built into
the (current) architecture (but see Section 4 for further discussion) .
Clearly these results are somewhat sensitive to settings of the parameters. One of
the most important parameters is the diagonal entry in the CPT. This controls the
relative desirability of having a disconnection against a transition in the tree that
involves a colour change. For example, if the diagonal entry in the CPT is reduced
to 0.95, the gap between the optimal and balanced trees in Figure 2(b) is decreased.
We have experimented with CPT entries of 0.90,0.95 and 0.99, but otherwise have
. not needed to explore the parameter space to obtain the results shown.
3.2
Generating from the prior and finding the MAP Tree in 2-d
We now turn our attention to 2-d images. Considering a 5 layer quad-tree node
arrangement gives a total of 256 leaf nodes or a 16x16 pixel image. A structural
plot of such a tree generated from the prior is shown in figure 3.
Each sub-plot is a slice through the tree showing the nodes on successive levels.
The boxes represent a single node on the current level and their shading indicates
the tree to which they belong. Nodes in the parent layer above are superimposed
as circles and the lines emanating from them shows their connectivity. Black circles
with a smaller white circle inside are used to indicate root nodes. Thus in the
example above we see that the forest consists of five trees, four of whose roots lie
at level 3 (which between them account for most of the black in the image, Figure
3(f?, while the root node at level 1 is responsible for the background.
4The parameters are the same as above, except that
encourage disconnections at this level.
anull
in level 3 was set to 10.0 to
639
DTs: Dynamic Trees
(a)
(b)
(c)
(e)
(d)
(f)
Figure 3: Plot of the MAP Dynamic Tree of the accompanying image (f).
Broadly speaking the parameters for the 2-d DTs were set to be similar to the I-d
trees of the previous section, except that the disconnection affinities were set to
favour disconnections higher up the tree, and to values for the leaf level such that
leaf disconnection probabilities tend to zero. In practice this resulted in all leaves
being connected to parent nodes (which is desirable as we believe that single-pixel
objects are unlikely). The (3 values increase with tree depth so that lower levels
nodes choose parents from a tighter neighbourhood. The 7ft and M t values were
unchanged, and again we consider binary valued nodes.
A suite of 600 images were created by sampling DTs from the above prior and then
generating 5 images from each. Figure 3(f) shows an example of an image generated
by the DT and it can be seen that the "blockiness" exhibited by balanced TSBNs
is not present .
. ':. .
(a)
(b)
Figure 4: (a) Comparison of the MAP DT log posterior against that of the quad-tree
for 600 images, (b) tree generated from the "part-time-node-employment" prior.
640
C. K. I. Williams and N. J Adams
The MAP Dynamic Tree for each of these images was found by Simulated Annealing
using the same exponential strategy described earlier, and their log posteriors are
compared with those of the balanced TSBN in the plot 4(a). The line denotes the
boundary of equal log posterior and the location of all the points above this clearly
shows that in every case the MAP tree found has a higher posterior.
4
Discussion
Above we have demonstrated that DT models have greater translation invariance
and do not exhibit the blockiness of the balanced TSBN model. We also see that
Simulated Annealing methods are successful at finding trees that have high posterior
probability.
We now discuss some extensions to the model. In the work above we have kept the
balanced tree arrangement of nodes. However, this could be relaxed, giving rise to
roughly equal numbers of nodes at the various levels (cf stationary wavelets). This
would be useful (a) for providing better translation invariance and (b) to avoid
slight shortages of hidden units that can occur when patterns that are "misaligned"
wrt the balanced tree are presented. In this case the prior over Z would need to be
adjusted to ensure a high proportion of tree-like structures, by generating the z's
and x's in layers, so that the z's can be contingent on the states of the units in the
layer above. We have devised a prior of this nature and called it the "part-timeemployment" prior as the nodes can decide whether or not they wish to be employed
in the tree structure or remain redundant and inactive. An example tree generated
from this prior is shown in figure 4(b); we plan to explore this direction further
in on-going research. Other research directions include the learning of parameters
in the networks (e.g. using EM), and the introduction of additional information
at the nodes; for example one might use real-valued variables in addition to the
multinomial variables considered above. These additional variables might be used to
encode information such as that concerning the instantiation parameters of objects.
Acknowledgements
This work stems from a conversation between CW and Zoubin Gharahmani at the Isaac
Newton Institute in October 1997. We thank Zoubin Ghahramani, Geoff Hinton and Peter
Dayan for helpful conversations, and the Isaac Newton Institute for Mathematical Sciences
(Cambridge, UK) for hospitality during the "Neural Networks and Machine Learning" programme. NJA is supported by an EPSRC research studentship, and the work of CW is
partially supported by EPSRC grant GR/L03088, Combining Spatially Distributed Predictions From Neural Networks.
References
Bouman, C. A. and M. Shapiro (1994). A Multiscale Random Field Model for Bayesian
Image Segmentation. IEEE Transactions on Image Processing 3(2),162-177.
Dayan, P., G. E. Hinton, R. M. Neal, and R. S. Zemel (1995). The Helmholtz Machine.
Neural Computation 7(5)r 889-904.
Luettgen, M. R. and A. S. Willsky (1995). Likelihood Calculation for a Class of
Multiscale Stocahstic Models, with Application to Texture Discrimination. IEEE
Trans. Image Processing 4(2?, 194-207.
Pearl, J. (1988). Probabilistic Reasoning in Intelligent Systems: Networks of Plausible
Inference. San Mateo, CA: Morgan Kaufmann.
| 1535 |@word proportion:1 crucially:1 shading:1 configuration:4 current:2 comparing:2 readily:1 visible:1 designed:1 plot:5 v:1 stationary:1 discrimination:1 leaf:5 fewer:1 accepting:1 node:38 location:2 successive:2 preference:1 five:3 mathematical:1 become:1 consists:1 inside:1 introduce:2 roughly:1 multi:1 tsbn:16 quad:2 considering:1 mass:1 null:2 lowest:1 tying:1 kind:1 finding:3 suite:1 every:1 exactly:1 ensured:1 uk:5 control:1 unit:7 normally:1 grant:1 positive:1 xv:1 ak:1 black:10 might:2 mateo:1 specifying:1 misaligned:1 responsible:1 practice:1 xr:3 zoubin:3 context:1 www:1 restriction:1 map:9 demonstrated:1 williams:4 attention:1 starting:2 qh:1 exr:1 losing:1 helmholtz:1 bottom:1 ft:1 epsrc:2 ensures:1 connected:3 ordering:1 highest:2 balanced:19 intuition:2 dynamic:13 personal:1 employment:2 creation:1 technically:1 division:1 easily:1 geoff:1 various:1 distinct:1 effective:1 describe:1 emanating:1 zemel:1 whose:2 larger:1 valued:2 plausible:1 otherwise:1 grammar:1 think:1 advantage:1 net:2 product:1 combining:1 translate:1 moved:1 parent:16 r1:1 generating:4 adam:4 object:6 derive:1 illustrate:1 ac:3 nearest:2 involves:1 indicate:1 concentrate:1 direction:2 f1:1 tighter:1 adjusted:1 extension:1 accompanying:1 considered:2 applicable:2 sensitive:3 clearly:3 hospitality:1 always:1 aim:1 desirability:1 rather:2 avoid:1 tsbns:5 derived:1 encode:1 indicates:1 superimposed:1 likelihood:1 contrast:2 posteriori:1 inference:7 helpful:1 dayan:3 unlikely:1 fis:1 hidden:2 going:1 semantics:1 pixel:5 plan:1 marginal:1 field:3 construct:1 equal:3 having:2 f3:1 sampling:4 np:2 stimulus:4 report:1 intelligent:1 neighbour:2 resulted:1 individual:1 stationarity:1 possibility:2 alignment:1 blocky:3 capable:1 encourage:1 tree:66 indexed:1 circle:3 plotted:1 bouman:2 column:1 earlier:1 logp:3 entry:3 successful:1 conducted:2 gr:1 referring:1 ehi:1 sensitivity:1 retain:3 probabilistic:1 informatics:1 connectivity:1 again:1 luettgen:2 choose:2 ek:2 style:1 account:1 depends:1 crossed:1 cpts:2 root:8 picked:1 break:1 offdiagonal:1 kaufmann:1 conducting:1 bayesian:2 converged:1 ed:3 against:2 blockiness:2 isaac:2 associated:2 mi:3 popular:1 conversation:2 segmentation:3 higher:2 dt:10 ayer:1 arranged:1 evaluated:1 though:1 box:1 hand:3 parse:1 christopher:1 multiscale:2 lack:1 propagation:1 artifact:1 believe:2 effect:1 concept:1 true:1 assigned:1 spatially:1 neal:1 white:3 attractive:1 during:1 branching:1 hill:1 temperature:4 reasoning:1 image:26 multinomial:2 exponentially:1 belong:2 interpretation:2 slight:1 cambridge:1 gibbs:2 ai:1 had:1 posterior:20 belongs:1 binary:3 seen:1 morgan:1 dai:2 greater:2 somewhat:1 relaxed:1 employed:1 additional:2 contingent:1 redundant:1 ii:1 full:1 desirable:1 infer:1 stem:1 calculation:1 cross:1 devised:1 concerning:1 prediction:1 mrf:2 represent:1 lea:1 proposal:1 addition:2 background:1 annealing:6 decreased:2 exhibited:1 tend:1 call:2 structural:1 affect:1 fit:1 zi:5 architecture:7 idea:2 itr:1 favour:1 whether:1 inactive:1 colour:1 suffer:1 peter:1 speaking:1 cpt:4 useful:3 shortage:1 reduced:1 http:1 specifies:3 shapiro:2 generate:1 notice:1 broadly:1 shall:1 four:4 clean:1 kept:1 sum:1 run:3 named:1 decide:1 forrest:1 layer:14 display:1 occur:2 ri:1 structured:3 poor:1 remain:1 across:1 smaller:1 em:1 labellings:1 explained:1 taken:2 turn:1 discus:1 needed:1 wrt:1 end:1 ckiw:1 hierarchical:2 away:1 appropriate:1 nicholas:1 neighbourhood:1 alternative:1 denotes:3 top:1 running:1 cf:1 ensure:1 include:1 newton:2 giving:1 ghahramani:2 unchanged:1 sweep:1 objective:1 question:1 arrangement:2 occurs:1 strategy:2 usual:2 diagonal:4 exhibit:1 affinity:8 cw:2 link:1 separate:1 simulated:5 thank:1 participate:1 willsky:2 providing:1 ql:1 october:1 rise:1 markov:1 hinton:2 communication:1 discovered:1 arbitrary:1 namely:1 specified:3 sentence:1 connection:1 dts:19 pearl:3 trans:1 bar:2 pattern:4 built:1 belief:6 natural:6 indicator:1 scheme:2 axis:2 created:2 deemed:1 unnormalised:2 prior:18 acknowledgement:1 relative:3 par:1 expect:1 limitation:1 degree:1 translation:5 naked:1 supported:2 free:1 allow:1 disconnection:5 institute:3 taking:1 edinburgh:1 slice:1 overcome:1 depth:2 boundary:1 transition:2 studentship:1 distributed:1 author:1 made:4 adaptive:1 san:1 programme:1 far:1 transaction:1 approximate:2 instantiation:3 search:1 why:1 table:1 reality:1 qz:1 nature:1 nja:1 ca:1 obtaining:3 symmetry:1 forest:5 orphan:1 anc:1 investigated:1 constructing:1 main:1 child:4 x16:1 favoured:1 sub:2 position:6 wish:2 xh:1 exponential:1 lie:1 clamped:1 wavelet:1 showing:1 experimented:1 essential:1 texture:1 gap:1 led:1 simply:1 explore:2 partially:1 conditional:1 hard:2 change:1 included:1 determined:1 except:2 averaging:2 total:1 called:1 invariance:3 accepted:1 avoiding:1 ex:1 |
586 | 1,536 | Divisive Normalization, Line Attractor
Networks and Ideal Observers
Sophie Deneve l Alexandre Pougetl, and P.E. Latham 2
Institute for Computational and Cognitive Sciences,
Georgetown University, Washington, DC 20007-2197
2Dpt of Neurobiology, UCLA, Los Angeles, CA 90095-1763, U.S.A.
1 Georgetown
Abstract
Gain control by divisive inhibition, a.k.a. divisive normalization,
has been proposed to be a general mechanism throughout the visual cortex. We explore in this study the statistical properties
of this normalization in the presence of noise. Using simulations,
we show that divisive normalization is a close approximation to a
maximum likelihood estimator, which, in the context of population
coding, is the same as an ideal observer. We also demonstrate analytically that this is a general property of a large class of nonlinear
recurrent networks with line attractors. Our work suggests that
divisive normalization plays a critical role in noise filtering, and
that every cortical layer may be an ideal observer of the activity in
the preceding layer.
Information processing in the cortex is often formalized as a sequence of a linear
stages followed by a nonlinearity. In the visual cortex, the nonlinearity is best described by squaring combined with a divisive pooling of local activities. The divisive
part of the nonlinearity has been extensively studied by Heeger and colleagues [1],
and several authors have explored the role of this normalization in the computation
of high order visual features such as orientation of edges or first and second order
motion[ 4]. We show in this paper that divisive normalization can also playa role in
noise filtering. More specifically, we demonstrate through simulations that networks
implementing this normalization come close to performing maximum likelihood estimation. We then demonstrate analytically that the ability to perform maximum
likelihood estimation, and thus efficiently extract information from a population of
noisy neurons, is a property exhibited by a large class of networks.
Maximum likelihood estimation is a framework commonly used in the theory of
ideal observers. A recent example comes from the work of Itti et al., 1998, who have
shown that it is possible to account for the behavior of human subjects in simple
discrimination tasks. Their model comprised two distinct stages: 1) a network
105
Divisive Normalization. Line Attractor Networks and Ideal Observers
which models the noisy response of neurons with tuning curves to orientation and
spatial frequency combined with divisive normalization, and 2) an ideal observer (a
maximum likelihood estimator) to read out the population activity of the network.
Our work suggests that there is no need to distinguish between these two stages,
since, as we will show, divisive normalization comes close to providing a maximum
likelihood estimation. More generally, we propose that there may not be any part
of the cortex that acts as an ideal observer for patterns of activity in sensory areas
but, instead , that each cortical layer acts as an ideal observer of the activity in the
preceding layer.
1
The network
Our network is a simplified model of a cortical hypercolumn for spatial frequency
and orientation. It consists of a two dimensional array of units in which each unit
is indexed by its preferred orientation, 8i , and spatial frequency, >'j.
1.1
LGN model
Units in the cortical layer are assumed to receive direct inputs from the lateral
geniculate nucleus (LG N). Here we do not model explicitly the LG N, but focus
instead on the pooled LGN input onto each cortical unit . The input to each unit
is denoted aij' We distinguish between the mean pooled LGN input, fij(8, >'), as
a function of orientation, 8, and spatial frequency, >., and the noise distribution
around this mean, P(aijI8, >.).
In response to a stimulus of orientation, 8, spatial frequency, >., and contrast, G,
the mean LGN input onto unit ij is a circular Gaussian with a small amount of
spontaneous activity, 1/:
J'J,,(8 ,/\')
- KG
-
exp
(COS(>. - >'j) 2
~A
1+
cos(8 - 8i )
2
~o
-
1) +
1/,
(1)
where K is a constant. Note that spatial frequency is treated as a periodic variable;
this was done for convenience only and should have negligible effects on our results
as long as we keep>. far from 27m, n an integer.
On any given trial the LGN input to cortical unit ij, aij, is sampled from a Gaussian
noise distribution with variance ~;j:
(2)
In our simulations, the variance of the noise was either kept fixed (~'fj = ~2) or set
to the mean activity (~t = Jij(8, >.)). The latter is more consistent with the noise
that has been measured experimentally in the cortex. We show in figure I-A an
example of a noisy LGN pattern of activity.
1.2
Cortical Model: Divisive Normalization
Activities in the cortical layer are updated over time according to:
S. Deneve, A. Pouget and P. E. Latham
106
A.
CORTEX
:::::::: r:::L:: t ::;-:::;:::::
- r- , _ :-
:::~
'~
DO
0.1
0.2
0.3
0.4
0.5
0.8
0.7
0.1
D.'
1
Contrast
Figure 1: A- LGN input (bottom) and stable hill in the cortical network after
relaxation (top). The position of the stable hill can be used to estimate orientation
(0) and spatial frequency (5.). B- Inverse of the variance of the network estimate for
orientation using Gaussian noise with variance equal to the mean as a function of
contrast and number of iterations (0, dashed; 1, diamond; 2, circle; and 3, square).
The continuous curve corresponds to the theoretical upper bound on the inverse
of the variance (i.e. an ideal observer). C- Gain curve for contrast for the cortical
units after 1, 2 and 3 iterations.
(3)
where {Wij,kt} are the filtering weights, Oij(t) is the activity of unit ij at time t,
S is a constant, and J1. is what we call the divisive inhibition weight. The filtering
weights implement a two dimensional Gaussian filter:
Wij,kl
=
Wi-k,j - l
= Kwexp (COS[27!'(i
-2 k )/P]
~w~
where Kw is a constant,
there are p 2 units.
~w~
and
~WA
-1
+ cos[27!'(j ~ l)/Pl-1)
(4)
~WA
control the width of the filtering weights, and
On each iteration the activity is filtered by the weights, squared, and then normalized by the total local activity. Divisive normalization per se only involves the
squaring and division by local activity. We have added the filtering weights to obtain a local pooling of activity between cells with similar preferred orientations and
spatial frequencies. This pooling can easily be implemented with cortical lateral
connections and it is reasonable to think that such a pooling takes place in the
cortex.
Divisive Normalization, Line Attractor Networks and Ideal Observers
2
107
Simulation Results
Our simulations consist of iterating equation 3 with initial conditions determined by
the presentation orientation and spatial frequency. The initial conditions are chosen
as follows: For a given presentation angle, (}o, and spatial frequency, Ao, determine
the mean cortical activity, /ij((}o, AO), via equation 1. Then generate the actual
cortical activity, {aij}, by sampling from the distribution given in equation 2. This
serves as our set of initial conditions: Oij (t = 0) = aij'
Iterating equation 3 with the above initial conditions, we found that for very low
contrast the activity of all cortical units decayed to zero. Above some contrast
threshold, however, the activities converged to a smooth stable hill (see figure I-A
for an example with parameters (Jw(} = (Jw).. = (J(} = (J).. = I/V8, K = 74, C = 1,
J.L = 0.01). The width of the hill is controlled by the width of the filtering weights.
Its peak, on the other hand, depends on the orientation and spatial frequency of the
LGN input, (}o and Ao. The peak can thus be used to estimate these quantities (see
figure I-A). To compute the position of the final hill, we used a population vector
estimator [3] although any unbiased estimator would work as well. In all cases we
looked at, the network produced an unbiased estimate of (}o and Ao.
In our simulations we adjusted (Jw(} and (Jw).. so that the stable hill had the same
profile as the mean LGN input (equation 1). As a result, the tuning curves of the
cortical units match the tuning curves specified by the pooled LGN input. For this
case, we found that the estimate obtained from the network has a variance close
to the theoretical minimum, known as the Cramer-Rao bound [3]. For Gaussian
noise of fixed variance, the variance of the estimate was 16.6% above this bound,
compared to 3833% for the population vector applied directly to the LGN input.
In a ID network (orientation alone), these numbers go to 12.9% for the network
versus 613% for population vector. For Gaussian noise with variance proportional
to the mean, the network was 8.8% above the bound, compared to 722% for the
population vector applied directly to the input. These numbers are respectively 9%
and 108% for the I-D network. The network is therefore a close approximation to
a maximum likelihood estimator, i.e., it is close to being an ideal observer of the
LGN activity with respect to orientation and spatial frequency.
As long as the contrast, C, was superthreshold, large variations in contrast did not
affect our results (figure I-B). However, the tuning of the network units to contrast
after reaching the stable state was found to follow a step function whereas, for real
neurons, the curves are better described by a sigmoid [2]. Improved agreement
with experiment was achieved by taking only 2-3 iterations, at which point the
performance of the network is close to optimal (figure I-B) and the tuning curves to
contrast are more realistic and closer to sigmoids (figure I-C). Therefore, reaching
a stable state is not required for optimal performance, and in fact leads to contrast
tuning curves that are inconsistent with experiment.
3
Mathematical Analysis
We first prove that line attractor networks with sufficiently small noise are close
approximations to a maximum likelihood estimator. We then show how this result
applies to our simulations with divisive normalization.
S. Deneve, A. Pouget and P. E. Latham
J08
3.1
General Case: Line Attractor Networks
Let On be the activity vector (denoted by bold type) at discrete time, n, for a set
of P interconnected units. We consider a one dimensional network, i.e., only one
feature is encoded; generalization to multidimensional networks is straightforward.
A generic mapping for this network may be written
(5)
where H is a nonlinear function. We assume that this mapping admits a line
attractor, which we denote G(O), for which G(O) = H(G(O)) where 0 is a continuous
variable. 1 Let the initial state of the network be a function of the presentation
parameter, 00 , plus noise,
00
= F(Oo)
+N
(6)
where F(Oo) is the function used to generate the data (in our simulations this
would correspond to the mean LGN input, equation 1). Iterating the mapping,
equation 5, leads eventually to a point on the line attractor. Consequently, as
n -+ 00 , On -+ G(O) . The parameter 0 provides an estimate of 00 .
To determine how well the network does we need to find fJO :::: 0 - 00 as a function
of the noise, N, then average over the noise to compute the mean and variance of
fJO. Because the mapping, equation 5, is nonlinear, this cannot be done exactly. For
small noise, however, we can take a perturbative approach and expand around a
point on the attractor. For line at tractors there is no general method for choosing
which point on the attractor to expand around. Our approach will be to expand
around an arbitrary point, G( 0), and choose 0 by requiring that the quadratic terms
be finite. Keeping terms up to quadratic order, equation 6 may be written
G(O)
In .
+ fJo n .
fJo o +
(7)
n-l
I.: (Jm . fJo
~
o) .
H" . (J m . fJo o )
,
(8)
m=O
where J(O) == [8G (o)H(G(0))f is the Jacobian (the subscript T means transpose),
H" is the Hessian of H evaluated at G(O) and a "." represents the standard dot
product.
Because the mapping, equation 5, admits a line attractor , J has one eigenvalue
equal to 1 and all others less than 1. Denote the eigenvector with eigenvalue 1 as
y and its adjoint v t : J . v = v and JT . v t = yt. It is not hard to show that y =
8oG(0), up to a multiplicative constant. Since J has an eigenvalue equal to 1, to
avoid the quadratic term in Eq. 8 approaching infinity as n -+ 00 we require that
lim I
n-too
n .
fJo o = O.
(9)
IThe line attractor is, in fact , an idealization ; for P units the attractors associated with
equation 5 consists of P isolated points. However, for P large, the attractors are spaced
closely enough that they may be considered a line.
Divisive Normalization. Line Attractor Networks and Ideal Observers
109
This equations has an important consequence: it implies that, to linear order,
limn-too 60 n = 0 (see equation 8), which in turn implies that 0 00 = G(O) which,
~nally, implies that 0 = O. Consequently we can find the network estimator of 00 ,
0, by computing O. We now turn to that task.
It is straightforward to show that JOO = vv t . Combining this expression for J with
equation 9, using equation 7 to express 600 in terms of 00 and G(O), and, finally
using equation 6 to express 00 in terms of the initial mean activity, F(Oo), and the
noise, N, we find that
v t (0) . [F(Oo) - G(O)
Using 00
=0-
+ N] = O.
(10)
60 and expanding F(Oo) to first order in 60 then yields
60
= vt(O) . [N + F(O) -
G(O)]
vt(O) . F'(O)
.
(11)
As long as v t is orthogonal to F(O) - G(O), (60) = 0 and the estimator is unbiased.
This must be checked on a case by case basis, but for the circularly symmetric
networks we considered orthogonality is satisfied.
We can now calculate the variance of the network estimate, (60)2. Assuming v t .
[F(O) - G(O)] = 0, equation 11 implies that
2
vt.R?v t
(60) = [v t . F'F
'
(12)
where a prime denotes a derivative with respect to 0 and R is the covariance matrix
of the noise, R = (NN). The network is equivalent to maximum likelihood when
this variance is equal to the Cramer-Rao bound [3], (60)bR. If the noise, N, is
Gaussian with a covariance matrix independent of 0, this bound is equal to:
2
(60)CR
=
1
F'. R - l . F'
(13)
For independent Gaussian noise of fixed variance, (T2, and zero covariance, the
variance of the network estimate, equation 12, becomes (T2 1(IF'1 2 cos2 f-L) where f-L
is the angle between v t and F'. The Cramer-Rao bound, on the other hand, is
equal to (T2 IIF'1 2 . These expressions differ only by cos 2 J1., which is 1 if F ex v t . In
addition, it is close to 1 for networks that have identical input and output tunin
curves, F(O) = G(O), and the Jacobian, J, is nearly symmetric, so that v ::::: v
(recall that v = G'). If these last two conditions are satisfied, the network comes
close to being a maximum likelihood estimator.
1
3.2
Application to Divisive Normalization
Divisive normalization is a particular example of the general case considered above.
For simplicity, in our simulations we chose the input and output tuning curves to
be equal (F = G in the above notation), which lead to a value of 0.87 for cos2 f-L
(evaluated numerically). This predicted a variance 15% above the Cramer-Rao
S. Deneve, A. Pouget and P E. Latham
110
bound for independent Gaussian noise with fixed variance, consistent with the 16%
we obtained in our simulations. The network also handles fairly well other noise
distributions, such as Gaussian noise with variance proportional to the mean, as
illustrated by our simulations.
4
Conclusions
We have recently shown that a subclass of line attractor networks can be used as
maximum likelihood estimators[3]. This paper extend this conclusion to a much
wider class of networks, namely, any network that admits a line (or, by straightforward extension of the above analysis, a higher dimensional) attractor. This is true
in particular for networks using divisive normalization, a normalization which is
thought to match quite closely the nonlinearity found in the primary visual cortex
and MT.
Although our analysis relies on the existence of an attractor, this is not a requirement for obtaining near optimal noise filtering. As we have seen, 2-3 iterations
are enough to achieve asymptotic performance (except at contrasts barely above
threshold). What matters most is that our network implement a sequence of low
pass filtering to filter out the noise, followed by a square nonlinearity to compensate
for the widening of the tuning curve due to the low pass filter, and a normalization
to weaken contrast dependence. It is likely that this process would still clean up
noise efficiently in the first 2-3 iterations even if activity decayed to zero eventually,
that is to say, even if the hills of activity were not stable states. This would allow us
to apply our approach to other types of networks, including those lacking circular
symmetry and networks with continuously clamped inputs.
To conclude, we propose that each cortical layer may read out the activity in the
preceding layer in an optimal way thanks to the nonlinear pooling properties of
divisive normalization, and, as a result, may behave like an ideal observer. It is
therefore possible that the ability to read out neuronal codes in the sensory cortices
in an optimal way may not be confined to a few areas like the parietal or frontal
cortex, but may instead be a general property of every cortical layer.
References
[1] D. Heeger. Normalization of cell responses in cat striate cortex. Visual Neuroscience, 9:181- 197,1992.
[2] L. Itti, C. Koch, and J. Braun. A quantitative model for human spatial vision threshold on the basis of non-linear interactions among spatial filters. In
R. Lippman, J. Moody, and D. Touretzky, editors, Advances in Neural Information Processing Systems, volume 11. Morgan-Kaufmann, San Mateo, 1998.
[3] A. Pouget, K. Zhang, S. Deneve, and P. Latham. Statistically efficient estimation
using population coding. Neural Computation, 10:373- 401, 1998.
[4] E. Simoncelli and D. Heeger. A model of neuronal responses in visual area MT.
Vision Research, 38(5):743- 761 , 1998.
| 1536 |@word trial:1 cos2:2 simulation:11 covariance:3 initial:6 nally:1 perturbative:1 written:2 must:1 realistic:1 j1:2 discrimination:1 alone:1 filtered:1 provides:1 zhang:1 mathematical:1 direct:1 consists:2 prove:1 behavior:1 actual:1 jm:1 becomes:1 notation:1 what:2 kg:1 eigenvector:1 quantitative:1 every:2 multidimensional:1 act:2 subclass:1 braun:1 exactly:1 control:2 unit:15 negligible:1 local:4 consequence:1 id:1 subscript:1 plus:1 chose:1 studied:1 mateo:1 suggests:2 co:5 statistically:1 implement:2 lippman:1 area:3 thought:1 onto:2 close:10 convenience:1 cannot:1 context:1 equivalent:1 yt:1 go:1 straightforward:3 formalized:1 simplicity:1 pouget:4 estimator:10 array:1 population:8 handle:1 variation:1 updated:1 spontaneous:1 play:1 agreement:1 fjo:7 bottom:1 role:3 calculate:1 ithe:1 division:1 basis:2 easily:1 cat:1 distinct:1 choosing:1 quite:1 encoded:1 say:1 ability:2 think:1 noisy:3 final:1 sequence:2 eigenvalue:3 propose:2 interconnected:1 jij:1 product:1 interaction:1 combining:1 achieve:1 adjoint:1 los:1 requirement:1 wider:1 oo:5 recurrent:1 measured:1 ij:4 eq:1 implemented:1 predicted:1 involves:1 come:4 implies:4 differ:1 fij:1 closely:2 filter:4 human:2 implementing:1 require:1 ao:4 generalization:1 adjusted:1 extension:1 pl:1 around:4 cramer:4 sufficiently:1 considered:3 exp:1 koch:1 mapping:5 estimation:5 geniculate:1 gaussian:10 reaching:2 avoid:1 cr:1 og:1 focus:1 likelihood:11 contrast:13 squaring:2 nn:1 expand:3 wij:2 lgn:13 among:1 orientation:13 denoted:2 spatial:14 fairly:1 equal:7 washington:1 sampling:1 iif:1 kw:1 represents:1 identical:1 nearly:1 others:1 stimulus:1 t2:3 few:1 attractor:18 circular:2 kt:1 edge:1 closer:1 orthogonal:1 indexed:1 circle:1 isolated:1 theoretical:2 weaken:1 rao:4 comprised:1 too:2 periodic:1 combined:2 thanks:1 decayed:2 peak:2 continuously:1 moody:1 squared:1 satisfied:2 choose:1 cognitive:1 derivative:1 itti:2 account:1 coding:2 pooled:3 bold:1 matter:1 explicitly:1 depends:1 multiplicative:1 observer:13 square:2 variance:17 who:1 efficiently:2 kaufmann:1 correspond:1 spaced:1 yield:1 produced:1 converged:1 touretzky:1 checked:1 colleague:1 frequency:12 associated:1 gain:2 sampled:1 recall:1 lim:1 alexandre:1 higher:1 follow:1 response:4 improved:1 jw:4 done:2 evaluated:2 stage:3 hand:2 nonlinear:4 effect:1 normalized:1 unbiased:3 true:1 requiring:1 analytically:2 read:3 symmetric:2 illustrated:1 width:3 hill:7 demonstrate:3 latham:5 motion:1 fj:1 recently:1 sigmoid:1 mt:2 volume:1 extend:1 numerically:1 tuning:8 nonlinearity:5 had:1 dot:1 joo:1 stable:7 cortex:11 inhibition:2 playa:1 recent:1 prime:1 vt:3 seen:1 minimum:1 morgan:1 preceding:3 determine:2 dashed:1 simoncelli:1 smooth:1 match:2 long:3 compensate:1 controlled:1 vision:2 iteration:6 normalization:23 achieved:1 cell:2 confined:1 receive:1 whereas:1 addition:1 limn:1 exhibited:1 pooling:5 subject:1 inconsistent:1 integer:1 call:1 j08:1 near:1 presence:1 ideal:13 enough:2 affect:1 approaching:1 br:1 angeles:1 expression:2 hessian:1 v8:1 generally:1 iterating:3 se:1 amount:1 extensively:1 generate:2 neuroscience:1 per:1 discrete:1 express:2 threshold:3 clean:1 kept:1 deneve:5 relaxation:1 idealization:1 inverse:2 angle:2 place:1 throughout:1 reasonable:1 layer:9 bound:8 followed:2 distinguish:2 quadratic:3 activity:24 orthogonality:1 infinity:1 ucla:1 performing:1 according:1 wi:1 equation:18 turn:2 eventually:2 mechanism:1 serf:1 apply:1 generic:1 existence:1 dpt:1 top:1 denotes:1 added:1 quantity:1 looked:1 primary:1 dependence:1 striate:1 lateral:2 barely:1 assuming:1 code:1 providing:1 lg:2 perform:1 diamond:1 upper:1 neuron:3 finite:1 behave:1 parietal:1 neurobiology:1 dc:1 arbitrary:1 namely:1 hypercolumn:1 kl:1 specified:1 connection:1 required:1 pattern:2 including:1 critical:1 treated:1 widening:1 oij:2 extract:1 georgetown:2 tractor:1 asymptotic:1 lacking:1 filtering:9 proportional:2 versus:1 nucleus:1 consistent:2 editor:1 last:1 keeping:1 transpose:1 aij:4 allow:1 vv:1 institute:1 taking:1 curve:11 cortical:17 sensory:2 author:1 commonly:1 san:1 simplified:1 far:1 preferred:2 keep:1 assumed:1 conclude:1 continuous:2 ca:1 expanding:1 obtaining:1 symmetry:1 did:1 noise:25 profile:1 neuronal:2 position:2 heeger:3 clamped:1 jacobian:2 jt:1 explored:1 admits:3 consist:1 circularly:1 sigmoids:1 explore:1 likely:1 visual:6 applies:1 corresponds:1 relies:1 presentation:3 consequently:2 experimentally:1 hard:1 specifically:1 determined:1 except:1 sophie:1 total:1 pas:2 divisive:21 latter:1 frontal:1 ex:1 |
587 | 1,537 | Maximum Conditional Likelihood via
Bound Maximization and the CEM
Algorithm
Tony J ebara and Alex Pentland
Vision and Modeling, MIT Media Laboratory, Cambridge MA
http://www.rnedia.rnit.edu/ ~ jebara
{ jebara,sandy }~rnedia.rnit.edu
Abstract
We present the CEM (Conditional Expectation Maximi::ation) algorithm as an extension of the EM (Expectation M aximi::ation)
algorithm to conditional density estimation under missing data. A
bounding and maximization process is given to specifically optimize
conditional likelihood instead of the usual joint likelihood. We apply the method to conditioned mixture models and use bounding
techniques to derive the model's update rules . Monotonic convergence, computational efficiency and regression results superior to
EM are demonstrated.
1
Introduction
Conditional densities have played an important role in statistics and their merits
over joint density models have been debated. Advantages in feature selection , robustness and limited resource allocation have been studied. Ultimately, tasks such
as regression and classification reduce to the evaluation of a conditional density.
However, popularity of maximumjoint likelihood and EM techniques remains strong
in part due to their elegance and convergence properties . Thus , many conditional
problems are solved by first estimating joint models then conditioning them . This
results in concise solutions such as the N adarya- Watson estimator [2], Xu's mixture
of experts [7], and Amari's em-neural networks [1]. However, direct conditional
density approaches [2, 4] can offer solutions with higher conditional likelihood on
test data than their joint counter-parts.
495
Maximum Conditional Likelihood via Bound Maximization and CEM
1~~1
01
'"
(a) La
= -4.2
L;
!i.
= -2.4
15
(b) Lb = -5 .2 Lb
20
= -1.8
Figure 1: Average Joint (x, y) vs. Conditional (ylx) Likelihood Visualization
Pop at [6] describes a simple visualization example where 4 clusters must be fit with
2 Gaussian models as in Figure 1. Here, the model in (a) has a superior joint likelihood (La> Lb) and hence a better p(x, y) solution. However, when the models are
conditioned to estimate p(ylx), model (b) is superior (Lb > L~). Model (a) yields
a poor unimodal conditional density in y and (b) yields a bi-modal conditional
density. It is therefore of interest to directly optimize conditional models using conditionallikelihood. We introduce the CEM (Conditional Expectation Maximization)
algorithm for this purpose and apply it to the case of Gaussian mixture models.
2
EM and Conditional Likelihood
For joint densities, the tried and true EM algorithm [3] maximizes joint likelihood
over data. However , EM is not as useful when applied to conditional density estimation and maximum conditional likelihood problems. Here, one typically resorts to
other local optimization techniques such as gradient descent or second order Hessian
methods [2]. We therefore introduce CEM, a variant of EM , which targets conditional likelihood while maintaining desirable convergence properties. The CEM
algorithm operates by directly bounding and decoupling conditional likelihood and
simplifies M-step calculations.
In EM, a complex density optimization is broken down into a two-step iteration
using the notion of missing data. The unknown data components are estimated via
the E-step and a simplified maximization over complete data is done in the M-step.
In more practical terms, EM is a bound maximization : the E-step finds a lower
bound for the likelihood and the M-step maximizes the bound.
M
P(Xi, Yi18)
=L
p(m, Xi, Yi1 8 )
(1)
m=l
Consider a complex joint density p(Xi , Yi 18) which is best described by a discrete
(or continuous) summation of simpler models (Equation 1) . Summation is over the
'missing components' m .
L:~llog(P(Xi' Yi 18 t ?) -Iog(p(xi ' YiI 8t - 1 ))
l:!.l
>
",\,N
L...i=l
h I
p(m ,X. ,Y.le t )
L...m=l im og p(m,X"Y.let I)
",\,M
h
h
p(m,x. ,y.le t- I )
were im = ",\,M
L...,,=I p(n ,x"y .Ie t- I )
By appealing to Jensen's inequality, EM obtains a lower bound for the incremental
log-likelihood over a data set (Equation 2) . Jensen's inequality bounds the logarithm of the sum and the result is that the logarithm is applied to each simple
(2)
T. Jebara and A. Pentland
496
model p(m, Xi , yd8) individually. It then becomes straightforward t.o compute the
derivatives with respect to e and set to zero for maximization (M-step) .
AJ
,\"M
"
I 8) = =c""-7
L...m=IP(m,xi,y;j8)
p( Yi IXi, 8)
- = L..: p( m , Y i Xi , M+------'-----'-
(3)
Lm=IP(m,XiI 8 )
m=l
However, the elegance of EM is compromised when we consider a conditioned density
as in Equation :3. The corresponding incremental conditional log-likelihood, L:l.lc, is
shown in Equation 4.
L~llog(p(Yilxi' 8 t )) -log(p(ydxi, 8 t I
L....
og
!=1
,\"N
LMm_I P(m ,X. ,Y.10)t
H
L;"=I p(m ,X. ,Y.10 t -
l )
-I
og
1)
LMn_IP(n ,X, I0 t)
(4)
M
Ln=1 p(n ,X.10 t -
l )
The above is a difference between a ratio of joints and a ratio of marginals. If
Jensen's inequality is applied to the second term in Equation 4 it yields an upper
bound since the term is subtracted (this would compromise convergence). Thus,
only the first ratio can be lower bounded with Jensen (Equation 5).
L:l.jC>~~h'
I
p(m,xi,YiI 8t ) -I
- L..: L..: 2m og (
18 t - 1) og
i=1 m=1
p m, Xi, Yi -
L~~lp(n,xiI8t)
M
Ln=1 p(n, XiI8 t - 1)
(5)
Note the lingering logarithm of a sum which prevents a simple M-Step. At this point,
one would resort to a Generalized EM (GEM) approach which requires gradient or
second-order ascent techniques for the M-step. For example, Jordan et al. overcome
the difficult M-step caused by EM with an Iteratively Re- Weighted Least Squares
algorithm in the mixtures of experts architecture [4].
3
Conditional Expectation Maximization
The EM algorithm can be extended by substituting Jensen's inequality for a different bound. Consider the upper variational bound of a logarithm x-I 2: log(x)
(which becomes a lower bound on the negative log). The proposed logarithm's
bound satisfies a number of desiderata: (1) it makes contact at the current operating point 1, (2) it is tangential to the logarithm, (3) it is a tight bound, (4)
it is simple and (5) it is the variational dual of the logarithm. Substituting this
linear bound into the incremental conditional log-likelihood maintains a true lower
bounding function Q (Equation 6).
The Mixture of Experts formalism [4J offers a graceful representation of a conditional
density using experts (conditional sub-models) and gates (marginal sub-models).
The Q function adopts this form in Equation 7.
1 The
current operating point is 1 since the
previous iteration's value e t - 1 .
e
t
model in the ratio is held fixed at the
Maximum Conditional Likelihood via Bound Maximization and CEM
L~l L~=1 {him(logp(Yilm,Xi,e t ) +logp(m,xile t )
where Zim = log(p(m,xi,Yile t -- 1 ))
497
- Zim) - riP(m , xile) +
and ri = (L~=1p(n'Xilet-1) )-1
ir}
(7)
Computing this Q function forms the CE-step in the Conditional Expectation Maximization algorithm and it results in a simplified M-step. Note the absence of the
logarithm of a sum and the decoupled models. The form here allows a more straightforward computation of derivatives with respect to e t and a more tractable M-Step.
For continuous missing data, a similar derivation holds.
At this point , without loss of generality, we specifically attend to the case of a conditioned Gaussian mixture model and derive the corresponding M-Step calculations.
This serves as an implementation example for comparison purposes.
4
CEM and Bound Maxinlization for Gaussian Mixtures
In deriving an efficient M-step for the mixture of Gaussians, we call upon more
bounding techniques that follow the CE-step and provide a monotonically convergent learning algori thm . The form ofthe condi tional model we will train is obtained
by conditioning a joint mixture of Gaussians. We write the conditional density
in a experts-gates form as in Equation 8. We use unnormalized Gaussian gates
N(x; p,~) = exp( - ~(x - p)T~-1 (x - p? since conditional models do not require
true marginal densities over x (i .e. that necessarily integrate to 1). Also, note that
the parameters of the gates (0:' , px , :E xx ) are independent of the parameters of the
experts (vm,rm,om).
Both gates and experts are optimized independently and have no variables in common. An update is performed over the experts and then over the gates. If each
of those causes an increase, we converge to a local maximum of conditional loglikelihood (as in Expectation Conditional Maximization [5]).
p(Ylx,8)
(8)
To update the experts , we hold the gates fixed and merely take derivatives of the Q
function with respect to the expert parameters (<l>m = {v m , rm, am} ) and set them
to O. Each expert is effectively decoupled from other terms (gates, other experts ,
etc .). The solution reduces to maximizing the log of a single conditioned Gaussian
and is analytically straightforward.
8Q(e t ,e(t-l?)
8<1>'"
(9)
Similarly, to update the gate mixing proportions, derivatives of the Q function are
taken with respect to O:'m and set to O. By holding the other parameters fixed , the
update equation for the mixing proportions is numerically evaluated (Equation 10).
N
N
O:'m := LriN(xi;P~,:E~x) le(l-I) {Lhim}-l
i=l
i=l
(10)
T. Jebara and A. Pentland
498
,
c,
O~
01
I
~
r
01
I
i
\
-
' .,
i
\,
',_
00
_1
( a)
... :
~, ,>/~
\
\
,f
at
//---~
',,---,
,1 \
f Function
-2
..!..:
_I
':li
:~
c'
-----
00
I
I?Ji
'--...
01'
I
~
''''\
I ~ jr' --...-.;==i
-J
j--- - -
' - ,- - 10
Ii
"
(b) Bound on f..L
( c) g Function
(d) Bound on I: xx
Figure 2: Bound Width Computation and Example Bounds
4.1
Bounding Gate Means
Taking derivatives of Q and setting to a is not as straightforward for the case of
the gate means (even though they are decoupled). What is desired is a simple
update rule (i.e. computing an empirical mean). Therefore, we further bound the
Q function for the M-step. The Q function is actually a summation of sub-elements
Qim and we bound it instead by a summation of quadratic functions on the means
(Equation 11).
N
Q(e t , e(t-1)) =
M
N
LL
i=l
Q(e t , e(t-1))im >
m=l
M
LL
kim - Wimllf..L~ - ciml1 2 (11)
i=l m=l
Each quadratic bound has a location parameter
cim
(a centroid), a scale parameter
Wim (narrowness), and a peak value at kim. The sum of quadratic bounds makes
e
contact with the Q function at the old values of the model t - 1 where the gate
mean was originally f..L':* and the covariance is I:':x*' To facilitate the derivation,
one may assume that the previous mean was zero and the covariance was identity
if the data is appropriately whitened with respect to a given gate.
The parameters of each quadratic bound are solved by ensuring that it contacts the
corresponding Qim function at t - 1 and they have equal derivatives at contact (i .e.
tangential contact) . Sol ving these constraints yields quadratic parameters for each
gate m and data point i in Equation 12 (kim is omitted for brevity) .
e
(12)
>
The tightest quadratic bound occurs when Wim is minimal (without violating the
inequality). The expression for Wim reduces to finding the minimal value, wim, as in
Equation 13 (here p2 = xT xd. The f function is computed numerically only once
and stored as a lookup table (see Figure 2(a)). We thus immediately compute the
optimal wim and the rest of the quadratic bound's parameters obtaining bounds as
in Figure 2(b) where a Qim is lower bounded.
1
* -_
Wim
.
rlCl'm
max
C
{ e-
1
-p
2
2
2
e- 2 C e CP
2
c
cp - 1 }
'
h?1m _
+- -
2
,
.
r l Cl'm e
1
- -p
2
2
h?1m
f( p) + - (13)
2
The gate means f..L~ are solved by maximizing the sum of the M x N parabolas which
bound Q. The update is f..L': = (2: wimCim) (2: wim)-l. This mean is subsequently
unwhitened to undo earlier data transformations.
Maximum Conditional Likelihood via Bound Maximization and CEM
499
[
,~,
:l
-1
"
~
0'-,
?. i
.
..... :
-
J ... .
(a) Data
(b) CEM p(ylx)
(c) CEM IC
(d) EM fit
(e) EM p(ylx)
(f) EM I C
Figure 3: Conditional Density Estimation for CEM and EM
4.2
Bounding Gate Covariances
Having derived the update equation for gate means, we now turn our attention
to the gate covariances. We bound the Q function with logarithms of Gaussians.
Maximizing this bound (a sum of log-Gaussians) reduces to the maximum-likelihood
estimation of a covariance matrix . The bound for a Qim sub-component is shown
in Equation 14. Once again, we assume the data has been appropriately whitened
with respect to the gate's previous parameters (the gate's previous mean is 0 and
previous covariance is identity). Equation 15 solves for the log-Gaussian parameters
(again p2 = XTXi).
Q( Dt,D(t-1));m > Iog (N) = k im
QQ.
_
-
",m
T
-1 Cim WimCimL..xx
Wim
Iog I",m
L.. xx I (14)
(15)
>
The computation for the minimal Wim simplifies to wim = riQ:mg(p) . The 9 function
is derived and plotted in Figure 2(c). An example of a log-Gaussian bound is
shown in Figure 2( d) a sub-component of the Q function. Each sub-component
corresponds to a single data point as we vary one gate 's covariance. All M x N
log-Gaussian bounds are computed (one for each data point and gate combination)
and are summed to bound the Q function in its entirety.
To obtain a final answer for the update of the gate covariances E~ we simply
maximize the sum of log Gaussians (parametrized by wim, kim, Cim). The update is
E~x = (2: WimCimCim T) (2: wim)-l. This covariance is subsequently unwhitened ,
inverting the whitening transform applied to the data.
5
]
'~ - -- -:----"
!'
'
1
Results
The CEM algorithm updates the conditioned mixture of Gaussians by computing
him and rim in the CE steps and interlaces these with updates on the experts,
mixing proportions, gate means and gate covariances. For the mixture of Gaussians ,
each CEM update has a computation time that is comparable with that of an EM
update (even for high dimensions). However, conditional likelihood (not joint) is
monotonically increased .
Consider the 4-cluster (x , y) data in Figure 3(a). The data is modeled with a conditional density p(ylx) using only 2 Gaussian models . Estimating the density with
CEM yields the p(ylx) shown in Figure 3(b). CEM exhibits monotonic conditional
likelihood growth (Figure 3(c)) and obtains a more conditionally likely model. In
T. Jebara and A. Pentland
500
Algorithm
Abalone
Table 1: Test Results. Class label regression accuracy data. (CNNO=cascadecorrelation, hidden units, CCN5=5 hidden LD=linear discriminant).
a
the EM case, a joint p(x, y) clusters the data as in Figure 3(d) . Conditioning it
yields the p(ylx) in Figure 3(e) . Figure 3(f) depicts EM's non-monotonic evolution
of conditional log-likelihood. EM produces a superior joint likelihood but an inferior conditional likelihood. Note how the CEM algorithm utilized limited resources
to capture the multimodal nature of the distribution in y and ignored spurious bimodal clustering in the x feature space. These properties are critical for a good
conditional density p(ylx).
For comparison , standard databases were used from DCI 2. Mixture models were
trained with EM and CEM , maximizingjoint and conditional likelihood respectively.
Regression results are shown in Table 1. CEM exhibited , monotonic conditional loglikelihood growth and out-performed other methods including EM with the same
2-Gaussian model (EM2 and CEM2).
6
Discussion
We have demonstrated a variant of EM called CEM which optimizes conditional
likelihood efficiently and monotonically. The application of CEM and bound maximization to a mixture of Gaussians exhibited promising results and better regression
than EM . In other work , a MAP framework with various priors and a deterministic
annealing approach have been formulated. Applications of the CEM algorithm to
non-linear regressor experts and hidden Markov models are currently being investigated . Nevertheless , many applications CEM remain to be explored and hopefully
others will be motivated to extend the initial results .
Acknowledgements
Many thanks to Michael Jordan and Kris Popat for insightful discussions.
References
[1] S. Amari. Information geometry of em and em algorithms for neural networks. Neural
Networks , 8(9), 1995 .
[2 ] C. Bishop. Neural Networks Jor Pattern Recognition. Oxford Press, 1996.
3
[ ] A. Dempster, N. Laird, and D. Rubin. Maximum likelihood from incomplete data via
the em algorithm. Journal oj the Royal Statistical Society, B39, 1977.
[4] M. Jordan and R. Jacobs. Hierarchical mixtures of experts and the em algorithm .
Neural Computation, 6:181 - 214, 1994.
[5] X. Meng and D. Rubin. Maximum likelihood estimation via the ecm algorithm : A
general framework. Biometrika , 80(2), 1993.
[6] A . Popat. Conjoint probabilistic subband modeling (phd. thesis). Technical Report
461, M.LT. Media Laboratory, 1997.
[7] 1. Xu , M. Jordan , and G. Hinton . An alternative model for mixtures of experts . In
Neural InJormation Processing Systems 7, 1995.
2http://www.ics.uci.edu/'''-'mlearn/MLRepository.html
| 1537 |@word proportion:3 tried:1 covariance:10 b39:1 jacob:1 concise:1 ld:1 initial:1 current:2 must:1 update:14 v:1 yi1:1 location:1 simpler:1 direct:1 introduce:2 becomes:2 estimating:2 bounded:2 xx:4 maximizes:2 medium:2 what:1 finding:1 transformation:1 xd:1 growth:2 biometrika:1 rm:2 unit:1 attend:1 local:2 oxford:1 meng:1 studied:1 limited:2 bi:1 practical:1 empirical:1 ving:1 selection:1 www:2 optimize:2 map:1 demonstrated:2 missing:4 maximizing:3 deterministic:1 straightforward:4 attention:1 independently:1 immediately:1 rule:2 estimator:1 deriving:1 notion:1 qq:1 target:1 rip:1 ixi:1 element:1 recognition:1 utilized:1 parabola:1 database:1 role:1 solved:3 capture:1 counter:1 sol:1 xile:2 dempster:1 broken:1 ultimately:1 trained:1 tight:1 compromise:1 upon:1 efficiency:1 multimodal:1 joint:14 various:1 derivation:2 train:1 loglikelihood:2 qim:4 amari:2 statistic:1 transform:1 laird:1 ip:2 final:1 injormation:1 advantage:1 mg:1 yii:2 uci:1 mixing:3 convergence:4 cluster:3 produce:1 incremental:3 derive:2 p2:2 solves:1 strong:1 entirety:1 subsequently:2 require:1 summation:4 im:4 extension:1 hold:2 ic:2 exp:1 lm:1 substituting:2 yilxi:1 vary:1 sandy:1 omitted:1 purpose:2 estimation:5 label:1 currently:1 wim:12 individually:1 him:2 weighted:1 mit:1 gaussian:11 ecm:1 og:5 derived:2 zim:2 likelihood:29 centroid:1 kim:4 am:1 tional:1 i0:1 typically:1 hidden:3 spurious:1 classification:1 dual:1 html:1 summed:1 marginal:2 equal:1 once:2 having:1 others:1 report:1 tangential:2 geometry:1 interest:1 evaluation:1 mixture:15 held:1 decoupled:3 incomplete:1 old:1 logarithm:9 re:1 desired:1 plotted:1 minimal:3 increased:1 formalism:1 modeling:2 earlier:1 logp:2 maximization:13 stored:1 answer:1 thanks:1 density:19 peak:1 ie:1 probabilistic:1 vm:1 regressor:1 michael:1 again:2 rlcl:1 algori:1 thesis:1 expert:16 resort:2 derivative:6 li:1 lookup:1 jc:1 caused:1 maximumjoint:1 performed:2 maintains:1 om:1 square:1 ir:1 accuracy:1 efficiently:1 yield:6 ofthe:1 kris:1 mlearn:1 elegance:2 rim:1 actually:1 higher:1 originally:1 violating:1 follow:1 dt:1 modal:1 done:1 evaluated:1 though:1 generality:1 hopefully:1 aj:1 facilitate:1 true:3 evolution:1 hence:1 analytically:1 laboratory:2 iteratively:1 conditionally:1 ll:2 width:1 inferior:1 unnormalized:1 abalone:1 mlrepository:1 generalized:1 complete:1 cp:2 variational:2 superior:4 common:1 ji:1 rnit:2 conditioning:3 extend:1 marginals:1 numerically:2 cambridge:1 similarly:1 operating:2 whitening:1 etc:1 ebara:1 optimizes:1 inequality:5 watson:1 yi:4 converge:1 maximize:1 monotonically:3 ii:1 unimodal:1 desirable:1 reduces:3 technical:1 calculation:2 offer:2 iog:3 ensuring:1 variant:2 regression:5 desideratum:1 whitened:2 vision:1 expectation:6 iteration:2 bimodal:1 condi:1 annealing:1 appropriately:2 rest:1 exhibited:2 ascent:1 undo:1 jordan:4 call:1 fit:2 architecture:1 reduce:1 simplifies:2 expression:1 motivated:1 hessian:1 cause:1 ignored:1 useful:1 ylx:9 http:2 estimated:1 popularity:1 xii:1 discrete:1 write:1 nevertheless:1 ce:3 merely:1 sum:7 ydxi:1 comparable:1 bound:37 played:1 convergent:1 quadratic:7 constraint:1 alex:1 ri:1 graceful:1 px:1 combination:1 poor:1 jr:1 describes:1 cascadecorrelation:1 em:31 remain:1 appealing:1 lp:1 taken:1 ln:2 resource:2 visualization:2 remains:1 equation:17 turn:1 merit:1 tractable:1 serf:1 gaussians:8 tightest:1 apply:2 hierarchical:1 subtracted:1 alternative:1 robustness:1 gate:25 clustering:1 tony:1 maintaining:1 subband:1 society:1 contact:5 occurs:1 usual:1 exhibit:1 gradient:2 parametrized:1 discriminant:1 modeled:1 ratio:4 difficult:1 holding:1 dci:1 negative:1 implementation:1 unknown:1 upper:2 markov:1 descent:1 pentland:4 extended:1 hinton:1 lb:4 thm:1 jebara:5 xtxi:1 inverting:1 optimized:1 pop:1 pattern:1 max:1 including:1 oj:1 royal:1 critical:1 ation:2 cim:3 prior:1 acknowledgement:1 loss:1 j8:1 allocation:1 conjoint:1 integrate:1 rubin:2 taking:1 overcome:1 dimension:1 jor:1 adopts:1 simplified:2 obtains:2 em2:1 cem:23 gem:1 xi:13 continuous:2 compromised:1 table:3 promising:1 nature:1 decoupling:1 obtaining:1 investigated:1 complex:2 necessarily:1 cl:1 bounding:7 xu:2 depicts:1 lc:1 sub:6 debated:1 down:1 xt:1 bishop:1 insightful:1 jensen:5 popat:2 explored:1 effectively:1 phd:1 conditioned:6 unwhitened:2 lt:1 simply:1 likely:1 prevents:1 monotonic:4 corresponds:1 satisfies:1 ma:1 conditional:42 identity:2 formulated:1 absence:1 specifically:2 operates:1 llog:2 called:1 la:2 brevity:1 |
588 | 1,538 | Robot Docking using Mixtures of Gaussians
Matthew Williamson*
Roderick Murray-Smith t
Volker Hansen t
Abstract
This paper applies the Mixture of Gaussians probabilistic model, combined with Expectation Maximization optimization to the task of summarizing three dimensional range data for a mobile robot. This provides
a flexible way of dealing with uncertainties in sensor information, and allows the introduction of prior knowledge into low-level perception modules. Problems with the basic approach were solved in several ways: the
mixture of Gaussians was reparameterized to reflect the types of objects
expected in the scene, and priors on model parameters were included
in the optimization process. Both approaches force the optimization to
find 'interesting' objects, given the sensor and object characteristics. A
higher level classifier was used to interpret the results provided by the
model, and to reject spurious solutions.
1 Introduction
This paper concerns an application of the Mixture of Gaussians (MoG) probabilistic model
(Titterington et aI., 1985) for a robot docking application. We use the ExpectationMaximization (EM) approach (Dempster et aI., 1977) to fit Gaussian sub-models to a sparse
3d representation of the robot's environment, finding walls, boxes, etc .. We have modified
the MoG formulation in three ways to incorporate prior knowledge about the task, and the
sensor characteristics: the parameters of the Gaussians are recast to constrain how they fit
the data, priors on these parameters are calculated and incorporated into the EM algorithm,
and a higher level processing stage is included which interprets the fit of the Gaussians on
the data, detects misclassifications, and providing prior information to guide the modelfitting.
The robot is equipped with a LIDAR 3d laser range-finder (PIAP, 1995) which it uses to
identify possible docking objects. The range-finder calculates the time of flight for a light
pulse reflected off objects in the scene. The particular LIDAR used is not very powerful,
making objects with poor reflectance (e.g., dark, shiny, or surfaces not perpendicular to the
*Corresponding author: MIT AI Lab, Cambridge, MA, USA. rna t t@ai . rni t . edu
tDept. of Mathematical Modelling, Technical University of Denmark. rod@imm. dtu. dk
tDaimlerChrysler, Alt-Moabit 96a, Berlin, Germany. hansen@dbag.bIn. dairnierbenz . com
946
M M Williamson, R. Murray-Smith and V. Hansen
laser beam) invisible. The scan pattern is also very sparse, especially in the vertical direction, as shown in the scan of a wall in Figure 1. However, if an object is detected, the range
returned is accurate (?1-2cm). When the range data is plotted in Cartesian space it forms
a number of sparse clusters, leading naturally to the use of MoG clustering algorithms to
make sense of the scene. While the Gaussian assumption is not an ideal model of the data,
the generality of MoG, and its ease of implementation and analysis motivated its use over a
more specialized approach. The sparse nature of the data inspired the modifications to the
MoG formulation described in this paper.
Model-based object recognition from dense range images has been widely reported (see
(Arman and Aggarwal, 1993) for a review), but is not relevant in this case given the sparseness of the data. Denser range images could be collected by combining multiple scans, but
the poor visibility of the sensor hampers the application of these techniques. The advantage
of the MoG technique is that the segmentation is "soft", and perception proceeds iteratively
during learning. This is especially useful for mobile robots where evidence accumulates
over time, and the allocation of attention is time and state-dependent. The EM algorithm is
useful since it is guaranteed to converge to a local maximum.
The following sections of the paper describe the re-parameterization of the Gaussians to
model plane-like clusters, the formulation of the priors, and the higher level processing
which interprets the clustered data in order to both move the robot and provide prior information to the model-fitting algorithm.
-<>.2
e
-0.
~ -0.6
~-08
-,
-,..
2
Figure 1: Plot showing data from a LIDAR scan of a wall, plotted in Cartesian space. The
robot is located at the origin, with the y axis pointing forward, x to the right, and z up. The
sparse scan pattern is visible, as well as the visibility constraint: the wall extends beyond
where the scan ends, but is invisible to the LIDAR due to the orientation of the wall
2 Mixture of Gaussians model
The range-finder returns a set of data, each of which is a position in Cartesian space Xi =
(Xi, Yi, Zi). The complete set of data D = {Xl ... XN} is modeled as being generated by a
mixture density
M
P(xn)
=L
P(xn Ii, JLi, E i , 1l'i)P( i),
i=l
where we use a Gaussian as the sub-model, with mean JLi, variance Ei and weight 1l'i' which
makes the probability of a particular data point:
M
P(xnIJL, E, 1l') =
~ (21l')3/:jE I1/2 exp ( -~(Xn i
JLi)TE;l(xn - JLi))
Robot Docking Using Mixtures of Gaussians
947
Given a set of data D, the most likely set of parameters is found using the EM algorithm.
This algorithm has a number of advantages, such as guaranteed convergence to a local
minimum, and efficient computational performance.
In 3D Cartesian space, the Gaussian sub-models form ellipsoids, where the size and orientation are determined by the covariance matrix ~~. In the general case, the EM algorithm
can be used to learn all the parameters of ~i. The sparseness of the LIDAR data makes
this parameterization inappropriate, as various odd collections of points could be clustered
together. By changing the parameterization of ~~ to better model plane-like structures, the
system can be improved. The reparameterization is most readily expressed in terms of the
eigenvalues Ai and eigenvectors ~ of the covariance matrix ~i = ~Ai ~ -I.
The covariance matrix of a normal approximation to a plane-like vertical structure will
have a large eigenvalue in the z direction, and in the x-y plane one large and one small
I = ~T = ~,
eigenvalue. Since ~i is symmetrical, the eigenvectors are orthogonal,
and ~i can be written:
v:-
o
where Oi is the angle of orientation of the ith sub-model in the x-y plane, ai scales the
cluster in the x and y directions, and bi scales in the z direction. The constant, controls
the aspect ratio of the ellipsoid in the x-y plane. I
The optimal values of these parameters (a, b) are found using EM, first calculating the
probability that data point Xn is modeled by Gaussian i, (h tn ) for every data point Xn and
every Gaussian i,
hin =
7ril~il-1/2 exp (-~(Xn - fli)T~il(Xn - fli))
--~M~--------~~~--------~--------~--
Li==1 7ril~~I - 1/2exp (-~(Xn - fldT~il(Xn - fli))'
This "responsibility" is then used as a weighting for the updates to the other parameters,
{) _ ~ t -I ( 2 Ln htn(Xnl - flil)(X n 2 - fli2) )
t 2 an
Ln htn[(Xnl - fl~I)2 - (Xn2 - fli2)2]
(r - l)((xnl - flid sin 0 + (Xn2 - fl~2) COSO)2 + (Xnl - flid 2 + (Xn2 - fli2)2
Ln hinxn
Ln h tn '
fli
Ln hin(
2, Ln hin '
b_
t -
Ln h in (Xn3 - fln3)2
Ln h tn
'
where Xnl is the first element of Xn etc. and ( corresponds to the projection of the data into
the plane ofthe cluster. It is im~ortant to update the means fli first, and use the new values
to update the other parameters. Figure 2 shows a typical model response on real LIDAR
data.
2.1
Practicalities of application, and results
Starting values for the model parameters are important, as EM is only guaranteed to find a
local optimum. The Gaussian mixture components are initialized with a large covariance,
allowing them to pick up data and move to the correct positions. We found that initializing
the means fli to random data points, rather than randomly in the input space, tended to
1By
experimentation, a value of'Y of 0.01 was found to be reasonable for this application.
2Intuition for the Oi update can be obtained by considering that (Xnl - fltl) is the x component of
the distance between Xn and /.Li, which is IXn - /.Ld cos and similarly (Xn2 - /.Li2) is IXn - /.Li Isin
so tan 2() = sin 20 = 2 sin 0 cos 0 = 2(xn1 -1'.1 )(xn 2 -1'.2) .
e,
cos 20
cos 2 0-sin 2 0
(X n 1-l'i1 )2 -(Xn2 -1'.2)2
e,
M. M. Williamson, R. Murray-Smith and V. Hansen
948
O '+ ~~
1 ;Ui?h?
"
----..-~
...
+
?
?
Figure 2: Example of clustering of the 3d data points. The left hand graph shows the view
from above (the x-y plane), and the right graph shows the view from the side (the y-z
plane), with the robot positioned at the origin. The scene shows a box at an oblique angle,
with a wall behind. The extent of the plane-like Gaussian sub-models is illustrated using
the ellipses, which are drawn at a probability of 0.5.
work better, especially given the sensor characteristics-if the LIDAR returned a range
measurement, it was likely to be part of an interesting object.
Despite the accuracy of measurement, there are still outlying data points, and it is impossible to fully segment the data into separate objects. One simple solution we found was
to define a "junk" Gaussian. This is a sub-model placed in the center of the data, with a
large covariance ~. This Gaussian then becomes responsible for the outliers in the data (i.e.
sparsely distributed data over the whole scene, none of which are associated with a specific
object), allowing the object-modeling Gaussians to work undistracted.
The use of EM with the a, b, e parameterization found and represented plane-like data clusters better than models where all the elements of the covariance matrix were free to adapt.
It also tended to converge faster, probably due to the reduced numbers of parameters in the
covariance matrix (3 as opposed to 6). Although the algorithm is constrained to find planes,
the parameterization was flexible enough to model other objects such as thin vertical lines
(say from a table leg). The only problem with the algorithm was that it occasionally found
poor local minimum solutions, such as illustrated in Figure 3. This is a common problem
with least squares based clustering methods (Duda and Hart, 1973) .
O.
o.
O.
OB
07
07
06
06
os
os
04
04
03
??
02
?
-,
01
0
-o.s
os
..
03
0.2
01
I ..%.6
-04
-02
02
04
06
08
Figure 3: Two examples of 'undesirable' local minimum solutions found by EM. Both
graphs show the top view of a scene of a box in front of a wall. The algorithm has incorrectly clustered the box with the left hand side of the wall.
949
Robot Docking Using Mixtures ofGaussians
3 Incorporating prior information
As well as reformulating the Gaussian models to suit our application, we also incorporated
prior knowledge on the parameters of the sub-models. Sensor characteristics are often
well-defined, and it makes sense to use these as early as possible in perception, rather than
dealing with their side-effects at higher levels of reasoning. Here, e.g., the visibility constraint, by which only planes which are almost perpendicular to the lidar rays are visible,
could be included by writing P(x n ) = I:~~l P(xnli, f3t)P(i)P(visiblelf3i), the updates
could be recalculated, and the feature immediately brought into the modeling process. In
addition, prior knowledge about the locations and sizes of objects, maybe from other sensors, can be used to influence the modeling procedure. This allows the sensor to make
better use of the sparse data.
For a model with parameters f3 and data D, Bayes rule gives:
P(f3)
P(,8ID) = P(D)
II P(xn lf3)?
Normally the logarithm of this is taken, to give the log-likelihood, which in the case of
mixtures of Gaussians is
L(DIf3) = log(p({/-li, 7ri,ai,bi ,6Q)) -log(p(D)) + LlogLp(xnli,/-li,7ri,ai,bi,Oi)
n
To include the parameter priors in the EM algorithm, distributions for the different parameters are chosen, then the log-likelihood is differentiated as usual to find the updates to the parameters (McMichael, 1995). The calculations are simplified if the
priors on all the parameters are assumed to be independent, p( {/-li, 7rt , ai, bt , Od) =
It p(/-ldp( 7ri)P( ai)p(bdp( Od?
The exact form of the prior distributions varies for different parameters, both to capture different behavior and for ease of implementation. For the element means (/-li),
a flat distribution over the data is used, specifying that the means should be among
the data points. For the element weights, a multinomial Dirichlet prior can be used,
p(7ri la) = n::~1J n~l 7rf. When the hyperparameter a > 0, the algorithm favours
weights around 1/NI, and when -1 < a < 0, weights close to 0 or 1. 3 The expected
value of ai (written as ai) can be encoded using a truncated inverse exponential prior
(McMichael, 1995), setting p(ailai) = Kexp(-at/(2ai)), where K is a normalizing
factor. 4 The prior for bi has the same form. Priors for Ot were not used, but could be useful
to capture the visibility constraint. Given these distributions, the updates to the parameters
become
I:n h in + a
/-li
I: nI:j h jn + a
o'i
I:n h in(/, + a;
2 I: nh in
bt = I:n h in (X n3 -
/-ln3)
I:n h in
2
-
+ bt .
The update for /-li is the same as before, the prior having no effect. The update for at and
bt forces them to be near ai and bi , and the update for 7ri is affected by the hyperparameter
a.
The priors on ai and bi had noticeable effects on the models obtained. Figure 4 shows the
results from two fits, starting from identical initial conditions. By adjusting the size of the
prior, the algorithm can be guided into finding different sized clusters. Large values of the
prior are shown here to demonstrate its effect.
3In this paper we make little use of the Q priors, but introducing separate Q;'S for each object
could be a useful next step for scenes with varying object sizes.
4To deal with the case when a, = 0, the prior is truncated, setting p(a;!a,) = 0 when a, < Perit .
950
..
M M Williamson. R. Murray-Smith and V. Hansen
~
.'
.
"
.....
\....t
,
~
1,
4'
~.
.,
.
.~
?
f
f.1J
~
;
.
~
~,
\....t
-
~
)
~
6JiiZC3!'
.
..
~.'
.,
.,
~
?
~
~'.
.
~
f
,
f:>
~
;
.
...
.
@
,.
.. ) ., ..
'.'
Figure 4: Example of the action of the priors on ai and bi . The photograph shows a
visual image of the scene: a box in front of a wall, and the priors were chosen to prefer a
distribution matching the wall. The two left hand graphs show the top and side view of the
scene clustered without priors, while the two right hand graphs use priors on ai and bi . The
priors give a preference for large values of ai and bi , so biasing the optimization to find a
mixture component matching the whole wall as opposed to just the top of it.
4 Classification and diagnosis
FEATURES
SENSOR
MODEL FITIING
DATA
EM ALGORITHM
PRIOR
HIGHER LEVEL
MOVE COMMAND
PROCESSING
FOR ROBOT
INFORMATION
Figure 5: Schematic of system
This section describes how higher-level processing can be used to not only interpret the
clusters fitted by the EM algorithm, but also affect the model-fitting using prior information.
The processes of model-fitting and analysis are thus coupled, and not sequential.
The results of the model fitting are primarily processed to steer the robot. Once the cluster
has been recognized as a boxlwaIVetc., the location and orientation are used to calculate
a move command. To perform the object-recognition, we used a simple classifier on a
feature vector extracted from the clustered data. The labels used were specific to docking,
and commonly clustered objects - boxes, walls, thin vertical lines. but also included labels
for clustering errors (like those shown in Figure 3). The features used were the values of the
parameters ai, bi , giving the size of the clusters, but also measures of the visibility of the
clusters, and the skewness of the within-cluster data. The classification used simple models
of the probability distributions of the features fi' given the objects OJ (i.e. P(hIOj)),
using a set of training data. In addition to moving the robot, the classifier can modify the
behavior of the model fitting algorithm. If a poor clustering solution is found, EM can be
re-run with slightly different initial conditions. If the probable locations or sizes of objects
are known from previous scans, or indeed from other sensors, then these can constrain the
clustering through priors, or provide initial means.
Robot Docking Using Mixtures ofGaussians
951
5 Summary
This paper shows that the Mixture of Gaussians architecture combined with EM optimization and the use of parameter priors can be used to segment and analyze real data from the
3D range-finder of a mobile robot. The approach was successfully used to guide a mobile
robot towards a docking object, using only its range-finder for perception.
For the learning community this provides more than an example of the application of a
probabilistic model to a real task. We have shown how the usual Mixture of Gaussians
model can be parameterized to include expectations about the environment in a way which
can be readily extended. We have included prior knowledge at three different levels: 1.
The use of problem-specific parameterization of the covariance matrix to find expected
patterns (e.g. planes at particular angles). 2. The use of problem-specific parameter priors
to automatically rule-out unlikely objects at the lowest level of perception. 3. The results of
the clustering process were post-processed by higher-level classification algorithms which
interpreted the parameters of the mixture components, diagnosed typical misclassification,
provided new priors for future perception, and gave the robot control system new targets.
It is expected that the basic approach can be fruitfully applied to other sensors, to problems which track dynamically changing scenes, or to problems which require relationships
between objects in the scene to be accounted for and interpreted. A problem common
to all modeling approaches is that it is not trivial to determine the number and types of
clusters needed to represent a given scene. Recent work with Markov-Chain Monte-Carlo
approaches has been successfully applied to mixtures of Gaussians (Richardson and Green,
1997), allowing a Bayesian solution to this problem, which could provide control systems
with even richer probabilistic information (a series of models conditioned on number of
clusters).
Acknowledgements
All authors were employed by Daimler-Benz AG during stages of the work. R. MurraySmith gratefully acknowledges the support of Marie Curie TMR grant FMBICT96 I 369.
References
Arman, F. and Aggarwal, J. K. (1993). Model-based object recognition in dense-range
images-a review. ACM Computing Surveys, 25 (1), 5-43.
Dempster, A. P., Laird, N. M., and Rubin, D. B. (1977). Maximum likelihood from incomplete data via the EM algorithm. J. Royal Statistical Society Series B, 39, 1-38.
Duda, R. O. and Hart, P. E. (1973). Pattern Classification and Scene Analysis. New York,
Wiley.
McMichael, D. W. (1995). Bayesian growing and pruning strategies for MAP-optimal
estimation of gaussian mixture models. In 4th lEE International Con! on Artificial
Neural Networks, pp. 364-368.
PIAP (1995) . PIAP impact report on TRC lidar performance. Technical Report 1, Industrial Research Institute for Automation and Measure ments, 02-486 Warszawa, AI.
Jerozolimskie 202, Poland.
Richardson, S. and Green, P. J. (1997). On Bayesian anaysis of mixtures with an unknown
number of components. Journal of the Royal Statistical Society B, 50 (4), 700-792.
Titterington, D., Smith, A., and Makov, U. (1985). Statistical Analysis of Finite Mixture
Distributions. Chichester, John Wiley & Sons.
| 1538 |@word duda:2 pulse:1 covariance:8 pick:1 ld:1 initial:3 series:2 com:1 od:2 trc:1 written:2 readily:2 john:1 visible:2 visibility:5 plot:1 update:10 parameterization:6 plane:14 ith:1 smith:5 oblique:1 provides:2 location:3 preference:1 mathematical:1 shiny:1 become:1 fitting:5 ray:1 indeed:1 expected:4 behavior:2 growing:1 inspired:1 detects:1 automatically:1 little:1 equipped:1 inappropriate:1 considering:1 becomes:1 provided:2 lowest:1 cm:1 interpreted:2 skewness:1 titterington:2 finding:2 ag:1 every:2 classifier:3 control:3 normally:1 grant:1 before:1 local:5 modify:1 despite:1 accumulates:1 id:1 dynamically:1 specifying:1 co:4 ease:2 range:12 perpendicular:2 bi:10 responsible:1 procedure:1 reject:1 projection:1 matching:2 undesirable:1 close:1 impossible:1 writing:1 influence:1 map:1 center:1 attention:1 starting:2 survey:1 immediately:1 rule:2 reparameterization:1 jli:4 target:1 tan:1 exact:1 us:1 origin:2 element:4 recognition:3 located:1 coso:1 sparsely:1 module:1 solved:1 initializing:1 capture:2 calculate:1 intuition:1 roderick:1 dempster:2 environment:2 ui:1 segment:2 various:1 represented:1 laser:2 describe:1 monte:1 detected:1 artificial:1 encoded:1 widely:1 richer:1 denser:1 say:1 richardson:2 laird:1 advantage:2 eigenvalue:3 relevant:1 combining:1 convergence:1 cluster:13 optimum:1 object:24 odd:1 noticeable:1 expectationmaximization:1 direction:4 guided:1 correct:1 bin:1 require:1 clustered:6 wall:12 probable:1 im:1 around:1 normal:1 exp:3 recalculated:1 matthew:1 pointing:1 early:1 estimation:1 label:2 hansen:5 successfully:2 mit:1 brought:1 sensor:11 gaussian:12 rna:1 modified:1 rather:2 volker:1 mobile:4 varying:1 command:2 modelling:1 likelihood:3 industrial:1 summarizing:1 sense:2 dependent:1 bt:4 unlikely:1 spurious:1 i1:2 germany:1 ldp:1 among:1 flexible:2 orientation:4 classification:4 constrained:1 once:1 f3:2 having:1 identical:1 thin:2 future:1 report:2 primarily:1 randomly:1 hamper:1 suit:1 lf3:1 chichester:1 mixture:19 light:1 behind:1 chain:1 accurate:1 ln3:1 orthogonal:1 incomplete:1 logarithm:1 initialized:1 re:2 plotted:2 fitted:1 soft:1 modeling:4 steer:1 maximization:1 introducing:1 f3t:1 fruitfully:1 front:2 reported:1 varies:1 combined:2 density:1 international:1 probabilistic:4 off:1 lee:1 together:1 tmr:1 reflect:1 opposed:2 leading:1 return:1 li:9 makov:1 automation:1 view:4 lab:1 responsibility:1 analyze:1 bayes:1 curie:1 square:1 ni:2 oi:3 il:3 variance:1 characteristic:4 accuracy:1 identify:1 ofthe:1 bayesian:3 none:1 carlo:1 tended:2 pp:1 naturally:1 associated:1 xn1:1 junk:1 con:1 adjusting:1 knowledge:5 segmentation:1 positioned:1 higher:7 reflected:1 response:1 improved:1 formulation:3 box:6 diagnosed:1 generality:1 just:1 stage:2 flight:1 hand:4 ei:1 o:3 usa:1 effect:4 ril:2 reformulating:1 iteratively:1 illustrated:2 deal:1 sin:4 during:2 complete:1 demonstrate:1 invisible:2 tn:3 hin:3 reasoning:1 image:4 fi:1 common:2 specialized:1 multinomial:1 flid:2 nh:1 interpret:2 measurement:2 cambridge:1 ai:21 ofgaussians:2 similarly:1 gratefully:1 had:1 moving:1 robot:18 surface:1 etc:2 recent:1 occasionally:1 yi:1 minimum:3 employed:1 recognized:1 converge:2 determine:1 ii:2 xn2:5 multiple:1 aggarwal:2 technical:2 faster:1 adapt:1 calculation:1 hart:2 post:1 finder:5 ellipsis:1 calculates:1 schematic:1 impact:1 basic:2 hioj:1 expectation:2 mog:6 represent:1 beam:1 addition:2 htn:2 ot:1 probably:1 near:1 ideal:1 enough:1 affect:1 fit:4 zi:1 misclassifications:1 architecture:1 gave:1 interprets:2 rod:1 favour:1 motivated:1 returned:2 york:1 action:1 useful:4 eigenvectors:2 maybe:1 dark:1 processed:2 daimler:1 reduced:1 track:1 li2:1 diagnosis:1 hyperparameter:2 affected:1 drawn:1 changing:2 marie:1 isin:1 graph:5 run:1 angle:3 inverse:1 uncertainty:1 powerful:1 parameterized:1 extends:1 almost:1 reasonable:1 ob:1 prefer:1 fl:2 guaranteed:3 kexp:1 constraint:3 constrain:2 scene:13 ri:5 flat:1 n3:1 aspect:1 poor:4 describes:1 slightly:1 em:15 son:1 making:1 modification:1 leg:1 outlier:1 taken:1 ln:8 benz:1 needed:1 end:1 b_:1 gaussians:14 experimentation:1 murraysmith:1 differentiated:1 jn:1 top:3 clustering:7 include:2 dirichlet:1 calculating:1 giving:1 reflectance:1 practicality:1 murray:4 especially:3 society:2 move:4 strategy:1 rt:1 usual:2 distance:1 separate:2 berlin:1 collected:1 extent:1 trivial:1 denmark:1 modeled:2 relationship:1 ellipsoid:2 providing:1 ratio:1 implementation:2 unknown:1 perform:1 allowing:3 vertical:4 markov:1 finite:1 incorrectly:1 reparameterized:1 truncated:2 extended:1 incorporated:2 community:1 beyond:1 proceeds:1 perception:6 pattern:4 biasing:1 recast:1 rf:1 oj:1 green:2 royal:2 misclassification:1 force:2 ixn:2 dtu:1 axis:1 acknowledges:1 coupled:1 docking:8 prior:35 review:2 acknowledgement:1 poland:1 fully:1 interesting:2 allocation:1 xnl:6 rni:1 rubin:1 bdp:1 summary:1 accounted:1 placed:1 free:1 guide:2 side:4 institute:1 sparse:6 distributed:1 calculated:1 xn:15 author:2 forward:1 collection:1 commonly:1 simplified:1 outlying:1 pruning:1 dealing:2 imm:1 symmetrical:1 assumed:1 xi:2 table:1 nature:1 learn:1 williamson:4 fli:6 dense:2 whole:2 je:1 wiley:2 sub:7 position:2 exponential:1 xl:1 weighting:1 specific:4 showing:1 dk:1 alt:1 ments:1 concern:1 evidence:1 incorporating:1 normalizing:1 sequential:1 te:1 conditioned:1 cartesian:4 sparseness:2 photograph:1 likely:2 visual:1 expressed:1 applies:1 corresponds:1 extracted:1 ma:1 acm:1 sized:1 towards:1 lidar:9 included:5 determined:1 typical:2 la:1 support:1 scan:7 incorporate:1 |
589 | 1,539 | Fisher Scoring and a Mixture of Modes
Approach for Approximate Inference and
Learning in Nonlinear State Space Models
Thomas Briegel and Volker Tresp
Siemens AG, Corporate Technology
Dept. Information and Communications
Otto-Hahn-Ring 6,81730 Munich, Germany
{Thomas.Briegel, Volker.Tresp} @mchp.siemens.de
Abstract
We present Monte-Carlo generalized EM equations for learning in nonlinear state space models. The difficulties lie in the Monte-Carlo E-step
which consists of sampling from the posterior distribution of the hidden
variables given the observations. The new idea presented in this paper is
to generate samples from a Gaussian approximation to the true posterior
from which it is easy to obtain independent samples. The parameters of
the Gaussian approximation are either derived from the extended Kalman
filter or the Fisher scoring algorithm. In case the posterior density is multimodal we propose to approximate the posterior by a sum of Gaussians
(mixture of modes approach). We show that sampling from the approximate posterior densities obtained by the above algorithms leads to better
models than using point estimates for the hidden states. In our experiment, the Fisher scoring algorithm obtained a better approximation of
the posterior mode than the EKF. For a multimodal distribution, the mixture of modes approach gave superior results.
1
INTRODUCTION
Nonlinear state space models (NSSM) are a general framework for representing nonlinear
time series. In particular, any NARMAX model (nonlinear auto-regressive moving average
model with external inputs) can be translated into an equivalent NSSM. Mathematically, a
NSSM is described by the system equation
(1)
where Xt denotes a hidden state variable, (t denotes zero-mean uncorrelated Gaussian noise
with covariance Qt and Ut is an exogenous (deterministic) input vector. The time-series
measurements Yt are related to the unobserved hidden states Xt through the observation
equation
(2)
where Vt is uncorrelated Gaussian noise with covariance lit. In the following we assume
that the nonlinear mappings fw{.) and gv{.) are neural networks with weight vectors w
and v, respectively. The initial state Xo is assumed to be Gaussian distributed with mean
ao and covariance Qo. All variables are in general multidimensional. The two challenges
404
T Briegel and V. Tresp
in NSSMs are the interrelated tasks of inference and learning. In inference we try to estimate the states of unknown variables Xs given some measurements Yb ... , Yt (typically
the states of past (s < t), present (s = t) or future (s > t) values of Xt) and in learning we
want to adapt some unknown parameters in the model (i.e. neural network weight vectors
wand v) given a set of measurements. 1 In the special case of linear state space models
with Gaussian noise, efficient algorithms for inference and maximum likelihood learning
exist. The latter can be implemented using EM update equations in which the E-step is
implemented using forward-backward Kalman filtering (Shumway & Stoffer, 1982). If
the system is nonlinear, however, the problem of inference and learning leads to complex
integrals which are usually considered intractable (Anderson & Moore, 1979). A useful
approximation is presented in section 3 where we show how the learning equations for
NSSMs can be implemented using two steps which are repeated until convergence. First in
the (Monte-Carlo) E-step, random samples are generated from the unknown variables (e.g.
the hidden variables Xt) given the measurements. In the second step (a generalized M-step)
those samples are treated as real data and are used to adapt Iw (.) and gv (.) using some
version of the backpropagation algorithm. The problem lies in the first step, since it is difficult to generate independent samples from a general multidimensional distribution. Since
it is difficult to generate samples from the proper distribution the next best thing might be
to generate samples using an approximation to the proper distribution which is the idea
pursued in this paper. The first thing which might come to mind is to approximate the
posterior distribution of the hidden variables by a multidimensional Gaussian distribution
since generating samples from such a distribution is simple. In the first approach we use the
extended Kalman filter and smoother to obtain mode and covariance ofthis Gaussian. 2 Alternatively, we estimate the mode and the covariance of the posterior distribution using an
efficient implementation of Fisher scoring derived by Fahrmeir and Kaufmann (1991) and
use those as parameters of the Gaussian. In some cases the approximation of the posterior
mode by a single Gaussian might be considered too crude. Therefore, as a third solution,
we approximate the posterior distribution by a sum of Gaussians (mixture of modes approach). Modes and covariances of those Gaussians are obtained using the Fisher scoring
algorithm. The weights of the Gaussians are derived from the likelihood of the observed
data given the individual Gaussian. In the following section we derive the gradient of the
log-likelihood with respect to the weights in I w (.) and gv (.). In section 3, we show that
the network weights can be updated using a Monte-Carlo E-step and a generalized M-step.
Furthermore, we derive the different Gaussian approximations to the posterior distribution
and introduce the mixture of modes approach. In section 4 we validate our algorithms using
a standard nonlinear stochastic time-series model. In section 5 we present conclusions.
2
THE GRADIENTS FOR NONLINEAR STATE SPACE MODELS
Given our assumptions we can write the joint probability of the complete data for t
1, ... , T as 3
p(Xr, Yr, Ur) = p(Ur) p(xo)
r
r
t=1
t=l
II p(Xt IXt-l. ut} II p(Yt IXt, ut}
(3)
1 In this paper we focus on the case s :::; t (smoothing and offline learning, respectively).
2Independently from our work, a single Gaussian approximation to the E-step using the EKFS
has been proposed by Ghahramani & Roweis (1998) for the special case of a RBF network. They
show that one obtains a closed form M-step when just adapting the linear parameters by holding
the nonlinear parameters fixed. Although avoiding sampling, the computational load of their M-step
seems to be significant.
3In the following, each probability density is conditioned on the current model. For notational
convenience, we do not indicate this fact explicitly.
Fisher Scoring and Mixture of Modes for Inference and Learning in NSSM
405
=
where UT
{Ul,"" UT} is a set of known inputs which means that p( UT) is irrelevant
in the following. Since only YT = {Yl,"" YT} and UT are observed, the log-likelihood
of the model is
log L = log J p(XT, YTIUT)p(UT) dXT ex log J p(XT, YTIUT ) dXT
(4)
with XT = {xo, ... , XT}. By inserting the Gaussian noise assumptions we obtain the
gradients of the log-likelihood with respect to the neural network weight vectors wand v,
respectively (Tresp & Hofmann, 1995)
T
810gL
8w
ex
810gL
8v
ex
~J!8fw(Xt-l,Ut)(
)
I
L.J
8w
Xt -fw(Xt-llUd p(Xt,Xt-l YT,UT)dxt-ldxt
t=1
~J8gv(Xt'Ut)(
)
L.J
8v
Yt-gv(Xt,ut} p(Xt!YT,UT)dxt.
(5)
t=1
3
3.1
APPROXIMATIONS TO THE E-STEP
Monte-Carlo Generalized EM Learning
The integrals in the previous equations can be solved using Monte-Carlo integration which
leads to the following learning algorithm.
xr
1. Generate S samples {xo, ... , };=1 from P(XT \YT , UT ) assuming the current
model is correct (Monte-Carlo E-Step).
2. Treat those samples as real data and update w new = wold
void + 1]&I~~L with stepsize 1] and
T
aIogL
8w
aIogL
8
v
ex
S t=1 s=1
ex
=
5
~2:2:8fw(Xt-l,udl
T
+ 1] &~! Land v new
8w
(x:-fW(X:_l,ud) (6)
Xt-l=t:_ 1
5
~~~8gv (Xt,Ut)1
( _ ("S
))
SL.JL.J
8
Yt gv Xt,Ut
t=1 s=1
V
Xt=i:;
(7)
(generalized M-step). Go back to step one.
The second step is simply a stochastic gradient step. The computational difficulties lie
in the first step. Methods which produce samples from multivariate distributions such as
Gibbs sampling and other Markov chain Monte-Carlo methods have (at least) two problems. First, the sampling process has to "forget" its initial condition which means that the
first samples have to be discarded and there are no simple analytical tools available to determine how many samples must be discarded . Secondly, subsequent samples are highly
correlated which means that many samples have to be generated before a sufficient amount
of independent samples is available. Since it is so difficult to sample from the correct
posterior distribution p(XT !YT, UT) the idea in this paper is to generate samples from an
approximate distribution from which it is easy to draw samples. In the next sections we
present approximations using a multivariate Gaussian and a mixture of Gaussians.
3.2
Approximate Mode Estimation Using the Extended Kalman Filter
Whereas the Kalman filter is an optimal state estimator for linear state space models the
extended Kalman filter is a suboptimal state estimator for NSSMs based on locallinearizations of the nonlinearities. 4 The extended Kalman filter and smoother (EKFS) algorithm is
4 Note that we do not include the parameters in the NSSM as additional states to be estimated as
done by other authors, e.g. Puskorius & Feldkamp (1994).
T. Briegel and V. Tresp
406
a forward-backward algorithm and can be derived as an approximation to posterior mode
estimation for Gaussian error sequences (Sage & Melsa, 1971). Its application to our framework amounts to approximating x~ode ~ x~KFS where x~KFS is the smoothed estimate
of Xt obtained from forward-backward extended Kalman filtering over the set of measurements YT and x~ode is the mode of the posterior distribution p( Xt IYT , UT). We use x~KFS
as the center of the approximating Gaussian. The EKFS also provides an estimate of the
error covariance of the state vector at each time step t which can be used to form the covariance matrix of the approximating Gaussian. The EKFS equations can be found in Anderson
& Moore (1979). To generate samples we recursively apply the following algorithm. Given
xLI is a sample from the Gaussian approximation of p(xt-IIYT, UT) at time t - 1 draw
from p(XtIXt-1 = X:_I' YT, UT). The last conditional density is Gaussian
a sample
with mean and covariance calculated from the EKFS approximation and the lag-one error
covariances derived in Shumway & Stoffer (1982), respectively.
xt
3.3
Exact Mode Estimation Using the Fisher Scoring Algorithm
If the system is highly nonlinear, however, the EKFS can perform badly in finding the
posterior mode due to the fact that it uses a first order Taylor series expansion of the nonlinearities fw (.) and gv(.) (for an illustration, see Figure 1). A u!:>cful- and computationally
tractable - alternative to the EKFS is to compute the "exact" posterior mode by maximizing
logp(XT IYrr, UT) with respect to XT. A suitable way to determine a stationary point of
the log posterior, or equivalently, of P(XT, YTIUT) (derived from (3) by dropping P(UT))
FS old
. to app Iy rlS
D' h
.
W'It h the current estImate
.
XT
.
IS
er scormg.
'
we get a better estImate
X~s,new = X;S,old + 1] J for the unknown state sequence XT where J is the solution of
(8)
with the score function s(XT
S(XT)
)
= alogp(::~YTIUT) and the expected information matrix
= E[_a210~1X;{fIUT'J.5
T
T
By extending the arguments given in Fahrmeir &
Kaufmann (1991) to nonlinear state space models it turns out that solving equation (8) e.g. to compute the inverse of the expected information matrix - can be performed by
Cholesky decomposition in one forward and backward pass. 6 The forward-backward steps
can be implemented as a fast EKFS-Iike algorithm which has to be iterated to obtain the
maximum posterior estimates x~ode
x;S (see Appendix). Figure 1 shows the estimate
obtained by the Fisher scoring procedure for a bimodal posterior density. Fisher scoring
is successful in finding the "exact" mode, the EKFS algorithm is not. Samples of the
approximating Gaussian are generated in the same way as in the last section.
=
3.4
The Mixture of Modes Approach
The previous two approaches to posterior mode smoothing can be viewed as single Gaussian approximations of the mode of p(XTIYT, UT). In some cases the approximation of
the posterior density by a single Gaussian might be considered too crude, in particular if
the posterior distribution is multimodal. In this section we approximate the posterior by a
weighted sum of m Gaussians p(XT IYT, UT) ~ :E~I okp(XT Ik) where p(XT Ik) is the
k-th Gaussian. If the individual Gaussians model the different modes we are able to model
multimodal posterior distributions accurately. The approximations of the individual modes
are local maxima of the Fisher scoring algorithm which are f~)Und by starting the algorithm
using different initial conditions. Given the different Gaussians, the optimal weighting facp(YTlk)p(k)jp(YT) where p(YTlk)
jp(YTIXT)P(XTlk)dXT is the
tors are ok
=
=
SNote that the difference between the Fisher scoring and the Gauss-Newton update is that in the
fonner we take the expectation of the information matrix.
6The expected information matrix is a positive definite blOCk-tridiagonal matrix.
Fisher Scoring and Mixture of Modes for Inference and Learning in NSSM
407
likelihood of the data given mode k. If we approximate that integral by inserting the Fisher
scoring solutions x;S,k for each time step t and linearize the nonlinearity gv (.) about the
Fisher scoring solutions, we obtain a closed form solution for computing the ok (see Appendix). The resulting estimator is a weighted sum of the m single Fisher scoring estimates
x~M
L::=l ok x;s,k. The mixture of modes algorithm can be found in the Appendix.
For the learning task samples of the mixture of Gaussians are based on samples of each of
the m single Gaussians which are obtained the same way as in subsection 3.2.
=
4
EXPERIMENTAL RESULTS
In the first experiment we want to test how well the different approaches can approximate
the posterior distribution of a nonlinear time series (inference). As a time-series model we
chose
1 2
= 0.5 Xt-l + 25 1 +Xt-l
2
+ 8eas ( 1.2(t -I},) g(xt} = 20Xt,
(9)
xt_ 1
the covariances Qt = 10, lit = 1 and initial conditions ao = 0 and Qo = 5 which is
f(Xt-l, Ut}
considered a hard inference problem (Kitagawa, 1987). At each time step we calculate the
expected value of the hidden variables Xt, t = 1, ... , 400 based on a set of measurements
Y400 = {Yl, ... , Y400} (which is the optimal estimator in the mean squared sense) and based
on the different approximations presented in the last section. Note that for the single mode
approximation, x~ode is the best estimate of Xt based on the approximating Gaussian. For
the mixture of modes approach, the best estimate is L:~l ok x;S,k where x;S,k is the mode
of the k-th Gaussian in the dimension of Xt. Figure 2 (left) shows the mean squared error
(MSE) of the smoothed estimates using the different approaches. The Fisher scoring (FS)
algorithm is significantly better than the EKFS approach. In this experiment, the mixture of
modes (MM) approach is significantly better than both the EKFS and Fisher scoring. The
reason is that the posterior probability is multimodal as shown in Figure 1.
In the second experiment we used the same time-series model and trained a neural network to approximate fw (.) where all covariances were assumed to be fixed and known.
For adaptation we used the learning rules of section 3 using the various approximations
to the posterior distribution of XT . Figure 2 (right) shows the results. The experiments
show that truly sampling from the approximating Gaussians gives significantly better results than using the expected value as a point estimate. Furthermore, using the mixture
of modes approach in conjunction with sampling gave significantly better results than the
approximations using a single Gaussian . When used for inference, the network trained using the mixture of modes approach was not significantly worse than the true model (5%
significance level, based on 20 experiments).
5
CONCLUSIONS
In our paper we presented novel approaches for inference and learning in NSSMs. The
application of Fisher scoring and the mixture of modes approach to nonlinear models as
presented in our paper is new. Also the idea of sampling from an approximation to the
posterior distribution of the hidden variables is presented here for the first time. Our results
indicate that the Fisher scoring algorithm gives better estimates of the expected value of
the hidden variable than the EKFS based approximations. Note that the Fisher scoring algorithm is more complex in requiring typically 5 forward-backward passes instead of only
one forward-backward pass for the EKFS approach. Our experiments also showed that if
the posterior distribution is multi modal, the mixture of modes approach gives significantly
better estimates if compared to the approaches based on a single Gaussian approximation.
Our learning experiments show that it is important to sample from the approximate distributions and that it is not sufficient to simply substitute point estimates. Based on the
T. Briegel and V. Tresp
408
0.4,---- - - - - - - - - - - - - - - ---,
0 . 2 , - - -- -- - - - - - - - - - - - - - - - .
0 . 18
0.16
,
, ,
,
0.14
0 . 35
..
0.3
1t 0.25
.8
0.2
1.0.15
0.05
,
,,- ....
,
o - 'o---~~~-=-=--~o--~--~~-~
t =-= 2 9 5
Figure 1: Approximations to the posterior distribution p( x t iY400, U400) for t = 294 and t =
295. The continuous line shows the posterior distribution based on Gibbs sampling using
1000 samples and can be considered a close approximation to the true posterior. The EKFS
approximation (dotted) does not converge to a mode. The Fisher scoring solution (dashdotted) finds the largest mode. The mixture of modes approach with 50 modes (dashed)
correctly finds the two modes.
sampling approach it is also possible to estimate hyperparameters (e.g. the covariance matrices) which was not done in this paper. The approaches can also be extended towards
online learning and estimation in various ways (e.g. missing data problems).
Appendix: Mixture of Modes Algorithm
The mixture of modes estimate x~M is derived as a weighted sum of k = 1, . .. ,m individual Fisher
scoring (mode) estimates x;S ,k. For m = 1 we obtain the Fisher scoring algorithm of subsection 3.3.
First, one performs the set of forward recursions (t = 1, ... , T) for each single mode estimator k.
",k
""'tit-I
=
Btk
p(~FS , k)"'k
tXt_I
",k
""'t-llt-I
""'t-Ilt-I
pT( , FS ,k)+Q
tXt_I
t
pT('FS ,k)(",k
)-1
tXt_I ""'tit-I
(10)
(11 )
(12)
,FS ,k)
St ( x t
k
'Yt
with the initialization :E~lo = Qo, 'Yo
ing recursions (t = T , .. . , 1)
k T
+ Bt
k
(13)
'Yt-l
= So (X~S , k). Then, one performs the set of backward smooth:E kt - l l t - l
(Dkt-I )-1
-
Bk:Ek
Bkt T
t
tit-I
:E~_1
(D~_d-l
0:_1
( Dtk 1)-1 'Yt-l
k
=
+ B;:E~ B; T
+ Bkok
t t
(14)
(15)
(16)
= 8fw(Xt_l
' U') I
G ( Z ) -- &YdXt,Utll
( ) - &logp(Xr,YrIUT) I
a d
&Xt_l
Xt_l=Z'
&Xt
Xt=Z, St Z &Xt
Xt=Z n
initialization o} = :E}'Y}. The k individual mode estimates x;S ,k are obtained by iterative application of the update rule X~S , k := '7 Ok + X~S , k with stepsize '7 where X~S , k = {X~S , k, .. . ,X~S,k}
with Pt(z)
t
and Ok = {o~ , ... , o} }. After convergence we obtain the mixture of modes estimate as the weighted
. h'ttng coe ffi Clents
.
k := 0'0k were
h
k(
T -,
1 "' , 0)
sum X,MM
= ",m
6k=1 0' k~FS
x t ' k WIt. h welg
0'
O't t =
t
are computed recursively starting with a uniform prior O'} =
(N(xlp,:E) stands for a Gaussian
with center p and covariance :E evaluated at x):
.k
k
O't
=
O'~+IN(Ytlgv(xfs,k, ur),
nn
(17)
(18)
Fisher Scoring and Mixture of Modes for Inference and Learning in NSSM
0.8
e
0.7
8
o.s
7
f:
';"4
~3
2
409
~
0.5
~O.4
~
0.3
0.2
0 . "1
0
0
Figure 2: Left (inference): The heights of the bars indicate the mean squared error between
the true Xt (which we know since we simulated the system) and the estimates using the
various approximations. The error bars show the standard deviation derived from 20 repetitions of the experiment. Based on the paired t-test, Fisher scoring is significantly better
than the EKFS and all mixture of modes approaches are significantly better than both EKFS
and Fisher scoring based on a 1% rejection region. The mixture of modes approximation
with 50 modes (MM 50) is significantly better than the approximation using 20 modes. The
improvement of the approximation using 20 modes (MM 20) is not significantly better than
the approximation with 10 (MM 10) modes using a 5% rejection region.
Right (learning): The heights of the bars indicate the mean squared error between the true
fw (.) (which is known) and the approximations using a multi-layer perceptron with 3 hidden units and T = 200. Shown are results using the EKFS approximation, (left) the Fisher
scoring approximation (center) and the mixture of modes approximation (right). There are
two bars for each experiment: The left bars show results where the expected value of x t
calculated using the approximating Gaussians are used as (single) samples for the generalized M-step - in other words - we use a point estimate for Xt. Using the point estimates, the
results of all three approximations are not significantly different based on a 5% significance
level. The right bars shows the result where S = 50 samples are generated for approximating the gradient using the Gaussian approximations. The results using sampling are all
significantly better than the results using point estimates (l % significance level). The sampling approach using the mixture of modes approximation is significantly better than the
other two sampling-based approaches (l % significance level). If compared to the inference
results of the experiments shown on the left, we achieved a mean squared error of 6.02 for
the mixture of modes approach with 10 modes which is not significantly worse than the
results the with the true model of 5.87 (5% significance level).
References
Anderson, B. and Moore, J. (1979) Optimal Filtering, Prentice-Hall, New Jersey.
Fahnneir, L. and Kaufmann, H. (1991) On Kalman Filtering. Posterior Mode Estimation and Fisher
Scoring in Dynamic Exponential Family Regression, Metrika, 38, pp. 37-60.
Ghahramani, Z. and Roweis, S. (1999) Learning Nonlinear Stochastic Dynamics using the Generalized EM ALgorithm, Advances in Neural Infonnation Processing Systems 11, eps. M. Keams, S.
Solla, D. Cohn, MIT Press, Cambridge, MA.
Kitagawa, G. (1987) Non-Gaussian State Space Modeling of Nonstationary Time Series (with Comments), JASA 82, pp. 1032-1063.
Puskorius, G. and Feldkamp, L. (1994) NeurocontroL of Nonlinear Dynamical Systems with KaLman
Filter Trained Recurrent Networks, IEEE Transactions on Neural Networks, 5:2, pp. 279-297.
Sage, A. and Melsa, J. (1971) Estimation Theory with Applications to Communications and Control,
McGraw-Hill, New York.
Shumway, R. and Stoffer, D. (1982) Time Series Smoothing and Forecasting Using the EM Algorithm, Technical Report No. 27, Division of Statistics, UC Davis.
Tresp, V. and Hofmann, R. (1995) Missing and Noisy Data in NonLinear Time-Series Prediction,
Neural Networks for Signal Processing 5, IEEE Sig. Proc. Soc., pp. 1-10.
| 1539 |@word dtk:1 version:1 seems:1 covariance:14 decomposition:1 recursively:2 initial:4 series:10 score:1 past:1 current:3 must:1 subsequent:1 hofmann:2 gv:8 update:4 stationary:1 pursued:1 yr:1 metrika:1 regressive:1 provides:1 height:2 welg:1 ik:2 consists:1 introduce:1 xt_l:3 expected:7 multi:2 feldkamp:2 ag:1 unobserved:1 finding:2 multidimensional:3 control:1 unit:1 before:1 positive:1 local:1 treat:1 might:4 chose:1 initialization:2 block:1 definite:1 backpropagation:1 xr:3 procedure:1 adapting:1 significantly:14 word:1 get:1 convenience:1 close:1 prentice:1 equivalent:1 deterministic:1 yt:17 center:3 maximizing:1 go:1 missing:2 starting:2 independently:1 wit:1 estimator:5 rule:2 updated:1 pt:3 exact:3 us:1 sig:1 observed:2 solved:1 calculate:1 region:2 solla:1 und:1 dynamic:2 trained:3 solving:1 xtixt:1 tit:3 division:1 translated:1 multimodal:5 joint:1 various:3 jersey:1 fast:1 dkt:1 monte:8 lag:1 otto:1 statistic:1 noisy:1 online:1 sequence:2 analytical:1 propose:1 adaptation:1 inserting:2 roweis:2 validate:1 convergence:2 extending:1 produce:1 generating:1 ring:1 derive:2 linearize:1 recurrent:1 qt:2 ex:5 soc:1 implemented:4 come:1 indicate:4 correct:2 filter:7 stochastic:3 ao:2 secondly:1 mathematically:1 kitagawa:2 mm:5 considered:5 hall:1 mapping:1 tor:1 estimation:6 proc:1 iw:1 infonnation:1 largest:1 repetition:1 tool:1 weighted:4 mit:1 gaussian:31 ekf:1 volker:2 conjunction:1 derived:8 focus:1 yo:1 notational:1 improvement:1 likelihood:6 sense:1 inference:14 nn:1 typically:2 bt:1 hidden:10 keams:1 germany:1 udl:1 smoothing:3 integration:1 special:2 uc:1 sampling:13 lit:2 rls:1 future:1 report:1 individual:5 highly:2 stoffer:3 mixture:27 truly:1 puskorius:2 chain:1 kt:1 integral:3 kfs:3 taylor:1 old:2 iiyt:1 modeling:1 logp:2 deviation:1 uniform:1 successful:1 tridiagonal:1 too:2 ixt:2 st:2 density:6 yl:2 iy:1 squared:5 worse:2 external:1 ek:1 nonlinearities:2 de:1 explicitly:1 performed:1 try:1 closed:2 exogenous:1 iike:1 kaufmann:3 xli:1 iterated:1 accurately:1 carlo:8 app:1 llt:1 pp:4 subsection:2 ut:25 ea:1 back:1 ok:6 modal:1 yb:1 done:2 wold:1 evaluated:1 anderson:3 furthermore:2 just:1 until:1 qo:3 cohn:1 nonlinear:17 mode:57 requiring:1 true:6 moore:3 davis:1 generalized:7 hill:1 complete:1 performs:2 novel:1 superior:1 jp:2 jl:1 eps:1 measurement:6 significant:1 cambridge:1 gibbs:2 nonlinearity:1 moving:1 okp:1 posterior:33 multivariate:2 showed:1 irrelevant:1 vt:1 scoring:28 additional:1 determine:2 converge:1 ud:1 dashed:1 ii:2 smoother:2 signal:1 corporate:1 ing:1 smooth:1 technical:1 adapt:2 paired:1 prediction:1 regression:1 expectation:1 bimodal:1 achieved:1 whereas:1 want:2 ode:4 void:1 pass:1 comment:1 briegel:5 thing:2 nonstationary:1 easy:2 gave:2 suboptimal:1 idea:4 ul:1 forecasting:1 f:7 york:1 useful:1 amount:2 generate:7 sl:1 exist:1 dotted:1 estimated:1 correctly:1 write:1 dropping:1 backward:8 sum:6 wand:2 inverse:1 ilt:1 family:1 draw:2 appendix:4 layer:1 alogp:1 badly:1 btk:1 argument:1 munich:1 em:5 ur:3 xo:4 computationally:1 equation:8 turn:1 ffi:1 mind:1 know:1 tractable:1 available:2 gaussians:12 apply:1 stepsize:2 alternative:1 thomas:2 substitute:1 denotes:2 include:1 newton:1 coe:1 bkt:1 ghahramani:2 hahn:1 approximating:8 xlp:1 gradient:5 simulated:1 reason:1 assuming:1 kalman:10 illustration:1 iyt:2 equivalently:1 difficult:3 holding:1 sage:2 implementation:1 proper:2 unknown:4 perform:1 observation:2 markov:1 discarded:2 extended:7 communication:2 smoothed:2 bk:1 able:1 bar:6 usually:1 dynamical:1 challenge:1 suitable:1 difficulty:2 treated:1 xtiyt:1 recursion:2 representing:1 technology:1 auto:1 tresp:7 prior:1 shumway:3 dxt:5 filtering:4 xtlk:1 jasa:1 sufficient:2 uncorrelated:2 land:1 lo:1 dashdotted:1 neurocontrol:1 gl:2 last:3 offline:1 perceptron:1 fahrmeir:2 mchp:1 distributed:1 calculated:2 dimension:1 stand:1 forward:8 author:1 transaction:1 approximate:12 obtains:1 mcgraw:1 assumed:2 alternatively:1 continuous:1 iterative:1 expansion:1 mse:1 complex:2 significance:5 noise:4 hyperparameters:1 repeated:1 xt_:1 exponential:1 lie:3 crude:2 third:1 weighting:1 load:1 xt:53 er:1 x:1 intractable:1 ofthis:1 ekfs:17 conditioned:1 rejection:2 forget:1 interrelated:1 simply:2 ma:1 conditional:1 viewed:1 rbf:1 towards:1 fisher:28 fw:9 hard:1 pas:2 gauss:1 experimental:1 siemens:2 cholesky:1 latter:1 dept:1 avoiding:1 correlated:1 |
590 | 154 | 81
WHAT SIZE NET GIVES VALID
GENERALIZATION?*
Eric B. Baum
Department of Physics
Princeton University
Princeton NJ 08540
David Haussler
Computer and Information Science
University of California
Santa Cruz, CA 95064
ABSTRACT
We address the question of when a network can be expected to
generalize from m random training examples chosen from some arbitrary probability distribution, assuming that future test examples
are drawn from the same distribution. Among our results are the
following bounds on appropriate sample vs. network size. Assume
o < ? $ 1/8. We show that if m > O( ~log~) random examples can be loaded on a feedforward network of linear threshold
functions with N nodes and W weights, so that at least a fraction
1 - t of the examples are correctly classified, then one has confidence approaching certainty that the network will correctly classify
a fraction 1 - ? of future test examples drawn from the same distribution. Conversely, for fully-connected feedforward nets with
one hidden layer, any learning algorithm using fewer than O(
random training examples will, for some distributions of examples
consistent with an appropriate weight choice, fail at least some
fixed fraction of the time to find a weight choice that will correctly
classify more than a 1 - ? fraction of the future test examples.
'!')
INTRODUCTION
In the last few years, many diverse real-world problems have been attacked by back
propagation. For example "expert systems" have been produced for mapping text
to phonemes [sr87], for determining the secondary structure of proteins [qs88], and
for playing backgammon [ts88].
In such problems, one starts with a training database, chooses (by making an educated guess) a network, and then uses back propagation to load as many of the
training examples as possible onto the network. The hope is that the network so designed will generalize to predict correctly on future examples of the same problem.
This hope is not always realized.
*
This paper will appear in the January 1989 issue of Neural Computation. For
completeness, we reprint this full version here, with the kind permission of MIT
Press. ? 1989, MIT Press
82
Baum and Haussler
We address the question of when valid generalization can be expected. Given a
training database of m examples, what size net should we attempt to load these
on? We will assume that the examples are drawn from some fixed but arbitrary
probability distribution, that the learner is given some accuracy parameter E, and
that his goal is to produce with high probability a feedforward neural network that
predicts correctly at least a fraction 1 - E of future examples drawn from the same
distribution. These reasonable assumptions are suggested by the protocol proposed
by Valiant for learning from examples [val84]. However, here we do not assume the
existence of any "target function"; indeed the underlying process generating the
examples may classify them in a stochastic manner, as in e.g. [dh73].
Our treatment of the problem of valid generalization will be quite general in that
the results we give will hold for arbitrary learning algorithms and not just for
back propagation. The results are based on the notion of capacity introduced by
Cover [cov65] and developed by Vapnik and Chervonenkis [vc7l], [vap82]. Recent
overviews of this theory are given in [dev88], [behw87b] and [poI84], from the various
perspectives of pattern recognition, Valiant's computational learning theory, and
pure probability theory, respectively. This theory generalizes the simpler counting
arguments based on cardinality and entropy used in [behw87a] and [dswshhj87], in
the latter case specifically to study the question of generalization in feedforward
nets (see [vap82] or [behw87b]).
The particular measures of capacity we use here are the maximum number of dichotomies that can be induced on m inputs, and the Vapnik-CheMlonenki. (Ve)
Dimen.ion, defined below. We give upper and lower bounds on these measures for
classes of networks obtained by varying the weights in a fixed feedforward architecture. These results show that the VC dimension is closely related to the number of
weights in the architecture, in analogy with the number of coefficients or "degrees
of freedom" in regression models. One particular result, of some interest independent of its implications for learning, is a construction of a near minimal size net
architecture capable of implementing all dichotomies on a randomly chosen set of
points on the n-hypercube with high probability.
Applying these results, we address the question of when a network can be expected
to generalize from m random training examples chosen from some arbitrary probability distribution, assuming that future test examples are drawn from the same
distribution. Assume 0 < E < 1/8. We show that ifm ~ O(~log.r:) random examples can be loaded on a feedforward network of linear threshold functions with
N nodes and W weights, so that at least a fraction 1 - j of the examples are correctly classified, then one has confidence approaching certainty that the network
will correctly classify a fraction 1 - E of future test examples drawn from the same
distribution. Conversely, for fully-connected feedforward nets with one hidden layer,
any learning algorithm using fewer than O( ~) random training examples will, for
some distributions of examples consistent with an appropriate weight choice, fail
at least some fixed fraction of the time to find a weight choice that will correctly
classify more than a 1 - E fraction of the future test examples.
What Size Net Gives Valid Generalization?
Ignoring the constant and logarithmic factors, these results suggest that the appropriate number of training examples is approximately the number of weights times
the inversel of the accuracy parameter E. Thus, for example, if we desire an accuracy level of 90%, corresponding to E
0.1, we might guess that we would need
about 10 times as many training examples as we have weights in the network. This
is in fact the rule of thumb suggested by Widrow [wid87], and appears to work fairly
well in practice. At the end of Section 3, we briefly discuss why learning algorithms
that try to minimize the number of non-zero weights in the network [rum87] [hin87]
may need fewer training examples.
=
DEFINITIONS
We use In to denote the natural logarithm and log to denote the logarithm base
2. We define an ezample as a pair (i, a), i E ~n, a E {-I, +1}. We define a
random sample as a sequence of examples drawn independently at random from
some distribution D on ~n X {-1, +1}. Let I be a function from ~n into {-1, +1}.
We define the error of I, with respect to D, as the probability a;/; I(i) for (i,a)
a random example.
Let F be a class of {-1, +l}-valued functions on ~n and let S be a set of m points
in ~n . A dichotomy of S induced by I E F is a partition of S into two disjoint
subsets S+ and S- such that I(i)
+1 for i E S+ and I(i)
-1 for i E S-.
By .6. F (S) we denote the number of distinct dichotomies of S induced by functions
I E F, and by .6.F(m) we denote the maximum of .6.F(S) over all S C ~n of
cardinality m. We say S is shattered by F if .6.F(S) = 2151 , i.e. all dichotomies of
S can be induced by functions in F. The Vapnik-CheMlonenkis (VC) dimension of
F, denoted VCdim(F), is the cardinality of the largest S C ~n that is shattered
by F, i.e. the largest m such that .6.F ( m) 2m ?
=
=
=
A feedforward net with input from ~n is a directed acyclic graph G with an ordered
sequence ofn source nodes (called inputs) and one sink (called the output). Nodes
of G that are not source nodes are called computation nodes, nodes that are neither
source nor sink nodes are called hidden nodes. With each computation node n.
there is associated a function" : ~inde't'ee(n,) ~ {-I, +1}, where indeg7'ee(n.) is
the number of incoming edges for node n,. With the net itself there is associated
a function I : ~n ~ {-I, +1} defined by composing the I,'s in the obvious way,
assuming that component i of the input i is placed at the it" input node.
A Jeedlorward architecture is a class of feedforward nets all of which share the
same underlying graph. Given a graph G we define a feedforward architecture by
a class of functions F, from ~'nde't'ee(n,)
associating to each computation node
n,
1
It should be noted that our bounds differ significantly from those given in [dev88] in
that the latter exhibit a dependence on the inverse of e2 ? This is because we derive
our results from Vapnik's theorem on the uniform relative deviation of frequencies
from their probabilities ([vap82], see Appendix A3 of [behw87b]), giving sharper
bounds as E approaches o.
83
84
Baum and Haussler
to {-I, +1}. The resulting architecture consists of all feedforward nets obtained by
choosing a particular function" from F, for each computation node ft,. We will
identify an architecture with the class offunctions computed by the individual nets
within the architecture when no confusion will arise.
CONDITIONS SUFFICIENT FOR VALID
GENERALIZATION
Theorem 1: Let F be a feedforward architecture generated by an underlying
graph G with N > 2 computation nodes and F, be the class of functions associated
with computation node ft, of G, 1 < i < N. Let d = E~l VCdim(Fl). Then
AF(m) < n~lAF,(m)::; (Nem/d)d for m > d, where e is the base of the natural
logarithm.
Proof: Assume G has n input nodes and that the computation nodes of G are
ordered so that node receives inputs only from input nodes and from computation
nodes nj, 1 < j ::; i - I . Let S be a set of m points in ~n. The dichotomy
induced on S by the function in node nl can be chosen in at most AFI (m) ways.
This choice determines the input to node nz for each of the m points in S. The
dichotomy induced on these m inputs by the function in node nz can be chosen
in at most AF:a(m) ways, etc. Any dichotomy of S induced by the whole network
can be obtained by choosing dichotomies for each of the ni's in this manner, hence
AF(m) < nf:l AF,(m).
n,
By a theorem of Sauer [sau72], whenever VCdim(F) = Ie < 00, AF(m) < (em/Ie)l
for all m > Ie (see also [behw87b]). Let ~ = VCdim(Fi), 1 < i < N. Thus
d Ef:l~. Then n~l AF,(m) < n~l(em/~)'" for m > d. Using the fact that
E~l -ailogai < logN whenever a. > 0, 1 < i < N, and E~l ai = I, and setting
ai
~/d, it is easily verified that n~l ~d. > (d/N)d. Hence n~l(em/di)d. <
(Nem/d)d.
=
=
Corollary 2: Let F be the class of all functions computed by feedforward nets
defined on a fixed underlying graph G with E edges and N > 2 computation
nodes, each of which computes a linear threshold function. Let W
E + N (the
total number of weights in the network, including one weight per edge and one
threshold per computation node). Then AF(m) < (Nem/W)W for all m > Wand
VCdim(F) < 2Wlog(eN).
=
Proof: The first inequality follows from directly from Theorem 1 using the fact that
VCdim(F) = Ie + 1 when F is the class of all linear threshold functions on ~l (see
e.g. [wd81]). For the second inequality, it is easily verified that for N > 2 and
m 2Wlog(eN), (N em/W)W < 2m. Hence this is an upper bound on VCdim(F).
=
Using VC dimension bounds given in [wd81], related corollaries can be obtained for
nets that use spherical and other types of polynomial threshold functions. These
bounds can be used in the following.
What Size Net Gives Valid Generalization?
Theorem 3 [vapS2} (see [behw87b), Theorem A3.3): Let F be a class offunctions2
on ~n, 0 < l' < 1,0 < ?,6 < 1. Let S be a random sequence of m examples drawn
independently according to the distribution D. The probability that there exists a
function in F that disagrees with at most a fraction (1 - 1')? of the examples in S
and yet has error greater than ? (w.r.t. D) is less than
From Corollary 2 and Theorem 3, we get:
Corollary 4: Given a fixed graph G with E edges and N linear threshold units
(i.e. W = E + N weights), fixed 0 < ? < 1/2, and m random training examples,
where
32W 1 32N
m>-n-,
?
?
if one can find a choice of weights so that at least a fraction 1- ?/2 of the m training
examples are correctly loaded, then one has confidence at least 1 - Se- 1?5W that
the net will correctly classify all but a fraction ? of future examples drawn from the
same distribution. For
m
64W I 64N
> --;-
n--;-,
the confidence is at least 1 - Se-em/S2.
Proof: Let l' = 1/2 and apply Theorem 3, using the bound on aF(m) given in
Corollary 2. This shows that the probability that there exists a choice of the weights
that defines a function with error greater than ? that is consistent with at least a
fraction 1 - ?/2 of the training examples is at most
When m = !ll!.ln!!K
this is S(2e 3fN'
E In!!K)W
which is less than Se- 1. 5W for N ->
e
e'
E
'
2 and ? < 1/2. When m > 84EW In 8~N, (2N em/W) W < e Em / S2 , so S(2N em/W) W
e- Em / 16 < Se-em/S2.
The constant 32 is undoubtably an overestimate. No serious attempt has been made
to minimize it. Further, we do not know if the log term is unavoidable. Nevertheless,
even without these terms, for nets with many weights this may represent a considerable number of examples. Such nets are common in cases where the complexity
of the rule being learned is not known in advance, so a large architecture is chosen
2 We assume some measurability conditions on the class F. See [poI84], [behwS7b1 for
details.
85
86
Baum and Haussler
in order to increase the chances that the rule can be represented. To counteract the
concomitant increase in the size of the training sample needed, one method that
has been explored is the use of learning algorithms that try to use as little of the
architecture as possible to load the examples, e.g. by setting as many weights to
zero as possible, and by removing as many nodes as possible (a node can be removed
if all its incoming weights are zero.) [rumS7] [hin87]. The following shows that the
VC dimension of such a "reduced" architecture is not much larger than what one
would get if one knew a priori what nodes and edges could be deleted.
Corollary 5: Let F be the class of all functions computed by linear threshold
feedforward nets defined on a fixed underlying graph G with N' > 2 computation
nodes and E' ~ N' edges, such that at most E > 2 edges have non-zero weights
and at most N ~ 2 nodes have at least one incoming edge with a non-zero weight.
Let W = E + N. Then the conclusion of Corollary 4 holds for sample size
32W
32NE'
l
m>-n--f
f
Prool sketch: We can bound dF( m) by considering the number of ways the N nodes
and E edges that remain can be chosen from among those in the initial network. A
crude upper bound is (N')N (E')E. Applying Corollary 2 to the remaining network
gives dF(m) ~ (N')N(E')E(Nem/W)w. This is at most (N E'em/W)w. The rest
of the analysis is similar to that in Corollary 4.
This iridicates that minimizing non-zero weights may be a fruitful approach. Similar
approaches in other learning contexts are discussed in [hauSS] and [litSS].
CONDITIONS NECESSARY FOR
VALID GENERALIZATION
The following general theorem gives a lower bound on the number of examples
needed for distribution-free learning, regardless of the algorithm used.
Theorem 6 [ehkvS7] (see also [behw87b]): Let F be a class of {-I, +1}-valued
functions on ~n. with VCdim(F) > 2. Let A be any learning algorithm that takes
as input a sequence of {-I, +1}-labeled examples over ~n. and produces as output
a function from ~n. into {-I, +1}. Then for any 0 < f ~ l/S, 0 < 0 ~ l~ and
m
1- fl 1 VCdim(F) -1]
3 2e
'
e n7'
v
< maz [ -
there exists (1) a function I E F and (2) a distribution D on ~n X {-I, +1} for
which Prob((E, a) : a f. I(E)) = 0, such that given a random sample of size m
chosen according to D, with probability at least 0, A produces a function with error
greater than e.
What Size Net Gives Valid Generalization?
This theorem can be used to obtain a lower bound on the number of examples
needed to train a net, assuming that the examples are drawn from the worst-case
distribution that is consistent with some function realizable on that net. We need
only obtain lower bounds on the VC dimension of the associated architecture. In
this section we will specialize by considering only fully-connected networks of linear
threshold units that have only one hidden layer. Thus each hidden node will have an
incoming edge from each input node and an outgoing edge to the output node, and
no other edges will be present. In [b88] a slicing construction is given that shows
that a one hidden layer net of threshold units with n inputs and 2j hidden units
can shatter an arbitrary set of 2jn vectors in general position in ~". A corollary of
this result is:
Theorem 7: The class of one hidden layer linear threshold nets taking input from
~" with k hidden units has VC dimension at least 2L~Jn.
Note that for large k and n, 2 L~ Jn is approximately equal to the total number W
of weights in the network.
A special case of considerable interest occurs when the domain is restricted to
the hypercube: {+1,-1}". Lemma 6 of [lit88] shows that the class of Boolean?
functions on {+1, _I}" represented by disjunctive normal form expressions with k
terms, k < 0(2,,/2/Vn) , where each term is the conjunction of n/2 literals, has
VC dimension at least kn/4. Since these functions can be represented on a linear
threshold net with one hidden layer of k units, this provides a lower bound on the
VC dimension of this architecture. We also can use the slicing construction of [b88]
to give a lower bound approaching kn/2. The actual result is somewhat stronger in
that it shows that for large n a randomly chosen set of approximately kn/2 vectors
is shattered with high probability.
Theorem 8: With probability approaching 1 exponentially in n, a set S of m < 2,,/3
vectors chosen randomly and uniformly from {+1, _I}" can be shattered by the
one hidden layer architecture with 2rm/l(n(1 - 1~0,,))J1linear threshold units in
its hidden layer.
Prool,ketch: With probability approaching 1 exponentially in n no pair of vectors
in S are negations of each other. Assume n > eto. Let l' = In(l- I~O,,)J. Divide
S at random into m/1'1 disjoint subsets S1I ... , Srm/t'l each containing no more
than l' vectors. We will describe a set T of ?1 vectors as Iliceable if the vectors
in T are linearly independent and the subspace they span over the reals does not
contain any ?l vector other than the vectors in T and their negations. In [od188]
it is shown, for large n, that any random set of l' vectors has probability P
4(;)(~)" +0(( 110)") of not being sliceable. Thus the probability that some S. is not
sliceable is 0(mn2(~)"), which is exponentially small for m < 2,,/3. Hence with
probability approaching 1 exponentially in n, each S, is sliceable, 1 ~ i $ m/ 1'1.
r
=
Consider any Boolean function
I
on S and let
S:
= {i
E
S, : f(i)
r
= +1},
87
88
Baum and Haussler
r
1 < i < m/7'1. If Si is sliceable and no pair of vectors in S are negations of each
other then we may pass a plane through the points in
that doesn't contain any
other points in S. Shifting this plane parallel to itself slightly we can construct two
half spaces whose intersection forms a slice of~" containing
and no other points
in S. Using threshold units at the hidden layer recognizing these two half spaces,
with weights to the output unit +1 and -1 appropriately, the output unit receives
input +2 for any point in the slice and 0 for any point not in the slice. Doing this
for each
and thresholding at 1 implements the function f.
st
st
S:
We can now apply Theorem 6 to show that any neural net learning algorithm using
too few examples will be fooled by some reasonable distributions.
Corollary 9: For any learning algorithm training Ii net with k linear threshold
functions in its hidden layer, and 0 < l ~ 1/8, if the algorithm uses (a) fewer
than 2l lc/;'f,,-1 examples to learn a function from ~" to {-I, +1}, or (b) fewer
than l"lll/2J(mQ,:I:(1/!~~-10/(ln n?)J-1 examples to learn a function from {-I, +1}"
to {-I, +1}, for k ~ O(2 n / 3 ), then there exist distributions D for which (i) there
exists a choice of weights such that the network exactly classifies its inputs according
to D, but (ii) the learning algorithm will have probability at least .01 of finding a
choice of weights which in fact has error greater than E.
CONCLUSION
We have given theoretical lower and upper bounds on the sample size vs. net
size needed such that valid generalization can be expected. The exact constants we
have given in these formulae are still quite crudej it may be expected that the actual
values are closer to 1. The logarithmic factor in Corollary 4 may also not be needed,
at least for the types of distributions and architectures seen in practice. Widrow's
experience supports this conjecture [wid87]. However, closing the theoretical gap
lower bound on the worst case
between the O( ': log ~) upper bound and the (2 (
sample size for architectures with one hidden layer of threshold units remains an
interesting open problem. Also, apart from our upper bound, the case of multiple
hidden layers is largely open. Finally, our bounds are obtained under the assumption
that the node functions are linear threshold functions (or at least Boolean valued).
We conjecture that similar bounds also hold for classes of real valued functions such
as sigmoid functions, and hope shortly to establish this.
1f)
Acknowledgements
We would like to thank Ron Rivest for suggestions on improving the bounds given in
Corollaries 4 and 5 in an earlier draft of this paper, and Nick Littlestone for many
helpful comments. The research of E. Baum was performed by the Jet Propulsion Laboratory, California Institute of Technology, as part of its Innovative Space
What Size Net Gives Valid Generalization?
Technology Center, which is sponsored by the Strategic Defense Initiative Organization/Innovative Science and Technology through an agreement with the National
Aeronautics and Space Administration (NASA). D. Haussler gratefully acknowledges the support of ONR grant NOOOI4-86-K-0454. Part of this work was done
while E. Baum was visiting UC Santa Cruz.
References
[b88]BAUM, E. B., (1988) On the capabilities of multilayer perceptrons, J. of Complexity, 4, 1988, ppI93-215.
[behw87a]BLUMER, A., EHRENFEUCHT, A. HAUSSLER, D., WARMUTH, M.,
(1987), Occam's Razor, Int: Proc. Let., 24, 1987, pp377-380.
[behw87b]BLUMER, A., EHRENFEUCHT, A. HAUSSLER, D., WARMUTH, M.,
(1987), Learnability and the Vapnik-Chervonenkis dimension, UC Santa Cruz Tech.
Rep. UCSC-CRL-87-20 (revised Oct., 1988) and J. ACM, to appear.
[cov65]COVER, T., (1965), Geometrical and statistical properties of systems of
linear inequalities with applications to pattern recognition, IEEE Trans. Elect.
Comp., V14, pp326-334.
[dev88]DEVROYE, L., (1988), Automatic pattern recognition, a study of the probability of error, IEEE Trans. P AMI, V10, N4, pp530-543.
[dswshhj87]DENKER J., SCHWARTZ D., WITTNER B., SOLLA S., HOP FIELD
J., HOWARD R., JACKEL L., (1987), Automatic learning, rule extraction, and
generalization, Complex Systems 1 pp877-922.
[dh73]DUDA, R., HART, P., (1973), Pattern clallification and scene analysis, Wiley, New York.
[ehkv87]EHRENFEUCHT, A., llAUSSLER, D., KEARNS, M., VALIANT, L., (1987),
A general lower bound on the number of examples needed for learning, UC Santa
Cruz Tech. Rep. UCSC-CRL-87-26 and Information and Computation, to appear.
[hau88]HAUSSLER, D., (1988), Quantifying inductive bias: AI learning algorithms
and Valiant's learning framework, Artificial Intelligence, 36, 1988, pp177-221.
[hin87]HINTON, G., (1987), Connectionist learning procedures, Artificial Intelligence, to appear.
[lit88]LITTLESTONE, N., (1988) Learning quickly when irrelevant attributes abound:
a new linear threshold algorithm, Machine Learning, V2, pp285-318.
[odI88]ODLYZKO, A., (1988), On subspaces spanned by random selections of ?1
vectors, J. Comb. Th. A, V47, Nt, pp124-133.
[poI84]POLLARD, D., (1984), Convergence 0/ stochastic procelles, Springer-Verlag,
New York. '
89
90
Baum and Haussler
[qs88]QUIAN, N., SEJNOWSKI, T. J., (1988), Predicting the secondary structure
of globular protein using neural nets, Bull. Math. Biophys. 5, 115-137.
[rum87]RUMELHART, D., (1987), personal communication.
[sau72]SAUER, N., (1972), On the density of families of sets, J. Comb. Th. A,
V13, 145-147.
[sr87]SEJNOWSKI, T.J., ROSENBERG, C. R., (1987), NET Talk: a parallel network that learns to read aloud, Complex Systems, vi pp145-168.
[ts88]TESAURO G., SEJNOWSKI, T. J.,(1988), A 'neural' network that learns to
play backgammon, in Neural Information Procelling Sy,tem" ed. D.Z. Anderson,
AlP, NY, pp794-803.
[val84]VALIANT, L. G., (1984), A theory of the learnable, Comm. ACM V27, Nil
pp1l34-1142.
[vc71]VAPNIK, V.N., Chervonenkis, A. Ya., (1971), On the uniform convergence of
relative frequencies of events to their probabilities, Th. Probe and its Applications,
V17, N2, pp264-280.
[vap82]VAPNIK, V.N., (1982), E,timation of Dependence, Ba,ed on Empirical Data,
Springer Verlag, NY.
[wd81]WENOCUR, R. S., DUDLEY, R. M., (1981) Some special Vapnik-Chervonenkis
classes, Discrete Math., V33, pp313-318.
[wid87]WIDROW, B, (1987) ADALINE and MADALINE - 1963, Plenary Speech,
Vol I, Proc. IEEE 1st Int. Conf. on Neural Networks, San Diego, CA, pp143-158.
| 154 |@word briefly:1 version:1 polynomial:1 stronger:1 maz:1 duda:1 open:2 initial:1 chervonenkis:4 ketch:1 nt:1 si:1 yet:1 cruz:4 fn:1 partition:1 offunctions:1 designed:1 sponsored:1 v:2 half:2 fewer:5 guess:2 intelligence:2 warmuth:2 plane:2 completeness:1 provides:1 node:36 ron:1 draft:1 math:2 simpler:1 shatter:1 ucsc:2 initiative:1 consists:1 specialize:1 dimen:1 comb:2 manner:2 indeed:1 expected:5 nor:1 ifm:1 spherical:1 little:1 actual:2 lll:1 cardinality:3 considering:2 abound:1 classifies:1 underlying:5 rivest:1 what:8 kind:1 developed:1 finding:1 nj:2 certainty:2 nf:1 exactly:1 rm:1 schwartz:1 unit:11 val84:2 grant:1 appear:4 overestimate:1 educated:1 approximately:3 might:1 nz:2 conversely:2 directed:1 v27:1 practice:2 implement:1 procedure:1 empirical:1 significantly:1 confidence:4 protein:2 suggest:1 get:2 onto:1 selection:1 context:1 applying:2 fruitful:1 center:1 baum:9 regardless:1 independently:2 pure:1 slicing:2 rule:4 haussler:10 spanned:1 his:1 mq:1 notion:1 target:1 construction:3 play:1 diego:1 exact:1 us:2 agreement:1 rumelhart:1 recognition:3 predicts:1 database:2 labeled:1 ft:2 disjunctive:1 worst:2 s1i:1 v13:1 connected:3 solla:1 removed:1 comm:1 complexity:2 pp1l34:1 personal:1 eric:1 learner:1 sink:2 easily:2 various:1 represented:3 talk:1 train:1 distinct:1 mn2:1 describe:1 sejnowski:3 artificial:2 dichotomy:9 choosing:2 aloud:1 quite:2 whose:1 larger:1 valued:4 say:1 itself:2 sequence:4 net:31 adaline:1 eto:1 convergence:2 produce:3 generating:1 prool:2 derive:1 widrow:3 v10:1 differ:1 closely:1 attribute:1 stochastic:2 vc:8 alp:1 implementing:1 globular:1 vcdim:9 generalization:12 hold:3 normal:1 mapping:1 predict:1 proc:2 jackel:1 largest:2 hope:3 mit:2 always:1 varying:1 rosenberg:1 conjunction:1 corollary:13 backgammon:2 fooled:1 tech:2 realizable:1 helpful:1 shattered:4 hidden:16 issue:1 among:2 denoted:1 logn:1 priori:1 special:2 fairly:1 uc:3 equal:1 construct:1 field:1 extraction:1 hop:1 lit:1 tem:1 future:9 connectionist:1 serious:1 few:2 randomly:3 ve:1 national:1 individual:1 undoubtably:1 negation:3 attempt:2 freedom:1 organization:1 interest:2 nl:1 implication:1 edge:12 capable:1 closer:1 necessary:1 experience:1 sauer:2 afi:1 divide:1 logarithm:3 littlestone:2 theoretical:2 minimal:1 plenary:1 classify:6 earlier:1 boolean:3 cover:2 bull:1 strategic:1 deviation:1 subset:2 uniform:2 srm:1 recognizing:1 too:1 learnability:1 kn:3 chooses:1 st:3 density:1 ie:4 physic:1 quickly:1 unavoidable:1 containing:2 literal:1 conf:1 expert:1 coefficient:1 int:2 vi:1 performed:1 try:2 doing:1 start:1 parallel:2 capability:1 minimize:2 ni:1 accuracy:3 loaded:3 phoneme:1 largely:1 sy:1 identify:1 generalize:3 thumb:1 produced:1 comp:1 classified:2 whenever:2 ed:2 definition:1 frequency:2 obvious:1 e2:1 associated:4 proof:3 di:1 treatment:1 noooi4:1 ezample:1 back:3 nasa:1 appears:1 dh73:2 done:1 anderson:1 just:1 sketch:1 receives:2 propagation:3 defines:1 measurability:1 contain:2 inductive:1 hence:4 read:1 laboratory:1 ehrenfeucht:3 ll:1 razor:1 noted:1 elect:1 confusion:1 nde:1 geometrical:1 ef:1 fi:1 common:1 sigmoid:1 overview:1 exponentially:4 discussed:1 v14:1 ai:3 automatic:2 closing:1 gratefully:1 etc:1 base:2 aeronautics:1 recent:1 perspective:1 irrelevant:1 apart:1 tesauro:1 verlag:2 inequality:3 onr:1 rep:2 seen:1 greater:4 somewhat:1 ii:2 full:1 multiple:1 jet:1 af:8 wittner:1 hart:1 regression:1 multilayer:1 df:2 represent:1 ion:1 source:3 appropriately:1 rest:1 comment:1 induced:7 n7:1 ee:3 near:1 counting:1 feedforward:14 architecture:17 approaching:6 associating:1 administration:1 expression:1 quian:1 defense:1 pollard:1 york:2 speech:1 santa:4 se:4 reduced:1 exist:1 disjoint:2 correctly:10 per:2 diverse:1 discrete:1 vol:1 threshold:18 nevertheless:1 drawn:10 deleted:1 neither:1 verified:2 graph:7 fraction:13 year:1 wand:1 counteract:1 inverse:1 prob:1 family:1 reasonable:2 vn:1 appendix:1 bound:23 layer:12 fl:2 scene:1 argument:1 span:1 innovative:2 conjecture:2 department:1 according:3 remain:1 slightly:1 em:11 n4:1 making:1 restricted:1 ln:2 remains:1 discus:1 fail:2 needed:6 know:1 end:1 generalizes:1 apply:2 denker:1 probe:1 v2:1 appropriate:4 dudley:1 permission:1 shortly:1 existence:1 jn:3 remaining:1 vc71:1 giving:1 establish:1 hypercube:2 question:4 realized:1 occurs:1 dependence:2 visiting:1 exhibit:1 subspace:2 thank:1 capacity:2 propulsion:1 assuming:4 devroye:1 procelling:1 concomitant:1 minimizing:1 madaline:1 sharper:1 ba:1 upper:6 revised:1 howard:1 attacked:1 january:1 hinton:1 communication:1 arbitrary:5 david:1 introduced:1 pair:3 nick:1 california:2 learned:1 vap82:4 trans:2 address:3 suggested:2 below:1 pattern:4 including:1 shifting:1 event:1 natural:2 predicting:1 technology:3 ne:1 reprint:1 acknowledges:1 text:1 disagrees:1 acknowledgement:1 determining:1 relative:2 fully:3 inde:1 interesting:1 suggestion:1 analogy:1 acyclic:1 lit88:2 degree:1 sufficient:1 consistent:4 thresholding:1 playing:1 share:1 occam:1 placed:1 last:1 free:1 bias:1 institute:1 taking:1 slice:3 dimension:9 valid:10 world:1 computes:1 doesn:1 made:1 san:1 odlyzko:1 incoming:4 knew:1 timation:1 why:1 learn:2 ca:2 composing:1 ignoring:1 improving:1 complex:2 protocol:1 domain:1 linearly:1 whole:1 s2:3 arise:1 n2:1 en:2 ny:2 wiley:1 wlog:2 lc:1 position:1 crude:1 learns:2 theorem:14 ofn:1 removing:1 load:3 formula:1 learnable:1 explored:1 a3:2 exists:4 vapnik:8 valiant:5 biophys:1 gap:1 b88:3 entropy:1 intersection:1 logarithmic:2 desire:1 ordered:2 springer:2 determines:1 chance:1 acm:2 oct:1 nem:4 goal:1 blumer:2 quantifying:1 crl:2 considerable:2 specifically:1 uniformly:1 ami:1 lemma:1 kearns:1 called:4 total:2 secondary:2 pas:1 nil:1 ya:1 ew:1 perceptrons:1 support:2 latter:2 outgoing:1 princeton:2 |
591 | 1,540 | General-purpose localization of textured
?
?
Image
regions
Rutb Rosenboltz?
XeroxPARC
3333 Coyote Hill Rd.
Palo Alto, CA 94304
Abstract
We suggest a working definition of texture: Texture is stuff that is
more compactly represented by its statistics than by specifying the
configuration of its parts. This definition suggests that to fmd
texture we look for outliers to the local statistics, and label as
texture the regions with no outliers. We present a method, based
upon this idea, for labeling points in natural scenes as belonging to
texture regions, while simultaneously allowing us to label lowlevel, bottom-up cues for visual attention. This method is based
upon recent psychophysics results on processing of texture and
popout.
1 WHAT IS TEXTURE, AND WHY DO WE WANT TO
FIND IT?
In a number of problems in computer VlSlon and image processing, one must
distinguish between image regions that correspond to objects and those which
correspond to texture, and perform different processing depending upon the type of
region. Current computer vision algorithms assume one magically knows this
region labeling. But what is texture? We have the notion that texture involves a
pattern that is somehow homogeneous, or in which signal changes are "too
complex" to describe, so that aggregate properties must be used instead (Saund,
1998). There is by no means a firm division between texture and objects; rather, the
characterization often depends upon the scale of interest (Saund, 1998).
? Email: rruth@parc.xerox.com
818
R. Rosenholtz
Ideally the defmition of texture should probably depend upon the application. We
investigate a definition that we believe will be of fairly general utility: Texture is
stuff that seems to belong to the local statistics. We propose extracting several
texture features, at several different scales, and labeling as texture those regions
whose feature values are likely to have come from the local distribution.
Outliers to the local statistics tend to draw our attention (Rosenholtz, 1997, 1998).
The phenomenon is often referred to as "popout." Thus while labeling (locally)
statistically homogeneous regions as texture, we can simultaneously highlight
salient outliers to the local statistics. Our revised defmition is that texture is the
absence of popout.
In Section 2, we discuss previous work in both human perception and in fmding
texture and regions of interest in an image. In Section 3, we describe our method.
We present and discuss results on a number of real images in Section 4.
2 PREVIOUS WORK
See (Wolfe, 1998) for a review of the visual search literature. Popout is typically
studied using simple displays, in which an experimental subject searches for the
unusual, target item, among the other, distractor items. One typically attempts to
judge the "saliency," or degree to which the target pops out, by studying the
efficiency of search for that item. Typically popout is modeled by a relatively lowlevel operator, which operates independently on a number of basic features of the
image, including orientation, contrast/color, depth, and motion. In this paper, we
look only at the features of contrast and orientation.
Within the image-processing field, much of the work in fmding texture has defmed
as texture any region with a high luminance variance, e.g. Vaisey & Gersho (1992).
Unfortunately, the luminance variance in a region containing an edge can be as high
as that in a textured region. Won & Park (1997) use model fitting to detect image
blocks containing an edge, and then label blocks with high variance as containing
texture.
Recently, several computer vision researchers have also tackled this problem.
Leung & Malik (1996) found regions of completely deterministic texture. Other
researchers have used the defmition that if the luminance goes up and then down
again (or vice versa) it's texture (Forsyth et aI, 1996). However, this method will
treat lines as if they were texture. Also, with no notion of similarity within a texture
(also lacking in the image-processing work), one would mark a "fault" in a texture
as belonging to that texture. This would be unacceptable for a texture synthesis
application, in which a routine that tried to synthesize such a texture would most
likely fail to reproduce the (highly visible) fault. More recently, Shi and Malik
(1998) presented a method for segmenting images based upon texture features. Their
method performs extremely well at the segmentation task, dividing an image into
regions with internal similarity that is high compared to the similarity across
regions. However, it is difficult to compare with their results, since they do not
explicitly label a subset of the resulting regions as texture. Furthermore, this
method may also tend to mark a "fault" in a texture as belonging to that texture.
This is both because the method is biased against separating out small regions, and
because the grouping of a patch with one region depends as much upon the
difference between that patch and other regions as it does upon the similarity
between the patch and the given region.
Very little computer vision work has been done on attentional cues. Milanese et al
(1993) found salient image regions using both top-down information and a bottomup "conspicuity" operator, which marks a local region as more salient the greater the
General-Purpose Localization o/Textured Image Regions
819
difference between a local feature value and the mean feature value in the
surrounding region. However. for the same difference in means. a local region is
less salient when there is a greater variance in the feature values in the surrounding
region (Duncan & Humphreys. 1989; Rosenholtz. 1997). We use as our saliency
measure a test for outliers to the local distribution. This captures. in many cases. the
dependence of saliency on difference between a given feature value and the local
mean. relative to the local standard deviation. We will discuss our saliency measure
in greater detail in the following section.
3
FINDING TEXTURE AND REGIONS OF INTEREST
We compute multiresolution feature maps for orientation and contrast. and then look
for outliers in the local orientation and contrast statistics. We do this by fast
creating a 3-level Gaussian pyramid representation of the image. To extract
contrast. we filter the pyramid with a difference of circularly symmetric Gaussians.
The response of these filters will oscillate. even in a region with constant-contrast
texture (e.g. a sinewave pattern). We approximate a computation of the maximum
response of these filters over a small region by fast squaring the filter responses.
and then filtering the contrast energy with an appropriate Gaussian. Finally. we
threshold the contrast to eliminate low-contrast regions ("flat" texture). These
thresholds (one for each scale) were set by examining the visibility of sinewave
patterns of various spatial frequencies.
We compute orientation in a simple and biologically plausible way. using Bergen &
Landy's (1991) "back pocket model" for low-level computations:
1.
Filter the pyramid with horizontal. vertical. and ?45? oriented Gaussian second
derivatives.
2.
Compute opponent energy by squaring the filter outputs. pooling them over a
region 4 times the scale of the second derivative filters. and subtracting the
vertical from the horizontal response and the +45 0 from the _45 0 response.
3.
Normalize the opponent energy at each scale by dividing by the total energy in
the 4 orientation energy bands at that scale.
The result is two images at each scale of the pyramid. To a good approximation. in
regions which are strongly oriented. these images represent kcos(26) and ksin(26).
where 6 is the local orientation at that scale. and k is a value between 0 and 1 which
is related to the local orientation specificity. Orientation estimates from points with
low specificity tend to be very noisy. In images of white noise. 80% of the
estimates of k fall below 0.5. therefore with 80% confidence. an orientation
specificity of k>0.5 did not occur due to chance. We use this value to threshold out
orientation estimates with low "orientedness.??
We then estimate D, the local feature distribution, for each feature and scale, using
the method of Parzen windows. The blurring of the distribution estimate by the
Parzen window mimics uncertainty in estimates of feature values by the visual
system. We collect statistics over a local integration region. For texture processing.
the size of this region is ind.ependent of viewing distance, and is roughly lOS in
diameter, where S is the support of the Gaussian 2nd derivative filters used to extract
the texture features (Kingdom & Keeble, 1997; Kingdom et ai, 1995).
We next compute a non-parametric measure of saliency:
(
saliency = -IO~
P(v ID)
maxP(x ID)
%
,
)
(1)
R. Rosenholtz
820
Note that if D were Gaussian N(~,a2), this simplifies to
(X_tt)2
(2)
'1a 2
which should be compared to the standard parametric test for outliers, which uses
the measure (x - tt)/a. Our saliency measure is essentially a more general, nonparametric form of this measure (i.e. it does not assume a Gaussian distribution).
Points with saliency less than 0.5 are labeled as candidate texture points. If D were
Gaussian, this would correspond to feature estimates within one standard deviation
of the mean. Points with saliency greater than 3.1 are labeled as candidates for
bottom-up attentional cues. If D were Gaussian, this would correspond to feature
estimates more than 2.50 from the mean, a standard parametric test for outliers.
One could, of course, keep the raw saliency values, as a measure of the likelihood
that a region contained texture, rather than setting a hard threshold. We use a hard
threshold in our examples to better display the results. Both the texture images and
the region of interest images are median-filtered to remove extraneous points.
4 EXPERIMENTAL RESULTS
Figure 3 shows several example images. Figures 2, 3, and 4 show texture found at
each scale of processing. The striped and checkered patterns represent oriented and
homogeneous contrast texture, respectively. The absence of an image in any of
these figures means that no texture of the given type was found in that image at the
given scale. Note that we perform no segmentation of one texture from another.
For the building image, the algorithm labeled bricks and window panes as fme-scale
texture, and windows and shutters as coarser-scale texture. The leopard skin and
low-frequency stripes in the lower right comer of the leopard image were correctly
labeled as texture. In the desk image, the "wood" texture was correctly identified.
The regular pattern of windows were marked as texture in the hotel image. In the
house image, the wood siding, trees, and part of the grass were labeled as texture
(much of the grass was low contrast and labeled as "flat" texture). One of the
bushes is correctly identified as having coarser texture than the other has. In the
lighthouse image, the house sans window, fence, and tower were marked, as well as
a low-frequency oriented pattern in the clouds.
Figure 5 shows the regions of interest that were found (the striped and plaid patterns
here have no meaning but were chosen for maximum visibility). Most complex
natural scenes had few interesting low-level attentional areas. In the lighthouse
image, the life preserver is marked. In the hotel, curved or unusual angular
windows are identified as attentional cues, as well as the top of the building. Both
of these results are in agreement with psychophysical results showing that observers
quickly identify curved or bent lines among straight lines (reviewed in Wolfe,
1998). The simpler desk scene yields more intuitive results, with each of the 3
objects labeled, as well as the phone cord.
Bottom-up attentional cues are outliers to the local distribution of features, and we
have suggested that texture is the absence of such outliers. This definition captures
some of the intuition that texture is homogeneous and statistical in nature. We
presented a method for fmding contrast and orientation outliers, and results both on
localizing texture and on finding popout in natural images. For the simple desk
image, the algorithm highlights salient regions that correspond to our notions of the
important objects in the scene. On complicated natural scenes, its results are less
intuitive; suggesting that search in natural scenes makes use of higher-level
General-Purpose Localization o/Textured Image Regions
821
processing such as grouping into objects. This result should not be terribly
surprising, but serves as a useful check on simple low-level models of visual
attention. The algorithm does a good job of identifying textured regions at a
number of different scales, with the results perhaps more intuitive at finer scales.
Acknowledgments
This work was partially supported by an NRC postdoctoral award at NASA Ames.
Many thanks to David Marimont and Eric Saund for useful discussions.
References
J. R. Bergen and M. S. Landy (1991), "Computational modeling of visual texture
segmentation," Computational Models of Visual Processing, Landy and Movshon
(eds.), pp. 252-271, MIT Press, Cambridge, MA.
J. Duncan and G. Humphreys (1989), "Visual search and stimulus similarity,"
Psych. Review 96, pp. 433-458.
D. Forsyth, J. Malik, M. Fleck, H. Greenspan, T. Leung, S. Belongie, C. Carson,
and C. Bregler (1996), "Finding pictures of objects in collections of images," ECCV
Workshop on Object Representation, Cambridge.
F. A. A. Kingdom, D. Keeble, D., and B. Moulden (1995), "Sensitivity to
orientation modulation in micropattern-based textures," Vis. Res. 35, 1, pp. 79-91.
F. A. A. Kingdom and D. Keeble (1997), "The mechanism for scale invariance in
orientation-defined textures." Invest. Ophthal. and Vis . Sci. (Suppl.) 38, 4, p. 636.
T. K. Leung and J. Malik (1996), "Detecting, localizing, and grouping repeated
scene elements from an image," Proc. 4th European Con! On Computer Vision,
1064, 1, pp. 546-555, Springer-Verlag, Cambridge.
R. Milanese, H. Wechsler, S. Gil, J. -M. Bost, and T. Pun (1993), "Integration of
bottom-up and top-down cues for visual attention using non-linear relaxation,"
Proc. IEEE CVPR, pp. 781-785, IEEE Computer Society Press, Seattle.
R. Rosenholtz (1997), "Basic signal detection theory model does not explain search
among heterogeneous distractors." Invest. Ophthal. and Vis. Sci. (Suppl.) 38, 4, p.
687.
R. Rosenholtz (1998), "A simple saliency model explains a number of motion
popout phenomena." Invest. Ophthal. and Vis. Sci. (Suppl.) 39,4, p. 629.
E. Saund (1998), "Scale and the ShapelTexture Continuum," Xerox Internal
Technical Memorandum.
J. Shi and J. Malik (1998), "SelfInducing Relational Distance and its Application to
Image Segmentation," Proc. jth European Con! on Computer Vision, Burkhardt and
Neumann (eds.), 1406, 1, pp. 528-543, Springer, Freiburg.
J. Vaisey and A. Gersho (1992), "Image compression with variable block size
segmentation." IEEE Trans. Signal Processing 40,8, pp. 2040-2060.
J. M. Wolfe (1998), "Visual search: a review," Attention, H. Pashler (ed.), pp. 1374, Psychology Press Ltd., Hove, East Sussex, UK.
C. S. Won and D. K. Park (1997), "Image block classification and variable block
size segmentation using a model-fitting criterion," Opt. Eng. 36, 8, pp. 2204-2209.
822
R. Rosenholtz
Figure I: Original images.
(a)
(b)
Figure 2: Fine-scale texture. (a) oriented texture, (b) homogeneous contrast
texture.
-.
~"
.
~
-
&
(a)
....
I.
...
,
..'
1".
.~
?, ?
?
-
..
'
..
?
(b)
Figure 3: Medium-scale texture. (a) oriented texture, (b) homogeneous contrast
texture.
General-Purpose Localization of Textured Image Regions
823
(a)
(b)
Figure 4: Coarse-scale texture. (a) oriented texture, (b) homogeneous contrast
texture .
.. .
, I,
?
?
Figure 5: Regions of interest.
| 1540 |@word compression:1 seems:1 nd:1 tried:1 eng:1 configuration:1 current:1 com:1 surprising:1 must:2 visible:1 visibility:2 remove:1 grass:2 cue:6 item:3 filtered:1 detecting:1 coarse:1 characterization:1 ames:1 simpler:1 pun:1 unacceptable:1 fitting:2 roughly:1 distractor:1 little:1 window:7 alto:1 medium:1 what:2 psych:1 finding:3 stuff:2 uk:1 segmenting:1 local:17 treat:1 io:1 id:2 modulation:1 studied:1 specifying:1 suggests:1 collect:1 statistically:1 acknowledgment:1 block:5 area:1 confidence:1 regular:1 specificity:3 suggest:1 operator:2 pashler:1 deterministic:1 map:1 shi:2 go:1 attention:5 lowlevel:2 independently:1 moulden:1 identifying:1 notion:3 memorandum:1 target:2 homogeneous:7 us:1 agreement:1 wolfe:3 synthesize:1 element:1 stripe:1 coarser:2 labeled:7 bottom:4 cloud:1 capture:2 region:42 cord:1 intuition:1 ideally:1 depend:1 parc:1 localization:4 upon:8 division:1 efficiency:1 textured:6 completely:1 compactly:1 blurring:1 comer:1 eric:1 represented:1 various:1 surrounding:2 fast:2 describe:2 labeling:4 aggregate:1 firm:1 whose:1 plausible:1 cvpr:1 maxp:1 statistic:7 noisy:1 propose:1 subtracting:1 multiresolution:1 intuitive:3 normalize:1 los:1 invest:3 seattle:1 neumann:1 object:7 depending:1 job:1 dividing:2 involves:1 come:1 judge:1 plaid:1 filter:8 human:1 terribly:1 viewing:1 explains:1 opt:1 bregler:1 leopard:2 sans:1 milanese:2 continuum:1 a2:1 purpose:4 proc:3 label:4 palo:1 vice:1 mit:1 gaussian:8 rather:2 greenspan:1 fence:1 likelihood:1 check:1 contrast:15 detect:1 bergen:2 squaring:2 leung:3 typically:3 eliminate:1 reproduce:1 among:3 orientation:14 classification:1 extraneous:1 spatial:1 integration:2 psychophysics:1 fairly:1 field:1 having:1 park:2 look:3 mimic:1 stimulus:1 few:1 oriented:7 simultaneously:2 attempt:1 detection:1 interest:6 investigate:1 highly:1 edge:2 tree:1 re:1 brick:1 modeling:1 localizing:2 deviation:2 subset:1 coyote:1 examining:1 too:1 thanks:1 sensitivity:1 synthesis:1 parzen:2 quickly:1 again:1 containing:3 creating:1 derivative:3 suggesting:1 forsyth:2 explicitly:1 depends:2 vi:4 saund:4 observer:1 complicated:1 fme:1 variance:4 correspond:5 saliency:11 identify:1 yield:1 raw:1 researcher:2 finer:1 straight:1 explain:1 ed:3 email:1 definition:4 against:1 energy:5 hotel:2 frequency:3 pp:9 con:2 color:1 distractors:1 segmentation:6 pocket:1 routine:1 back:1 nasa:1 higher:1 response:5 done:1 strongly:1 furthermore:1 angular:1 working:1 horizontal:2 somehow:1 perhaps:1 believe:1 building:2 symmetric:1 white:1 ind:1 defmed:1 sussex:1 won:2 criterion:1 carson:1 hill:1 freiburg:1 tt:1 performs:1 motion:2 image:39 meaning:1 recently:2 belong:1 defmition:3 versa:1 cambridge:3 ai:2 rd:1 had:1 similarity:5 recent:1 phone:1 lighthouse:2 verlag:1 fault:3 life:1 greater:4 signal:3 technical:1 bent:1 award:1 basic:2 heterogeneous:1 vision:5 essentially:1 popout:7 represent:2 pyramid:4 suppl:3 want:1 fine:1 median:1 biased:1 probably:1 subject:1 tend:3 pooling:1 marimont:1 extracting:1 psychology:1 identified:3 idea:1 simplifies:1 shutter:1 utility:1 ltd:1 movshon:1 oscillate:1 useful:2 burkhardt:1 nonparametric:1 desk:3 locally:1 band:1 diameter:1 conspicuity:1 rosenholtz:7 gil:1 correctly:3 salient:5 threshold:5 luminance:3 relaxation:1 wood:2 uncertainty:1 patch:3 draw:1 duncan:2 distinguish:1 display:2 tackled:1 occur:1 striped:2 scene:7 flat:2 extremely:1 pane:1 relatively:1 xerox:2 belonging:3 magically:1 across:1 biologically:1 outlier:11 discus:3 fail:1 mechanism:1 know:1 gersho:2 serf:1 unusual:2 studying:1 gaussians:1 opponent:2 appropriate:1 original:1 top:3 landy:3 wechsler:1 society:1 psychophysical:1 malik:5 skin:1 parametric:3 dependence:1 distance:2 attentional:5 separating:1 sci:3 tower:1 bost:1 modeled:1 fmding:3 difficult:1 unfortunately:1 kingdom:4 perform:2 allowing:1 vertical:2 revised:1 curved:2 relational:1 david:1 pop:1 trans:1 suggested:1 below:1 pattern:7 perception:1 including:1 natural:5 picture:1 extract:2 review:3 literature:1 relative:1 lacking:1 highlight:2 interesting:1 filtering:1 degree:1 eccv:1 course:1 supported:1 jth:1 fall:1 depth:1 collection:1 approximate:1 checkered:1 keep:1 belongie:1 bottomup:1 postdoctoral:1 search:7 why:1 reviewed:1 nature:1 ca:1 complex:2 european:2 did:1 nrc:1 noise:1 repeated:1 fmd:1 referred:1 candidate:2 house:2 humphreys:2 down:3 sinewave:2 showing:1 grouping:3 workshop:1 circularly:1 texture:68 likely:2 visual:9 contained:1 partially:1 springer:2 chance:1 ma:1 marked:3 absence:3 change:1 hard:2 operates:1 total:1 invariance:1 experimental:2 east:1 fleck:1 internal:2 mark:3 support:1 bush:1 phenomenon:2 |
592 | 1,541 | Active Noise Canceling using Analog NeuroChip with On-Chip Learning Capability
Jung-Wook Cho and Soo-Young Lee
Computation and Neural Systems Laboratory
Department of Electrical Engineering
Korea Advanced Institute of Science and Technology
373-1 Kusong-dong, Yusong-gu, Taejon 305-701, Korea
sylee@ee.kaist.ac.kr
Abstract
A modular analogue neuro-chip set with on-chip learning capability is
developed for active noise canceling. The analogue neuro-chip set
incorporates the error backpropagation learning rule for practical
applications, and allows pin-to-pin interconnections for multi-chip
boards. The developed neuro-board demonstrated active noise
canceling without any digital signal processor. Multi-path fading of
acoustic channels, random noise, and nonlinear distortion of the loud
speaker are compensated by the adaptive learning circuits of the
neuro-chips. Experimental results are reported for cancellation of car
noise in real time.
1
INTRODUCTION
Both analog and digital implementations of neural networks have been reported.
Digital neuro-chips can be designed and fabricated with the help of well-established
CAD tools and digital VLSI fabrication technology [1]. Although analogue neurochips have potential advantages on integration density and speed over digital chips[2],
they suffer from non-ideal characteristics of the fabricated chips such as offset and
nonlinearity, and the fabricated chips are not flexible enough to be used for many
different applications. Also, much careful design is required, and the fabricated chip
characteristics are fairly dependent upon fabrication processes.
For the implementation of analog neuro-chips, there exist two different approaches, i.e.,
with and without on-chip learning capability [3,4], Currently the majority of analog
neuro-chips does not have learning capability, while many practical applications
require on-line adaptation to continuously changing environments, and must have online adaptation learning capability. Therefore neuro-chips with on-chip learning
capability are essential for such practical applications. Modular architecture is also
665
Active Noise Canceling with Analog On-Chip Learning Neuro-Chip
advantageous to provide flexibility of implementing many large complex systems from
same chips.
Although many applications have been studied for analog neuro-chips, it is very
important to find proper problems where analog neuro-chips may have potential
advantages over popular DSPs. We believe applications with analog input/output
signals and high computational requirements are those good problems. For example,
active noise controls [5] and adaptive equalizers [6,7] are good applications for analog
neuro-chips.
In this paper we report a demonstration of the active noise canceling, which may have
many applications in real world. A modular analog neuro-chip set is developed with
on-chip learning capability, and a neuro-board is fabricated from multiple chips with
PC interfaces for input and output measurements. Unlike our previous implementations
for adaptive equalizers with binary outputs [7], both input and output values are
analogue in this noise canceling.
..-.---1-1.... 0
II'
.
xl, ~t---(~.-~-.lI'"'"'"iI--- (~
Figure 1. Block diagram of a synapse cell
2
~
._._.- i')
Figure 2. Block diagram of a neuron cell
ANALOG NEURO-CHIP WITH ON-CHIP LEARNING
We had developed analog neuro-chips with error backpropagation learning capability.
With the modular architecture the developed analog neuro-chip set consists of a
synapse chip and a neuron chip.[8] The basic cell of the synapse chip is shown in
Figure 1. Each synapse cell receives two inputs, i.e., pre-synaptic neural activation x
and error correction term 8, and generates two outputs, i.e., feed-forward signal wx and
back-propagated error w8. Also it updates a stored weight w by the amount of x8.
Therefore, a synapse cell consists of three multiplier circuits and one analogue storage
for the synaptic weight. Figure 2 shows the basic cell in the neuron chip, which collects
signals from synapses in the previous layer and distributes to synapses in the following
layer. Each neuron body receives two inputs, i.e., post-synaptic neural activation 0 and
back-propagated error 8 from the following layer, and generates two outputs, i.e.,
Sigmoid-squashed neural activation 0 and a new backpropagated error 8 multiplied by
a bell-shaped Sigmoid-derivative. The backpropagated error may be input to the
synapse cells in the previous layer.
To provide easy connectivity with other chips, the two inputs of the synapse cell are
represented as voltage, while the two outputs are as currents for simple current
summation. On the other hand the inputs and outputs of the neuron cell are represented
as currents and voltages, respectively. For simple pin-to-pin connections between chips,
one package pin is maintained to each input and output of the chip. No time-
J.-W Cho and s.-Y. Lee
666
multiplexing is introduced, and no other control is required for multi-chip and multilayer systems. However, it makes the number of package pins the main limiting factor
for the number of synapse and neuron cells in the developed chip sets.
Although many simplified multipliers had been reported for high-density integration,
their performance is limited in linearity, resolution, and speed. For on-chip learning, it
is desirable to have high precision, and a faithful implementation of the 4-quadranr
Gilbert multipliers is used. Especially, the mUltiplier for weight updates in the synapse
cell requires high precision.[9] The synaptic weight is stored on a capacitor, and an
MaS switch is used to allow current flow from the multiplier to the capacitor during a
short time interval for weight adaptation. For applications like active noise controls [5]
and telecommunications [6,7], tapped analog delay lines are also designed and
integrated in the synapse chip. To reduce offset accumulation, a parallel analog delay
line is adopted. Same offset voltage is introduced for operational amplifiers at all
nodes [10] . Diffusion capacitors with 2.2 pF are used for the storage of the tapped
analog delay line.
In a synapse chip 250 synapse cells are integrated in a 25xl0 array with a 25-tap
analog delay line. Inputs may be applied either from the analog delay line or from
external pins in parallel. To select a capacitor in the cell for refresh, decoders are
placed in columns and rows. The actual size of the synapse cell is 14111m x 17911m,
and the size of the synapse chip is 5.05mm x 5.05mm. The chip is fabricated in a
0.811m single-poly CMOS process. On the other hand, the neuron chip has a very
simple structure, which consists of 20 neuron cells without additional circuits. The
Sigmoid circuit [3] in the neuron cell uses a differential pair, and the slope and
amplitude are controlled by a voltage-controlled resistor [II]. Sigmoid-derivative
circuit is also using differential pair with min-select circuit. The size of the neuron cell
is 177.2I1m x 62.4l1m.
Synapse
Chip
PC
Neuron
PC
Chip
,
~ _ ' Target
I
N t-.,.' ----t~
I
I
DSP
TMS320C51
N
I
Output
:
-:-r!1"iJ-[B~h:---'-:_._.-.____;--..; r:1-r:"1'
L!..r~
..._~Q_-I_""',_ _
I
Input
- L -_ _ _--I
ANN Board
Figure 3: Block diagram of the analog neuro-board
GDAB
tv.c : 32ch
ArIC : 48<:h
D1 : 16bitll
DO : 48bitll
Active Noise Canceling with Analog On-Chip Learning Neuro-Chip
667
Using these chip sets, an analog neuro-system is constructed. Figure 3 shows a brief
block diagram of the analog neuro-system, where an analogue neuro-board is
interfaced to a host computer through a GDAB (General Data Acquisition Board). The
GDAB board is specially designed for the data interface with the analogue neuro-chips.
The neuro-board has 6 synapse chips and 2 neuron chips with the 2-layer Perceptron
architecture. For test and development purposes, a DSP, ADC and DAC are installed
on the neuro-board to refresh and adjust weights.
Forward propagation time of the 2 layers Perceptron is measured as about 30 f..lsec.
Therefore the computation speed of the neuro-board is about 266 MCPS (Mega
Connections Per Second) for recall and about 200 MCUPS (Mega Connections
Updates Per Second) for error backpropagation learning. To achieve this speed with a
DSP, about 400 MIPS is required for recall and at least 600 MIPS for error-back
propagation learning.
C 1 (z)
Channel
Error
Signal
Noise
Source
Adaptive Filter
or
Multilayer Perceptron
Figure 4: Structure of a feedforward active noise canceling
3 ACTIVE NOISE CANCELING USING NEURO-CHIP
Basic architecture of the feed forward active noise canceling is shown in Figure 4. An
area near the microphone is called "quiet zone," which actually means noise should be
small in this area. Noise propagates from a source to the quiet zone through a
dispersive medium, of which characteristics are modeled as a finite impulse response
(FIR) filter with additional random noise. An active noise canceller should generate
electric signals for a loud speaker, which creates acoustic signals to cancel the noise at
the quiet zone. In general the electric-to-acoustic signal transfer characteristics of the
loud speaker is nonlinear, and the overall active noise canceling (ANC) system also
becomes nonlinear. Therefore, multilayer Perceptron has a potential advantage over
popular transversal adaptive filters based on linear-mean.-square (LMS) error
minimization.
Experiments had been conducted for car noise canceling. The reference signal for the
noise source was extracted from an engine room, while a compact car was running at
60 kmlhour. The difference of the two acoustic channels, i.e., H(z) = C1 (z) / C2 ( z) ,
addition noise n, and nonlinear characteristics of the loud speaker need be
compensated. Two different acoustic channels are used for the experiments. The first
channel Hl (z) = 0.894 + 0.447z- 1 is a minimum phase channel, while the second non-
J.-W Cho and S-Y. Lee
668
minimum
phase
channel
H2 (z)
= 0.174 + 0.6z -I + 0.6z -2 + 0.174z -3
characterizes
frequency-selective multipath fading with a deep spectral amplitude null. A simple
cubic distortion model was used for the characteristics of the loud speaker.[12] To
compare performance of the neuro-chip with digital processors, computer simulation
was first conducted with error backpropagation algorithm for a single hidden-layer
Perceptron as well as the LMS algorithm for a transversal adaptive filter. Then, the
same experimental data were provided to the developed neuro-board by a personal
computer through the GDAB.
o
ro
a::
-5
c
o
:.;:;
g -10 .
"0
CI.>
0:::
CI.>
.~
o -15
z
- 20 ~-----~------~----~------~--=-~
o
5
10
15
20
25
Signal-to-Distortion Ratio
(a)
o
15 -5
0:::
c
o
.......
~ -10
"0
CI.>
0:::
CI.>
.~
o
-15
z
-20~----~----~----~----~----~
o
5
10
15
20
Signal-to-Distortion Ratio
25
(b)
Figure 5: Noise Reduction Ratio (dB) versus Signal-to-Distortion Ratio (dB) for (a) a
simple acoustic channel HI (z) and (b) a multi-path fading acoustic channel H2 (z) _
Here, '+', '*', 'x', and '0' denote results ofLMS algorithm, neural networks simulation,
neural network simulation with 8-bit input quantization, and neuro-chips, respectively_
Active Noise Canceling with Analog On-Chip Learning Neuro-Chip
669
Results for the channels HI ( z) and H2 (z) are shown in Figures 5(a) and 5(b),
respectively. Each point in these figures denotes the result of one experiment with
different parameters. The horizontal axes represent Signal-to-Distortion Ratio (SDR)
of the speaker nonlinear characteristics. The vertical axes represent Noise Reduction
Ratio (NRR) of the active noise canceling systems. As expected, severe nonlinear
distortion of the loud speaker resulted in poor noise canceling for the LMS canceller.
However, the performance degradation was greatly reduced by neural network
canceller. With the neuro-chips the performance was worse than that of computer
simulation. Although the neuro-chip demonstrated active noise canceling and worked
better than LMS cancellers for very small SDRs, i.e. , very high nonlinear distortions,
its performance became saturated to -8 dB and -5 dB NRRs, respectively. The
performance saturation was more severe for the harder problem with the complicated
H 2 (z ) channel.
The performance degradation with neuro-chips may come from inherent limitations of
analogue chips such as limited dynamic ranges of synaptic weights and signals,
unwanted offsets and nonlinearity, and limited resolution of the learning rate and
sigmoid slope. [9] However, other side effects of the GDAB board, i.e., fixed resolution
of AID converters and D/A converters for data 110, also contributed to the performance
degradation. The input and output resolutions of the GDAB were J6 bit and 8 bit,
respectively. Unlike actual real-world systems the input values of the experimental
analogue neuro-chips are these 8-bit quantized values. As shown in Figures 5, results
of the computer simulation with 8-bit quantized target values showed much degraded
performance compared to the floating-point simulations. Therefore, a significant
portion of the poor performance in the experimental analogue system may be
contributed from the AID converters, and the analogue system may work better in real
world systems.
Actual acoustic signals are plotted in Figure 6. The top, middle, and bottom signals
denote noise , negated speaker signal, and residual noise at the quiet zone, respectively.
Figure 6: Examples of noise, negated loud-speaker canceling signal, and residual error
J.-w. Cho and s.-Y. Lee
670
4
CONCLUSION
In this paper we report an experimental results of active noise canceling using analogue
neuro-chips with on-chip learning capability. Although the its performance is limited
due to nonideal characteristics of analogue chip itself and also peripheral devices, it
clearly demonstrates feasibility of analogue chips for real world applications.
Acknowledgements
This research was supported by Korean Ministry of Information and Telecommunications.
References
[1] T. Watanabe, K. Kimura, M. Aold, T. Sakata & K. Ito (1993) A Single 1.5-V
Digital Chip for a 106 Synapse Neural Network, IEEE Trans. Neural Network,
VolA, No.3, pp.387-393.
[2J T. Morie and Y. Amemiya (1994) An All-Analog Expandable Neural Network
LSI with On-Chip Backpropagation Learning, IEEE Journal of Solid State
Circuits, vo1.29, No.9, pp.1086-1093.
[3J J.-W. Cho, Y. K. Choi, S.-Y. Lee (1996) Modular Neuro-Chip with On-Chip
Learning and Adjustable Learning Parameters, Neural Processing Letters, VolA,
No.1.
[4J J. Alspector, A. Jayakumar, S. Luna (1992) Experimental evaluation of learning in
neural microsystem, Advances in Neural Information Processing Systems 4, pp.
871-878 .
[5 J B. Widrow, et al. (1975) Adative Noise Cancelling: Principles and Applications,
Proceeding of IEEE, Vo1.63, No.12, pp.1692-1716.
[6] J. Choi, S.H. Bang, BJ. Sheu (1993) A Programmable Analog VLSI Neural
Network Processor for Communication Receivers, IEEE Transaction on Neural
Network, VolA, No.3, ppA84-495.
[7J J.-W. Cho and S.-Y. Lee (1998) Analog neuro-chips with on-chip learning
capability for adaptive nonlinear equalizer, Proc. lJCNN, pp. 581-586, May 4-9,
Anchorage, USA.
[8J J. Van der Spiegel, C. Donham, R. Etienne-Cummings, S. Fernando (1994) Large
scale analog neural computer with programmable architecture and programmable
time constants for temporal pattern analysis, Proc. ICNN, pp. 1830-1835.
[9J Y.K. Choi, K.H. Ahn, and S.Y. Lee (1996) Effects of multiplier offsets on onchip learning for analog neuro-chip, Neural Processing Letters, vol. 4, No.1, 1-8.
[1OJ T. Enomoto, T. Ishihara and M. Yasumoto (1982) Integrated tapped MaS
analogue delay line using switched-capacitor technique, Electronics Lertters,
Vo1.l8, pp.193-194.
[11 J P.B. Allen, D.R. Holberg (1987) CMOS Analog Circuit Design, Holt, Douglas
Rinehart and Winston.
[12J F. Gao and W.M. Snelgrove (1991) Adaptive linearization of a loudspeaker,
Proc. International Conference on Acoustics, Speech and Signal processing, pp.
3589-3592.
| 1541 |@word middle:1 advantageous:1 donham:1 simulation:6 solid:1 harder:1 reduction:2 electronics:1 nrr:1 current:4 cad:1 activation:3 must:1 refresh:2 wx:1 designed:3 update:3 device:1 short:1 quantized:2 node:1 anchorage:1 constructed:1 c2:1 differential:2 consists:3 expected:1 alspector:1 multi:4 actual:3 pf:1 becomes:1 provided:1 linearity:1 circuit:8 medium:1 null:1 developed:7 adc:1 fabricated:6 kimura:1 temporal:1 w8:1 unwanted:1 ro:1 demonstrates:1 control:3 engineering:1 installed:1 path:2 onchip:1 studied:1 mcups:1 collect:1 limited:4 range:1 practical:3 faithful:1 block:4 backpropagation:5 area:2 bell:1 pre:1 holt:1 storage:2 microsystem:1 gilbert:1 accumulation:1 demonstrated:2 compensated:2 resolution:4 rinehart:1 rule:1 array:1 limiting:1 target:2 us:1 tapped:3 bottom:1 electrical:1 environment:1 dynamic:1 personal:1 upon:1 creates:1 gu:1 chip:73 represented:2 modular:5 kaist:1 distortion:8 interconnection:1 sakata:1 spiegel:1 itself:1 online:1 advantage:3 adaptation:3 cancelling:1 flexibility:1 achieve:1 requirement:1 cmos:2 help:1 widrow:1 ac:1 measured:1 ij:1 come:1 filter:4 implementing:1 require:1 l1m:1 icnn:1 summation:1 correction:1 mm:2 bj:1 lm:4 purpose:1 proc:3 currently:1 tool:1 minimization:1 clearly:1 voltage:4 ax:2 dsp:3 greatly:1 dependent:1 integrated:3 hidden:1 vlsi:2 i1m:1 selective:1 overall:1 flexible:1 development:1 transversal:2 integration:2 fairly:1 shaped:1 cancel:1 report:2 inherent:1 resulted:1 floating:1 phase:2 sdr:1 amplifier:1 evaluation:1 adjust:1 severe:2 saturated:1 pc:3 korea:2 morie:1 plotted:1 column:1 delay:6 fabrication:2 conducted:2 reported:3 stored:2 cho:6 density:2 international:1 lee:7 dong:1 continuously:1 connectivity:1 fir:1 worse:1 external:1 luna:1 jayakumar:1 derivative:2 li:1 potential:3 characterizes:1 portion:1 capability:10 parallel:2 complicated:1 slope:2 square:1 degraded:1 became:1 characteristic:8 interfaced:1 expandable:1 j6:1 processor:3 synapsis:2 canceling:18 synaptic:5 acquisition:1 frequency:1 pp:8 propagated:2 popular:2 recall:2 car:3 dispersive:1 amplitude:2 actually:1 back:3 feed:2 cummings:1 response:1 synapse:17 hand:2 receives:2 horizontal:1 dac:1 nonlinear:8 propagation:2 impulse:1 believe:1 usa:1 effect:2 multiplier:6 yusong:1 laboratory:1 during:1 maintained:1 speaker:9 q_:1 allen:1 interface:2 sigmoid:5 sheu:1 analog:29 measurement:1 significant:1 cancellation:1 nonlinearity:2 had:3 ahn:1 showed:1 binary:1 der:1 minimum:2 additional:2 ministry:1 fernando:1 signal:19 ii:3 multiple:1 desirable:1 host:1 post:1 controlled:2 feasibility:1 neuro:39 basic:3 multilayer:3 represent:2 cell:17 c1:1 addition:1 interval:1 diagram:4 source:3 unlike:2 specially:1 db:4 flow:1 incorporates:1 capacitor:5 ee:1 near:1 ideal:1 feedforward:1 enough:1 easy:1 mips:2 switch:1 architecture:5 converter:3 reduce:1 suffer:1 speech:1 programmable:3 deep:1 amount:1 backpropagated:2 reduced:1 generate:1 exist:1 lsi:1 per:2 mega:2 vol:1 changing:1 douglas:1 diffusion:1 equalizer:3 package:2 letter:2 telecommunication:2 taejon:1 bit:5 layer:7 hi:2 winston:1 fading:3 worked:1 multiplexing:1 generates:2 lsec:1 speed:4 min:1 loudspeaker:1 department:1 tv:1 peripheral:1 poor:2 hl:1 aold:1 mcps:1 pin:7 adopted:1 multiplied:1 multipath:1 spectral:1 denotes:1 running:1 top:1 etienne:1 especially:1 squashed:1 quiet:4 majority:1 decoder:1 sylee:1 modeled:1 ratio:6 demonstration:1 korean:1 adative:1 implementation:4 design:2 proper:1 adjustable:1 contributed:2 negated:2 i_:1 vertical:1 neuron:12 finite:1 communication:1 dsps:1 introduced:2 pair:2 required:3 connection:3 tap:1 acoustic:9 engine:1 established:1 trans:1 pattern:1 saturation:1 oj:1 soo:1 analogue:15 residual:2 advanced:1 technology:2 brief:1 x8:1 acknowledgement:1 limitation:1 versus:1 digital:7 h2:3 ljcnn:1 switched:1 propagates:1 principle:1 row:1 l8:1 jung:1 placed:1 supported:1 side:1 allow:1 perceptron:5 institute:1 van:1 world:4 forward:3 adaptive:8 simplified:1 transaction:1 compact:1 active:17 receiver:1 channel:11 transfer:1 operational:1 anc:1 complex:1 poly:1 electric:2 main:1 noise:35 body:1 board:13 cubic:1 aid:2 precision:2 watanabe:1 resistor:1 xl:1 ito:1 young:1 choi:3 offset:5 essential:1 quantization:1 kr:1 ci:4 nonideal:1 linearization:1 gao:1 ch:1 loud:7 extracted:1 ma:2 ann:1 careful:1 bang:1 room:1 vo1:3 distributes:1 degradation:3 microphone:1 called:1 experimental:6 zone:4 select:2 xl0:1 d1:1 |
593 | 1,542 | Bayesian modeling of human concept learning
Joshua B. Tenenbaum
Department of Brain and Cognitive Sciences
Massachusetts Institute of Technology, Cambridge, MA 02139
jbt@psyche.mit.edu
Abstract
I consider the problem of learning concepts from small numbers of positive examples, a feat which humans perform routinely but which computers are rarely capable of. Bridging machine learning and cognitive
science perspectives, I present both theoretical analysis and an empirical
study with human subjects for the simple task oflearning concepts corresponding to axis-aligned rectangles in a multidimensional feature space.
Existing learning models, when applied to this task, cannot explain how
subjects generalize from only a few examples of the concept. I propose
a principled Bayesian model based on the assumption that the examples
are a random sample from the concept to be learned. The model gives
precise fits to human behavior on this simple task and provides qualitati ve
insights into more complex, realistic cases of concept learning.
1 Introduction
The ability to learn concepts from examples is one of the core capacities of human cognition.
From a computational point of view, human concept learning is remarkable for the fact that
very successful generalizations are often produced after experience with only a small number
of positive examples of a concept (Feldman, 1997). While negative examples are no doubt
useful to human learners in refining the boundaries of concepts, they are not necessary
in order to make reasonable generalizations of word meanings, perceptual categories, and
other natural concepts. In contrast, most machine learning algorithms require examples of
both positive and negative instances of a concept in order to generalize at all, and many
examples of both kinds in order to generalize successfully (Mitchell, 1997).
This paper attempts to close the gap between human and machine concept learning by
developing a rigorous theory for concept learning from limited positive evidence and
testing it against real behavioral data. I focus on a simple abstract task of interest to
both cognitive science and machine learning: learning axis-parallel rectangles in ?R m . We
assume that each object x in our world can be described by its values (XI, ... , xm) on m
real-valued observable dimensions, and that each concept C to be learned corresponds to a
conjunction of independent intervals (mini (C) ~ Xi ~ maXi (C? along each dimension
60
1. B. Tenenbaum
(b)
(a)
(e)
\ - ...... - . ~
r-------------.
I
-
I
I
I
I
+
:C +
I
~
" ...... ~ .
"
- >"
~
?
!
I
+
+
I
t.. ... ____ ... __ ...... ___,
Figure 1: (a) A rectangle concept C. (b-c) The size principle in Bayesian concept learning:
of the man y hypotheses consistent wi th the observed posi ti ve examples, the smallest rapidly
become more likely (indicated by darker lines) as more examples are observed.
i. For example, the objects might be people, the dimensions might be "cholesterol level"
and "insulin level", and the concept might be "healthy levels". Suppose that "healthy
levels" applies to any individual whose cholesterol and insulin levels are each greater than
some minimum healthy level and less than some maximum healthy level. Then the concept
"healthy levels" corresponds to a rectangle in the two-dimensional cholesterol/insulin space.
The problem of generalization in this setting is to infer, given a set of positive (+) and
negative (-) examples of a concept C, which other points belong inside the rectangle
corresponding to C (Fig. 1a.). This paper considers the question most relevant for cognitive
modeling: how to generalize from just a few positive examples?
In machine learning, the problem of learning rectangles is a common textbook example
used to illustrate models of concept learning (Mitchell, 1997). It is also the focus of stateof-the-art theoretical work and applications (Dietterich et aI., 1997). The rectangle learning
task is not well known in cognitive psychology, but many studies have investigated human
learning in similar tasks using simple concepts defined over two perceptually separable
dimensions such as size and color (Shepard, 1987). Such impoverished tasks are worth
our attention because they isolate the essential inductive challenge of concept learning in a
form that is analytically tractable and amenable to empirical study in human subjects.
This paper consists of two main contributions. I first present a new theoretical analysis
of the rectangle learning problem based on Bayesian inference and contrast this model's
predictions with standard learning frameworks (Section 2). I then describe an experiment
with human subjects on the rectangle task and show that, of the models considered, the
Bayesian approach provides by far the best description of how people actually generalize
on this task when given only limited positive evidence (Section 3). These results suggest
an explanation for some aspects of the ubiquotous human ability to learn concepts from just
a few positive examples.
2
Theoretical analysis
Computational approaches to concept learning. Depending on how they model a concept, different approaches to concept learning differ in their ability to generalize meaningfully from only limited positive evidence. Discriminative approaches embody no explicit
model of a concept, but only a procedure for discriminating category members from members of mutually exclusive contrast categories. Most backprop-style neural networks and
exemplar-based techniques (e.g. K -nearest neighbor classification) fall into this group,
along with hybrid models like ALCOVE (Kruschke, 1992). These approaches are ruled out
by definition; they cannot learn to discriminate positive and negative instances ifthey have
seen only positive examples. Distributional approaches model a concept as a probability
distribution over some feature space and classify new instances x as members of C if their
Bayesian Modeling ofHuman Concept Learning
61
estimated probability p(xIG) exceeds a threshold (J. This group includes "novelty detection" techniques based on Bayesian nets (Jaakkola et al., 1996) and, loosely, autoencoder
networks (Japkowicz et al., 1995). While p(xIG) can be estimated from only positive examples, novelty detection also requires negative examples for principled generalization, in
order to set an appropriate threshold (J which may vary over many orders of magnitude for
different concepts. For learning from positive evidence only, our best hope are algorithms
that treat a new concept G as an unknown subset of the universe of objects and decide how
to generalize G by finding "good" subsets in a hypothesis space H of possible concepts.
The Bayesian framework. For this task, the natural hypothesis space H corresponds to all
rectangles in the plane. The central challenge in generalizing using the subset approach is
that any small set of examples will typically be consistent with many hypotheses (Fig. Ib).
This problem is not unique to learning rectangles, but is a universal dilemna when trying to
generalize concepts from only limited positive data. The Bayesian solution is to embed the
hypothesis space in a probabilistic model of our observations, which allows us to weight
different consistent hypotheses as more or less likely to be the true concept based on the
particular examples observed. Specifically, we assume that the examples are generated by
random sampling from the true concept. This leads to the size principle: smaller hypotheses
become more likely than larger hypotheses (Fig. Ib - darker rectangles are more likely),
and they become exponentially more likely as the number of consistent examples increases
(Fig. lc). The size principle is the key to understanding how we can learn concepts from
only a few positive examples.
=
Formal treatment. We observe n positive examples X
{xCI), ... , x Cn )} of concept G
and want to compute the generalization/unction p(y E GIX), i.e. the probability that some
new object y belongs to G given the observations X. Let each rectangle hypothesis h be
denoted by a quadruple (11,/ 2,81,82), where Ii E [-00,00] is the location of h's lower-left
comer and 8i E [0,00] is the size of h along dimension i.
Our probabilistic model consists of a prior density p( h) and a likelihood function p( X Ih)
for each hypothesis h E H. The likelihood is determined by our assumption of randomly
sampled positive examples. In the simplest case, each example in X is assumed to be
independently sampled from a uniform density over the concept C. For n examples we
then have:
p(Xlh)
(1)
o otherwise,
where Ihl denotes the size of h. For rectangle (11,/2,81,82), Ihl is simply 8182 . Note that
because each hypothesis must distribute one unit mass oflikelihood over its volume for each
example
h p(xlh)dh
1), the probability density for smaller consistent hypotheses is
greater than for larger hypotheses, and exponentially greater as a function of n. Figs. Ib,c
illustrate this size principle for scoring hypotheses (darker rectang!es are more likely).
cJx
=
The appropriate choice of p( h) depends on our background knowledge. If we have no a
priori reason to prefer any rectangle hypothesis over any other, we can choose the scaleand location-invariant uninformative prior, p( h) = P(ll, 12, 81 ,82) = 1/(81,82), In any
realistic application, however, we will have some prior information. For example, we may
know the expected size O'i of rectangle concepts along dimension i in our domain, and then
use the associated maximum entropy prior P(ll, 12, 81,82) = exp{ -( 81/0'1 + 82/ 0'2)}.
The generalization function p(y E GIX) is computed by integrating the predictions of all
hypotheses, weighted by their posterior probabilities p( h IX):
p(y E GIX) =
r
p(y E Glh) p(hIX) dh,
(2)
lhEH
where from Bayes' theorem p(hIX) ex: p(Xlh)p(h) (normalized such that
fhEH p(hIX)dh = 1), and p(y E Clh) = 1 if y E hand 0 otherwise. Under the
J. B. Tenenbaum
62
uninformative prior, this becomes:
(3)
Here ri is the maximum distance between the examples in X along dimension i, and
di equals 0 if y falls inside the range of values spanned by X along dimension i, and
otherwise equals the distance from y to the nearest example in X along dimension i.
Under the expected-size prior, p(y E GIX) has no closed form solution valid for all n.
However, except for very small values of n (e.g. < 3) and ri (e.g. < 0'i/1O), the following
approximation holds to within 10% (and usually much less) error:
(4)
Fig. 2 (left column) illustrates the Bayesian learner's contours of equal probability of
generalization (at p = 0.1 intervals), for different values of nand ri. The bold curve
0.5, a natural boundary for generalizing the concept.
corresponds to p(y E GIX)
Integrating over all hypotheses weighted by their size-based probabilities yields a broad
gradient of generalization for small n (row 1) that rapidly sharpens up to the smallest
consistent hypothesis as n increases (rows 2-3), and that extends further along the dimension
with a broader range ri of observations. This figure reflects an expected-size prior with
0'1 = 0'2 = axiLwidthl2; using an uninformative prior produces a qualitatively similar plot.
=
Related work: MIN and Weak Bayes. Two existing subset approaches to concept learning
can be seen as variants of this Bayesian framework. The classic MIN algorithm generalizes
no further than the smallest hypothesis in H that includes all the positive examples (Bruner
et al., 1956; Feldman, 1997). MIN is a PAC learning algorithm for the rectangles task, and
also corresponds to the maximum likelihood estimate in the Bayesian framework (Mitchell,
1997). However, while it converges to the true concept as n becomes large (Fig. 2, row 3),
it appears extremely conservative in generalizing from very limited data (Fig. 2, row 1).
An earlier approach to Bayesian concept learning, developed independently in cognitive
psychology (Shepard, 1987) and machine learning (Haussler et al., 1994; Mitchell, 1997),
was an important inspiration for the framework of this paper. I call the earlier approach
weak Bayes, because it embodies a different generative model that leads to a much weaker
likelihood function than Eq. 1. While Eq. 1 came from assuming examples sampled
randomly from the true concept, weak Bayes assumes the examples are generated by an
arbitrary process independent of the true concept. As a result, the size principle for scoring
hypotheses does not apply; all hypotheses consistent with the examples receive a likelihood
of 1, instead of the factor of 1/lhln in Eq. 1. The extent of generalization is then determined
solely by the prior; for example, under the expected-size prior,
(5)
Weak Bayes, unlike MIN, generalizes reasonably from just a few examples (Fig. 2, row 1).
However, because Eq. 5 is independent of n or ri, weak Bayes does not converge to the
true concept as the number of examples increases (Fig. 2, rows 2-3), nor does it generalize
further along axes of greater variability. While weak Bayes is a natural model when the
examples really are generated independently of the concept (e.g. when the learner himself
or a random process chooses objects to be labeled "positive" or "negative" by a teacher), it
is clearly limited as a model oflearning from deliberately provided positive examples.
In sum, previous subset approaches each appear to capture a different aspect of how humans
generalize concepts from positive examples. The broad similarity gradients that emerge
Bayesian Modeling ofHuman Concept Learning
63
from weak Bayes seem most applicable when only a few broadly spaced examples have
been observed (Fig. 2, row 1), while the sharp boundaries of the MIN rule appear more
reasonable as the number of examples increases or their range narrows (Fig. 2, rows 2-3).
In contrast, the Bayesian framework guided by the size principle automatically interpolates
between these two regimes of similarity-based and rule-based generalization, offering the
best hope for a complete model of human concept learning.
3 Experimental data from human subjects
This section presents empirical evidence that our Bayesian model - but neither MIN nor
weak Bayes - can explain human behavior on the simple rectangle learning task. Subjects
were given the task of guessing 2-dimensional rectangular concepts from positive examples
only, under the cover story of learning about the range of healthy levels of insulin and
cholesterol, as described in Section 1. On each trial of the experiment, several dots
appeared on a blank computer screen. Subjects were told that these dots were randomly
chosen examples from some arbitrary rectangle of "healthy levels," and their job was to
guess that rectangle as nearly as possible by clicking on-screen with the mouse. The dots
were in fact randomly generated on each trial, subject to the constraints ofthree independent
variables that were systematically varied across trials in a (6 x 6 x 6) factorial design. The
three independent variables were the horizontal range spanned by the dots (.25, .5, 1, 2, 4,
8 units in a 24-unit-wide window), vertical range spanned by the dots (same), and number
of dots (2,3,4,6, 10,50). Subjects thus completed 216 trials in random order. To ensure
that subjects understood the task, they first completed 24 practice trials in which they were
shown, after entering their guess, the "true" rectangle that the dots were drawn from. I
The data from 6 subjects is shown in Fig. 3a, averaged across subjects and across the two
directions (horizontal and vertical). The extent d of subjects' rectangles beyond r, the
range spanned by the observed examples, is plotted as a function of rand n, the number
of examples. Two patterns of generalization are apparent. First, d increases monotonically
with r and decreases with n. Second, the rate of increase of d as a function of r is much
slower for larger values of n.
Fig. 3b shows that neither MIN nor weak Bayes can explain these patterns. MIN always
predicts zero generalization beyond the examples - a horizontal line at d = 0 - for all values
of rand n. The predictions of weak Bayes are also independent of rand n: d
0" log 2,
assuming subjects give the tightest rectangle enclosing all points y with p(y E G\X) > 0.5.
=
Under the same assumption, Figs. 3c,d show our Bayesian model's predicted bounds on
generalization using uninformative and expected-size priors, respectively. Both versions of
the model capture the qualitative dependence of d on rand n, confirming the importance of
the size principle in guiding generalization independent of the choice of prior. However, the
uninformative prior misses the nonlinear dependence on r for small n, because it assumes
an ideal scale invariance that clearly does not hold in this experiment (due to the fixed size
of the computer window in which the rectangles appeared). In contrast, the expected-size
prior naturally embodies prior knowledge about typical scale in its one free parameter 0". A
reasonable value of 0" =5 units (out of the 24-unit-wide window) yields an excellent fit to
subjects' average generalization behavior on this task.
4
Conclusions
In developing a model of concept learning that is at once computationally principled and
able to fit human behavior precisely, I hope to have shed some light on how people are able
I Because dots were drawn randomly, the "true" rectangles that subjects saw during practice were
quite variable and were rarely the "correct" response according to any theory considered here. Thus
it is unlikely that this short practice was responsible for any consistent trends in subjects' behavior.
64
1. B. Tenenbaum
to infer the correct extent of a concept from only a few positive examples. The Bayesian
model has two key components: (1) a generalization function that results from integrating
the predictions of all hypotheses weighted by their posterior probability; (2) the assumption
that examples are sampled from the concept to be learned, and not independently of the
concept as previous weak Bayes models have assumed. Integrating predictions over the
whole hypothesis space explains why either broad gradients of generalization (Fig. 2, row
1) or sharp, rule-based generalization (Fig. 2, row 3) may emerge, depending on how
peaked the posterior is. Assuming examples drawn randomly from the concept explains
why learners do not weight all consistent hypotheses equally, but instead weight more
specific hypotheses higher than more general ones by a factor that increases exponentially
with the number of examples observed (the size principle).
This work is being extended in a number of directions. Negative instances, when encountered, are easily accomodated by assigning zero likelihood to any hypotheses containing
them. The Bayesian formulation applies not only to learning rectangles, but to learning
concepts in any measurable hypothesis space - wherever the size principle for scoring
hypotheses may be applied. In Tenenbaum (1999), I show that the same principles enable
learning number concepts and words for kinds of objects from only a few positive examples. 2 I also show how the size principle supports much more powerful inferences than
this short paper could demonstrate: automatically detecting incorrectly labeled examples,
selecting relevant features, and determining the complexity of the hypothesis space. Such
inferences are likely to be necessary for learning in the complex natural settings we are
ultimately interested in.
Acknowledgments
Thanks to M. Bernstein, W. Freeman, S. Ghaznavi, W. Richards, R Shepard, and Y. Weiss for helpful
discussions. The author was a Howard Hughes Medical Institute Predoctoral Fellow.
References
Bruner, J. A., Goodnow,J. S., & Austin, G. J. (1956). A study of thinking. New York: Wiley.
Dietterich, T, Lathrop, R, & Lozano-Perez, T (1997). Solving the multiple-instance problem with
axis-parallel rectangles. ArtificiaL Intelligence 89(1-2), 31-71.
Feldman, J. (1997). The structure of perceptual categories. J. Math. Psych. 41, 145-170.
Haussler, D., Keams, M., & Schapire, R (1994). Bounds on the sample complexity of Bayesian
learning using infonnation theory and the VC-dimension. Machine Learning 14, 83-113.
Jaakkola, T., Saul, L., & Jordan, M. (1996) Fast learning by bounding likelihoods in sigmoid type
belief networks. Advances in NeuraL Information Processing Systems 8.
Japkowicz, N., Myers, C., & Gluck, M. (1995). A novelty detection approach to classification.
Proceedings of the 14th InternationaL Joint Conference on AritificaL InteLLigence.
Kruschke, J. (1992). ALCOVE: An exemplar-based connectionist model of category learning. Psych.
Rev. 99,22-44.
Mitchell, T (1997). Machine Learning. McGraw-Hill.
Muggleton, S. (preprint). Learning from positive data. Submitted to Machine Learning.
Shepard, R (1987). Towards a universal law of generalization for psychological science. Science
237,1317-1323.
Thnenbaum, J. B. (1999). A Bayesian Frameworkfor Concept Learning. Ph. D. Thesis, MIT
Department of Brain and Cognitive Sciences.
2In the framework of inductive logic programming, Muggleton (preprint) has independently
proposed that similar principles may allow linguistic grammars to be learned from positive data only.
Bayesian Modeling ofHuman Concept Learning
65
MIN
Bayes
weak Bayes
n=6
n= 12
Figure 2: Performance of three concept learning algorithms on the rectangle task.
(b) MIN and weak Bayes models
(a) Average data from 6 subjects
52.5
ie
2.5
2
2
~ 1.5
1.5
'0 1
C
~ 0.5
0.5
weak Bayes (0 :: 2)
"In
&
??
~
0
weak Bayes (0:: 1)
~--~~--~----~------
o
2
4
6
0
8
n examples
(c) Bayesian model (uninformative prior)
0
"In
MIN
"In
2
4
6
8
r: Range spanned by
2.5
(d) Bayesian model (expected-size prior)
2.5
2
2
n::2
1.5
1.5
n::3
n=4
n=6
n", 10
n= 50
o
2
4
6
8
2
4
6
8
Figure 3: Data from human subjects and model predictions for the rectangle task.
PART
II
NEUROSCIENCE
| 1542 |@word trial:5 version:1 sharpens:1 selecting:1 offering:1 existing:2 glh:1 blank:1 unction:1 assigning:1 must:1 realistic:2 confirming:1 dilemna:1 plot:1 generative:1 intelligence:2 guess:2 plane:1 core:1 short:2 cjx:1 provides:2 detecting:1 math:1 location:2 along:9 become:3 qualitative:1 consists:2 behavioral:1 inside:2 expected:7 embody:1 behavior:5 nor:3 brain:2 freeman:1 automatically:2 window:3 becomes:2 provided:1 mass:1 kind:2 psych:2 textbook:1 developed:1 finding:1 fellow:1 multidimensional:1 ti:1 shed:1 unit:5 medical:1 appear:2 positive:26 understood:1 xig:2 treat:1 quadruple:1 solely:1 might:3 limited:6 range:8 averaged:1 unique:1 responsible:1 acknowledgment:1 testing:1 practice:3 hughes:1 procedure:1 universal:2 empirical:3 word:2 integrating:4 suggest:1 cannot:2 close:1 measurable:1 xci:1 attention:1 kruschke:2 independently:5 rectangular:1 insight:1 haussler:2 rule:3 cholesterol:4 spanned:5 classic:1 suppose:1 programming:1 hypothesis:29 trend:1 richards:1 distributional:1 labeled:2 predicts:1 observed:6 preprint:2 capture:2 decrease:1 principled:3 complexity:2 ultimately:1 solving:1 learner:4 comer:1 easily:1 joint:1 routinely:1 fast:1 describe:1 artificial:1 whose:1 apparent:1 larger:3 valued:1 quite:1 otherwise:3 grammar:1 ability:3 insulin:4 myers:1 net:1 propose:1 aligned:1 relevant:2 rapidly:2 description:1 produce:1 converges:1 object:6 illustrate:2 depending:2 exemplar:2 nearest:2 job:1 eq:4 predicted:1 differ:1 direction:2 guided:1 correct:2 vc:1 human:18 enable:1 backprop:1 require:1 explains:2 generalization:19 really:1 hold:2 considered:2 exp:1 cognition:1 vary:1 smallest:3 applicable:1 infonnation:1 healthy:7 saw:1 successfully:1 alcove:2 weighted:3 reflects:1 hope:3 mit:2 clearly:2 always:1 jaakkola:2 broader:1 conjunction:1 linguistic:1 ax:1 focus:2 refining:1 likelihood:7 contrast:5 rigorous:1 helpful:1 inference:3 typically:1 unlikely:1 nand:1 keams:1 japkowicz:2 interested:1 classification:2 bruner:2 accomodated:1 stateof:1 denoted:1 priori:1 art:1 equal:3 once:1 sampling:1 broad:3 nearly:1 thinking:1 peaked:1 connectionist:1 few:8 randomly:6 ve:2 individual:1 attempt:1 detection:3 interest:1 light:1 perez:1 amenable:1 capable:1 necessary:2 experience:1 loosely:1 ruled:1 plotted:1 theoretical:4 psychological:1 instance:5 classify:1 modeling:5 column:1 earlier:2 cover:1 oflearning:2 subset:5 uniform:1 successful:1 teacher:1 chooses:1 thanks:1 density:3 international:1 gix:5 discriminating:1 ie:1 probabilistic:2 told:1 mouse:1 posi:1 thesis:1 central:1 containing:1 choose:1 xlh:3 cognitive:7 style:1 doubt:1 distribute:1 bold:1 includes:2 goodnow:1 depends:1 view:1 closed:1 bayes:17 parallel:2 contribution:1 yield:2 spaced:1 clh:1 generalize:10 weak:15 bayesian:24 produced:1 worth:1 submitted:1 explain:3 definition:1 against:1 naturally:1 associated:1 di:1 sampled:4 treatment:1 massachusetts:1 ihl:2 mitchell:5 color:1 knowledge:2 impoverished:1 actually:1 appears:1 higher:1 response:1 wei:1 rand:4 formulation:1 just:3 hand:1 horizontal:3 nonlinear:1 indicated:1 dietterich:2 concept:62 true:8 normalized:1 deliberately:1 inductive:2 analytically:1 inspiration:1 lozano:1 entering:1 jbt:1 ll:2 during:1 trying:1 hill:1 complete:1 demonstrate:1 meaning:1 common:1 sigmoid:1 shepard:4 exponentially:3 volume:1 belong:1 cambridge:1 feldman:3 ai:1 dot:8 similarity:2 posterior:3 perspective:1 belongs:1 came:1 joshua:1 scoring:3 seen:2 minimum:1 greater:4 novelty:3 converge:1 monotonically:1 ii:2 multiple:1 infer:2 exceeds:1 frameworkfor:1 muggleton:2 equally:1 prediction:6 variant:1 himself:1 receive:1 background:1 want:1 uninformative:6 interval:2 unlike:1 subject:19 isolate:1 meaningfully:1 member:3 seem:1 jordan:1 call:1 ideal:1 bernstein:1 fit:3 psychology:2 cn:1 bridging:1 interpolates:1 york:1 useful:1 factorial:1 tenenbaum:5 ph:1 category:5 simplest:1 schapire:1 ofthree:1 estimated:2 neuroscience:1 broadly:1 group:2 key:2 threshold:2 drawn:3 neither:2 rectangle:29 sum:1 powerful:1 extends:1 reasonable:3 decide:1 prefer:1 bound:2 encountered:1 constraint:1 precisely:1 ri:5 aspect:2 min:11 extremely:1 separable:1 department:2 developing:2 according:1 smaller:2 across:3 psyche:1 wi:1 rev:1 wherever:1 invariant:1 computationally:1 mutually:1 know:1 tractable:1 generalizes:2 tightest:1 apply:1 observe:1 appropriate:2 slower:1 denotes:1 assumes:2 ensure:1 completed:2 embodies:2 question:1 exclusive:1 dependence:2 guessing:1 gradient:3 distance:2 capacity:1 considers:1 extent:3 reason:1 assuming:3 mini:1 negative:7 design:1 enclosing:1 unknown:1 perform:1 predoctoral:1 vertical:2 observation:3 howard:1 incorrectly:1 extended:1 variability:1 precise:1 varied:1 arbitrary:2 sharp:2 learned:4 narrow:1 beyond:2 able:2 usually:1 pattern:2 xm:1 regime:1 appeared:2 challenge:2 explanation:1 belief:1 natural:5 hybrid:1 technology:1 axis:3 hix:3 autoencoder:1 prior:17 understanding:1 determining:1 law:1 remarkable:1 consistent:9 principle:12 story:1 systematically:1 row:10 austin:1 free:1 formal:1 weaker:1 allow:1 institute:2 neighbor:1 fall:2 wide:2 saul:1 emerge:2 boundary:3 dimension:11 curve:1 world:1 valid:1 contour:1 author:1 qualitatively:1 far:1 observable:1 mcgraw:1 feat:1 logic:1 assumed:2 xi:2 discriminative:1 why:2 learn:4 reasonably:1 investigated:1 complex:2 excellent:1 domain:1 main:1 universe:1 whole:1 bounding:1 fig:17 screen:2 darker:3 wiley:1 lc:1 guiding:1 explicit:1 clicking:1 perceptual:2 ib:3 ix:1 theorem:1 embed:1 specific:1 pac:1 maxi:1 evidence:5 essential:1 ih:1 importance:1 magnitude:1 perceptually:1 illustrates:1 gap:1 aritifical:1 gluck:1 entropy:1 generalizing:3 simply:1 likely:7 applies:2 corresponds:5 dh:3 ma:1 towards:1 man:1 specifically:1 determined:2 except:1 typical:1 miss:1 conservative:1 discriminate:1 invariance:1 e:1 experimental:1 lathrop:1 rarely:2 people:3 support:1 ex:1 |
594 | 1,543 | Learning Mixture Hierarchies
Nuno Vasconcelos
Andrew Lippman
MIT Media Laboratory, 20 Ames St, EI5-320M, Cambridge, MA 02139,
{nuno,lip} @media.mit.edu,
http://www.media.mit.edwnuno
Abstract
The hierarchical representation of data has various applications in domains such as data mining, machine vision, or information retrieval. In
this paper we introduce an extension of the Expectation-Maximization
(EM) algorithm that learns mixture hierarchies in a computationally efficient manner. Efficiency is achieved by progressing in a bottom-up
fashion, i.e. by clustering the mixture components of a given level in the
hierarchy to obtain those of the level above. This cl ustering requires onl y
knowledge of the mixture parameters, there being no need to resort to
intermediate samples. In addition to practical applications, the algorithm
allows a new interpretation of EM that makes clear the relationship with
non-parametric kernel-based estimation methods, provides explicit control over the trade-off between the bias and variance of EM estimates, and
offers new insights about the behavior of deterministic annealing methods
commonly used with EM to escape local minima of the likelihood.
1 Introduction
There are many practical applications of statistical learning where it is useful to characterize
data hierarchically. Such characterization can be done according to either top-down or
bottom-up strategies. While the former start by generating a coarse model that roughly
describes the entire space, and then successively refine the description by partitioning the
space and generating sub-models for each of the regions in the partition; the later start
from a fine description, and successively agglomerate sub-models to generate the coarser
descriptions at the higher levels in the hierarchy.
Bottom-up strategies are particularly useful when not all the data is available at once, or
when the dataset is so big that processing it as whole is computationally infeasible. This
is the case of machine vision tasks such as object recognition, or the indexing of video
databases. In object recognition, it is many times convenient to determine not only which
object is present in the scene but also its pose [2], a goal that can be attained by a hierarchical,
description where at the lowest level a model is learned for each object pose and all pose
models are then combined into a generic model at the top level of the hierarchy. Similarly,
Learning Mixture Hierarchies
607
for video indexing, one may be interested in learning a description for each frame and
then combine these into shot descriptions or descriptions for some other sort of high level
temporal unit [6).
In this paper we present an extension of the EM algorithm [I) for the estimation of hierarchical mixture models in a bottom-up fashion. It turns out that the attainment of this goal
has far more reaching consequences than the practical applications above. In particular,
because a kernel density estimate can be seen as a limiting case ofa mixture model (where
a mixture component is superimposed on each sample), this extension establishes a direct
connection between so-called parametric and non-parametric density estimation methods
making it possible to exploit results from the vast non-parametric smoothing literature [4)
to improve the accuracy of parametric estimates. Furthennore, the original EM algorithm
becomes a particular case of the one now presented, and a new intuitive interpretation becomes available for an important variation of EM (known as deterministic annealing) that
had previously been derived from statistical physics. With regards to practical applications,
the algorithm leads to computationally efficient methods for estimating density hierarchies
capable of describing data at different resolutions.
2 Hierarchical mixture density estimation
Our model consists of a hierarchy of mixture densities, where the data at a given level is
described by
cl
P(X) =
L 1I"~p(Xlz~ =
I , Md,
(I)
k= 1
where 1 is the level in the hierarchy (l = 0 providing the coarsest characterization of the
data), MI the mixture model at this level, C l the number of mixture components that
compose it, 11"~ the prior probability of the kth component, and z~ a binary variable that
takes the value 1 if and only if the sample X was drawn from this component. The only
restriction on the model is that if node j of levell + 1 is a child of node i of levell, then
11"1+1
J
=
11"1+111"1
jlk k'
(2)
where k is the parent of j in the hierarchy of hidden variables.
The basic problem is to compute the mixture parameters of the description at levell given
the knowledge of the parameters at level 1 + 1. This can also be seen as a problem of
clustering mixture components. A straightforward solution would be to draw a sample
from the mixture density at levell + 1 and simply run EM with the number of classes of
the level 1 to estimate the corresponding parameters. Such a solution would have at least
two major limitations. First, there would be no guarantee that the constraint of equation (2)
would be enforced, i.e. there would be no guarantee of structure in the resulting mixture
hierarchy, and second it would be computationally expensive, as all the models in the
hierarchy would have to be learned from a large sample. In the next section, we show that
this is really not necessary.
3 Estimating mixture hierarchies
The basic idea behind our approach is, instead of generating a real sample from the mixture
model at level L + 1, to consider a virtual sample generated from the same model, use EM
to find the expressions for the parameters of the mixture model of levell that best explain
this virtual sample, and establish a closed-fonn relationship between these parameters and
those of the model at level 1 + I. For this, we start by considering a virtual sample
X = {XI, .. . , X C l+ l } from ; \.11+1, where each of the Xi is a virtual sample from one of
N. Vasconcelos and A. Lippman
608
the C ' + 1 components of this model, with size Mi =
virtual points.
11'! N,
where N is the total number of
We next establish the likelihood for the virtual sample under the model M" For this, as is
usual in the EM literature, we assume that samples from different blocks are independent,
Le.
C 1+ 1
P(XIM,) =
II P(XiIM,),
(3)
i=1
but, to ensure that the constraint of equation (2) is enforced, samples within the same block
are assigned to the same component of M,. Assuming further that, given the knowledge
of the assignment the samples are drawn independently from the corresponding mixture
component, the likelihood of each block is given by
P(XiIMd
c1
c1
j = 1
j= 1
= 2: lI'~P(Xilzij = I,M,) = 2: 11'}
Mi
II p(XrlZij = I,M,),
(4)
m= 1
where Zij = Z!+I z; is a binary variable with value one if and only if the block Xi is assigned
is the mth data point in Xi. Combining equations (3)
to the jth component of M" and
and (4) we obtain the incomplete data likelihood, under M" for the whole sample
xr
C 1+ 1
P(XIM,)
=
c1
M.
II 2: 11'; II p(XrlZij = I,M,).
i= 1 j = 1
(5)
m= 1
This equation is similar to the incomplete data likelihood of standard EM, the main difference being that instead of having an hidden variable for each sample point, we now have
one for each sample block. The likelihood of the complete data is given by
C 1+ 1
P(X, ZIM,) =
c1
II II [lI'~P(Xilzij = 1, M,)f?i ,
(6)
i= 1 j = 1
where Z is a vector containing all the Zij, and the log-likelihood becomes
C 1+ 1
log P(X, ZIM,)
c1
= 2: 2: Zij 10g(1I';P(Xilzij = 1, M,).
(7)
i= 1 j = 1
Relying on EM to estimate the parameters of M, leads to the the following E-step
_
_
_
P(Xi IZij = I, M,)lI';
E[zijIXi,M,]-P(zij-lIXi , M,)-~ P(X .I. - I lA) I ' (8)
L..k
, Zzk - , l VI/ lI'k
The key quantity to compute is therefore P (Xi IZij = I, M,). Taking its logarithm
h ij
10gP(Xilzij
= I,M,)
1 M.
Mi[M . 2:logP(xrlzij
= I,M,)]
, i= 1
MiE M 1+ [log P(XIZij = 1, M,)],
1 .,
(9)
where we have used the law of large numbers, and EM 1+ 1? ? [x] is the expected value of x
according the ith mixture component of M'+ 1 (the one from which Xi was drawn). This
is an easy computation for most densities commonly used in mixture modeling. It can be
shown [5] that for the Gaussian case it leads to
.L
{ ~l _1~ l +I}]M.
I 1
[ g(J1.i+
,/L~ , E~)e-~trace (~J)~'
7r~
(10)
609
Learning Mixture Hierarchies
where 9(x, J1., E) is the expression for a Gaussian with mean J1. and covariance E.
The M-step consists of maximizing
C I +1
Q=
cl
L L hij 10g(1T~P(Xilzij =
1, Md)
(II)
i= 1 j = 1
subject to the constraint E j 7r~ = I. Once again, this is a relatively simple task for
common mixture models and in [5] we show that for the Gaussian case it leads to the
following parameter update equations
(12)
Ei hij MiJ1.~+ 1
Ei h ij Mi
(3)
E.~ .. M . [LhijMiE~+1 + LhijMi(J1.~+I-J1.;)(J1.!+1 -J1.;f] .(4)
,
'J
,
i
i
Notice that neither equation (10) nor equations (12) to (14) depend explicitly on the underlying sample Xi and can be computed directly from the parameters of Ml+l. The
algorithm is thus very efficient from a computational standpoint as the number of mixture
components in Ml+ 1 is typically much smaller than the size of the sample at the bottom of
the hierarchy.
4
Relationships with standard EM
There are interesting relationships between the algorithm derived above and the standard
EM procedure. The first thing to notice is that by making Mi = I and E~+ 1 = 0, the E and
M-steps become those obtained by applying standard EM to the sample composed of the
points J1.~+1
.
Thus, standard EM can be seen as a particular case of the new algorithm, that learns a two
level mixture hierarchy. An initial estimate is first obtained at the bottom of this hierarchy
by placing a Gaussian with zero covariance on top of each data point, the model at the
second level being then computed from this estimate. The fact that the estimate at the
bottom level is nothing more than a kernel estimate with zero bandwidth suggests that other
choices of the kernel bandwidth may lead to better overall EM estimates.
Under this interpretation, the E~+I become free parameters that can be used to control the
smoothness of the density estimates and the whole procedure is equivalent to the composition
of three steps: I) find the kernel density estimate that best fits the sample under analysis, 2)
draw a larger virtual sample from that density, and 3) compute EM estimates from this larger
sample. In section 5, we show that this can leave to significant improvements in estimation
accuracy, particularly when the initial sample is small, the free parameters allowing explicit
control over the trade-off between the bias and variance of the estimator.
Another interesting relationship between the hierarchical method and standard EM can
be derived by investigating the role of the size of the underlying virtual sample (which
determines M i ) on the estimates. Assuming Mi constant, Mi = M, Vi, it factors out of
all summations in equations (12) to (14), the contributions of numerator and denominator
canceling each other. In this case, the only significance of the choice of M is its impact on
the E-step. Assuming, as before, that E~+I = 0 we once again have the EM algorithm, but
where the class-conditional likelihoods of the E-step are now raised to the Mth power. If
610
N Vasconcelos and A. Lippman
M is seen as the inverse of temperature, both the E and M steps become those of standard
EM under deterministic annealing (DA) I [3] .
The DA process is therefore naturally derived from our hierarchical formulation, which
gives it a new interpretation that is significantly simpler and more intuitive than those
derived from statistical physics. At the start of the process M is set to zero, i.e. no virtual
samples are drawn from the Gaussian superimposed on the real dataset, and there is no
virtual data. Thus, the assignments hij of the E-step simply become the prior mixing
proportions 11"; and the M-step simply sets the parameters of all Gaussians in the model to
the sample mean and sample covariance of the real sample. As M increases, the number
of virtual points drawn from each Gaussian also increases and for M = 1 we have a single
point that coincides with the point on the real training sample. We therefore obtain the
standard EM equations. Increasing M further will make the E-step assignments harder (in
the limit of M = 00 each point is assigned to a single mixture component) because a larger
virtual probability mass is attached to each real point leading to much higher certainty with
regards to the reliability of the assignment.
Overall, while in the beginning of the process the reduced size of the virtual sample allows
the points in the real sample to switch from mixture to mixture easily, as M is increased
the switching becomes much less likely. The "exploratory" nature of the initial iterations
drives the process towards solutions that are globally good, therefore allowing it to escape
local minima.
5 Experimental results
In this section, we present experimental results that illustrate the properties of the hierarchical EM algorithm now proposed. We start by a simple example that illustrates how the
algorithm can be used to estimate hierarchical mixtures.
..
. ~.,...:.:~ ..:
, :...
:~~y :.~.::
~,~~,;,i
1!IO
- 100
-so
0
50
100
'?.i~ i ' ;~
'~'Tl;~:.
150
Figure I: Mixture hierarchy derived from the model shown in the left. The plot relative to each
level of the hierarchy is superimposed on a sample drawn from this model. Only the one-standard
deviation contours are shown for each Gaussian.
The plot on the left of Figure 1 presents a Gaussian mixture with 16 uniformly weighted
components. A sample with 1000 points was drawn from this model, and the algorithm
used to find the best descriptions for it at three resolutions (mixtures with 16, 4, and 2
Gaussian). These descriptions are shown in the figure. Notice how the mixture hierarchy
naturally captures the various levels of structure exhibited by the data.
This example suggests how the algorithm could be useful for applications such as object
recognition or image retrieval. Suppose that each of the Gaussians in the leftmost plot of
IDA is a technique drawn from analogies with statistical physics that avoids local maxima of
the likelihood function (in which standard EM can get trapped) by perfonning a succession of
optimizations at various temperatures [31.
611
Learning Mixture Hierarchies
(
I'?
j
'~~~~--H~~W~~~~
o-.I? ...,r'Itol
Figure 2: Object recognition task. Left: 8 of the 100 objects in the database. Right: computational
savings achieved with hierarchical recognition vs full search.
the figure describes how a given pose of a given object populates a 2-D feature space on
which object recognition is to be perfonned. In this case, higher levels in the hierarchical
representation provide a more generic description of the object. E.g. each of the Gaussians
in the model shown in the middle of the figure might provide a description for all the poses
in which the camera is on the same quadrant of the viewing sphere, while those in the
model shown in the right might represent views from the same hemisphere. The advantage,
for recognition or retrieval, of relying on a hierarchal structure is that the search can be
perfonned first at the highest resolution, where it is much less expensive, only the best
matches being considered at the subsequent levels.
Figure 2 illustrates the application of hierarchical mixture modeling to a real object recognition task. Shown on the left side of the figure are 8 objects from the 100 contained in the
Columbia object database [2]. The database consists of 72 views (obtained by positioning
the camera in 5? intervals along a circle on the viewing sphere), which were evenly separated into a training and a test set. A set of features was computed for each image, and a
hierarchical model was then learned for each object in the resulting feature space. While
the process could be extended to any number of levels, here we only report on the case of
a two-level hierarchy: at the bottom each image is described by a mixture of 8 Gaussians,
and at the top each mixture (also with 8 Gaussians) describes 3 consecutive views. Thus,
the entire training set is described by 3600 mixtures at the bottom resolution and 1200 at
the top.
Given an image of an object to recognize, recognition takes place by computing its projection
into the feature space, measuring the likelihood of the resulting sample according to each
of the models in the database, and choosing the most likely. The complexity of the process
is proportional to the database size. The plot on the left of Figure 2 presents the recognition
accuracy achieved with the hierarchical representation vs the corresponding complexity,
shown as a percent of the complexity required by full search. The full-search accuracy is
in this case 90%, and is also shown as a straight line in the graph. As can be seen from the
figure, the hierarchical search achieves the full search accuracy with less than 40% of its
complexity. We are now repeating this experiments with deeper trees, where we expect the
gains to be even more impressive.
We finalize by reporting on the impact of smoothing on the quality of EM estimates.
For this, we conducted the following Monte Carlo experiment: I) draw 200 datasets
Si, i = 1, ... ,200 from the model shown on the left of Figure 1, 2) fit each dataset with
EM, 3) measure the correlation coefficient Pi, i = 1, ... ,200 between each of the EM fits
and the original model, and 4) compute the sample mean pand variance p. The correlation
coefficient is defined by Pi = f f(x)h(x)dxIU f(x)dxf ii(X)dx), where f(x) is the
a
N. Vasconcelos and A. Lippman
612
8lC 10~
OD
-so
,
08
- -HI)
/
-so
- -
- - 100
?
200
?
4CXl
-
- - 300
J(I)
500
- - 1000
<00
-500
-- - 1000
-
o .o'-----'--~,
0--':15:-----:2'
O :---25
~--:'c30--=
36:--.....
40: :---'
45
- - - - - - - - -:: ..-- ----10
15
20
-
2S
~ . " -, -'
30
36
Figure 3: Results of the Monte Carlo experiment described on the text. Left: p as a function 17k.
Right: Up as a function of 17k. The various curves in each graph correspond to to different sample
sizes.
true model and fi(X) the ith estimate, and can be computed in closed form for Gaussian
mixtures. The experiment was repeated with various dataset sizes and various degrees of
smoothing (by setting the bandwidth of the underlying Gaussian kernel to
for various
values of O'k).
oil
Figure 3 presents the results of this experiment. It is clear, from the graph on the left, that
smoothing can have a significant impact on the quality of the EM estimates. This impact
is largest for small samples, where smoothing can provide up to a two fold improvement
estimation accuracy, but can be found even for large samples.
The kernel bandwidth allows control over the trade-off between the bias and variance of
the estimates. When O'k is zero (standard EM), bias is small but variance can be large, as
illustrated by the graph on the right of the figure. As O'k is increased, variance decreases at
the cost of an increase in bias (the reason why for large O'k aU lines in the graph of the left
meet at the same point regardless ofthe sample size). The point where p is the highest is the
point at which the bias-variance trade off is optimal. Operating at this point leads to a much
smaller dependence of the accuracy of the estimates on the sample size or, conversely, the
need for much smaller samples to achieve a given degree of accuracy.
References
[I] A. Dempster, N. Laird, and D. Rubin. Maximum-likelihood from Incomplete Data via
the EM Algorithm. J. of the Royal Statistical Society, B-39, 1977.
[2] H. Murase and S. Nayar. Visual Learning and Recognition of 3-D Objects from
Appearence. International Journal of Computer Vision, 14:5-24, 1995.
[3] K. Rose, E. Gurewitz, and G. Fox. Vector Quantization by Determinisc Annealing.
IEEE Trans. on Information Theory, Vol. 38, July 1992.
[4] J. Simonoff. Smoothing Methods in Statistics. Springer-Verlag, 1996.
[5] N. Vasconcelos and A. Lippman. Learning Mixture Hierarchies. Technical report,
MIT
Media
Laboratory,
1998.
Available
from
ftp:l/ftp.media.mit.eduJpub/nunolHierMix.ps.gz .
[6] N. Vasconcelos and A. Lippman. Content-based Pre-Indexed Video. In Proc. Int. Con!
Image Processing, Santa Barbara, California, 1997.
| 1543 |@word middle:1 proportion:1 covariance:3 xilzij:5 fonn:1 harder:1 shot:1 initial:3 zij:4 ida:1 od:1 si:1 dx:1 subsequent:1 partition:1 j1:8 plot:4 update:1 v:2 beginning:1 ith:2 provides:1 characterization:2 coarse:1 ames:1 node:2 simpler:1 along:1 direct:1 become:4 consists:3 combine:1 compose:1 introduce:1 manner:1 expected:1 roughly:1 behavior:1 nor:1 relying:2 globally:1 considering:1 increasing:1 becomes:4 dxf:1 estimating:2 underlying:3 medium:5 lixi:1 lowest:1 mass:1 guarantee:2 temporal:1 certainty:1 ofa:1 control:4 partitioning:1 unit:1 before:1 local:3 limit:1 consequence:1 switching:1 io:1 meet:1 might:2 au:1 suggests:2 conversely:1 practical:4 camera:2 block:5 xr:1 lippman:6 procedure:2 significantly:1 convenient:1 projection:1 pre:1 quadrant:1 get:1 applying:1 www:1 restriction:1 deterministic:3 xlz:1 equivalent:1 maximizing:1 appearence:1 straightforward:1 regardless:1 independently:1 resolution:4 insight:1 estimator:1 exploratory:1 variation:1 limiting:1 hierarchy:23 suppose:1 recognition:11 particularly:2 expensive:2 coarser:1 database:6 bottom:9 role:1 capture:1 region:1 trade:4 highest:2 decrease:1 rose:1 dempster:1 complexity:4 depend:1 efficiency:1 easily:1 various:7 separated:1 monte:2 choosing:1 larger:3 furthennore:1 statistic:1 gp:1 laird:1 advantage:1 combining:1 mixing:1 achieve:1 description:12 intuitive:2 parent:1 xim:2 p:1 generating:3 leave:1 object:16 ftp:2 illustrate:1 andrew:1 pose:5 ij:2 murase:1 mie:1 viewing:2 virtual:13 really:1 summation:1 extension:3 considered:1 major:1 achieves:1 consecutive:1 estimation:6 proc:1 largest:1 establishes:1 weighted:1 mit:5 gaussian:11 reaching:1 derived:6 zim:2 improvement:2 likelihood:11 superimposed:3 progressing:1 entire:2 typically:1 hidden:2 mth:2 interested:1 overall:2 smoothing:6 raised:1 once:3 saving:1 vasconcelos:6 having:1 placing:1 report:2 escape:2 composed:1 recognize:1 mining:1 mixture:41 behind:1 capable:1 necessary:1 fox:1 tree:1 incomplete:3 indexed:1 logarithm:1 circle:1 increased:2 modeling:2 measuring:1 logp:1 assignment:4 maximization:1 cost:1 deviation:1 conducted:1 characterize:1 combined:1 st:1 density:10 international:1 off:4 physic:3 again:2 successively:2 containing:1 resort:1 leading:1 li:4 coefficient:2 int:1 explicitly:1 vi:2 later:1 view:3 closed:2 start:5 sort:1 contribution:1 accuracy:8 pand:1 variance:7 succession:1 correspond:1 ofthe:1 carlo:2 finalize:1 drive:1 straight:1 explain:1 canceling:1 nuno:2 naturally:2 mi:8 con:1 gain:1 dataset:4 knowledge:3 higher:3 attained:1 formulation:1 done:1 correlation:2 ei:2 quality:2 oil:1 true:1 former:1 assigned:3 laboratory:2 illustrated:1 numerator:1 coincides:1 leftmost:1 agglomerate:1 complete:1 temperature:2 percent:1 image:5 fi:1 common:1 attached:1 interpretation:4 significant:2 composition:1 cambridge:1 smoothness:1 similarly:1 had:1 reliability:1 impressive:1 operating:1 hemisphere:1 barbara:1 hierarchal:1 verlag:1 binary:2 seen:5 minimum:2 determine:1 july:1 ii:8 full:4 levell:5 match:1 positioning:1 technical:1 offer:1 sphere:2 retrieval:3 impact:4 basic:2 denominator:1 vision:3 expectation:1 iteration:1 kernel:7 represent:1 achieved:3 c1:5 addition:1 fine:1 annealing:4 interval:1 standpoint:1 exhibited:1 subject:1 thing:1 intermediate:1 easy:1 switch:1 fit:3 bandwidth:4 idea:1 expression:2 useful:3 clear:2 santa:1 repeating:1 reduced:1 http:1 generate:1 notice:3 trapped:1 vol:1 key:1 drawn:8 neither:1 vast:1 graph:5 enforced:2 run:1 inverse:1 place:1 reporting:1 draw:3 hi:1 fold:1 refine:1 constraint:3 scene:1 coarsest:1 relatively:1 according:3 describes:3 smaller:3 em:31 making:2 indexing:2 jlk:1 computationally:4 equation:9 previously:1 turn:1 describing:1 available:3 gaussians:5 hierarchical:14 generic:2 original:2 top:5 clustering:2 ensure:1 exploit:1 establish:2 society:1 quantity:1 parametric:5 strategy:2 dependence:1 md:2 usual:1 kth:1 evenly:1 reason:1 assuming:3 relationship:5 providing:1 onl:1 hij:3 trace:1 allowing:2 datasets:1 extended:1 frame:1 populates:1 required:1 connection:1 california:1 learned:3 trans:1 royal:1 video:3 power:1 perfonned:2 improve:1 gz:1 columbia:1 gurewitz:1 text:1 prior:2 literature:2 relative:1 law:1 expect:1 interesting:2 limitation:1 proportional:1 analogy:1 c30:1 degree:2 rubin:1 pi:2 free:2 infeasible:1 jth:1 bias:6 side:1 deeper:1 taking:1 regard:2 curve:1 avoids:1 contour:1 simonoff:1 commonly:2 far:1 ml:2 investigating:1 xi:8 search:6 cxl:1 why:1 lip:1 nature:1 attainment:1 cl:3 domain:1 da:2 significance:1 hierarchically:1 main:1 big:1 whole:3 nothing:1 child:1 repeated:1 tl:1 fashion:2 lc:1 sub:2 explicit:2 learns:2 down:1 quantization:1 illustrates:2 simply:3 likely:2 visual:1 contained:1 springer:1 determines:1 ma:1 conditional:1 goal:2 towards:1 content:1 uniformly:1 called:1 total:1 experimental:2 la:1 perfonning:1 izij:2 nayar:1 |
595 | 1,544 | Exploratory Data Analysis Using Radial Basis
Function Latent Variable Models
Alan D. Marrs and Andrew R. Webb
DERA
St Andrews Road, Malvern
Worcestershire U.K. WR14 3PS
{marrs,webb}@signal.dera.gov.uk
@British Crown Copyright 1998
Abstract
Two developments of nonlinear latent variable models based on radial
basis functions are discussed: in the first, the use of priors or constraints
on allowable models is considered as a means of preserving data structure
in low-dimensional representations for visualisation purposes. Also, a
resampling approach is introduced which makes more effective use of
the latent samples in evaluating the likelihood.
1 INTRODUCTION
Radial basis functions (RBF) have been extensively used for problems in discrimination
and regression. Here we consider their application for obtaining low-dimensional representations of high-dimensional data as part of the exploratory data analysis process. There
has been a great deal of research over the years into linear and nonlinear techniques for
dimensionality reduction. The technique most commonly used is principal components
analysis (PCA) and there have been several nonlinear generalisations, each taking a particular definition of PCA and generalising it to the nonlinear situation.
One approach is to find surfaces of closest fit (as a generalisation of the PCA definition
due to the work of Pearson (1901) for finding lines and planes of closest fit). This has
been explored by Hastie and Stuetzle (1989), Tibshirani (1992) (and further by LeBlanc
and Tibshirani, 1994) and various authors using a neural network approach (for example,
Kramer, 1991). Another approach is one of variance maximisation subject to constraints
on the transformation (Hotelling, 1933). This has been investigated by Webb (1996), using
a transformation modelled as an RBF network, and in a supervised context in Webb (1998).
An alternative strategy also using RBFs, based on metric multidimensional scaling, is described by Webb (1995) and Lowe and Tipping (1996). Here, an optimisation criterion,
A. D. Marrs and A. R. Webb
530
termed stress, is defined in the transformed space and the weights in an RBF model determined by minimising the stress.
The above methods use a radial basis function to model a transformation from the highdimensional data space to a low-dimensional representation space. A complementary approach is provided by Bishop et al (1998) in which the structure of the data is modelled as
a function of hidden or latent variables. Termed generative topographic mapping (GTM),
the model may be regarded as a nonlinear generalisation of factor analysis in which the
mapping from latent space to data space is characterised by an RBF.
Such generative models are relevant to a wide range of applications including radar target
modelling, speech recognition and handwritten character recognition.
However, one of the problems with GTM that limits its practical use for visualising data on
manifolds in high dimensional space arises from distortions in the structure that it imposes.
This is acknowledged in Bishop et al (1997) where 'magnification factors' are introduced
to correct for the GTM's deficiency as a means of data visualisation.
This paper considers two developments: constraints on the permissible models and resampiing of the latent space. Section 2 presents the background to latent variable models;
Model constraints are discussed in Section 3. Section 4 describes a re-sampling approach
to estimation of the posterior pdf on the latent samples. An illustration is provided in Section 5.
2 BACKGROUND
Briefly, we shall re-state the basic GTM model, retaining the notation of Bishop et al
(1998). Let Ui, i = 1, ... , N}, ti E RP represent measurements on the data space variables; Z E R. represent the latent variables.
Let t be normally-distributed with mean y(z; W) and covariance matrix {3-1 I; y(z; W)
is a nonlinear transformation that depends on a set of parameters W. Specifically, we shall
assume a basis function model
M
L Wi<Pi(Z)
y(z; W) =
i=1
where the vectors Wi E R.D are to be determined through optimisation and {<Pi, i =
1, . . . , M} is a set of basis functions defined on the latent space.
The data distribution may be written
p(tIW,{3) =
!
(1)
p(tlz; W , {3)p(z)dz
where, under the assumptions of normality,
{3 )
p(tlz; W,{3) = ( 271"
D
/2
exp
{{3
-"2 ly (z; W) -
tll 2
}
Approximating the integral by a finite sum (assuming the functions p(z) and y(z) do not
vary too greatly compared with the sample spacing), we have
K
p(tIW,{3) = LPiP(tlzi; W,{3)
i =1
which may be regarded as a function of the parameters W and {3 that characterise y .
(2)
Exploratory Data Analysis Using Radial Basis Function Latent Variable Models
Given the data set {ti' i
531
= 1, ... ,N}, the log likelihood is given by
N
L(W,J3) = I)n[p(t;IW,J3)]
j=l
which may be maximised using a standard EM approach (Bishop et ai, 1998).
In this case, we have
Pj
1
= N
N
L~n
(3)
n=l
as the re-estimate of the mixture component weights, Pj, at the (m
~
n
_
p~m)p(tnIZi; w(m), J3(m?)
-
EiP~m)p(tnlzi;W(m),J3(m?)
+ 1) step, where
(4)
and (.) (m) denotes values at the mth step. Note that Bishop et al (1998) do not re-estimate
P;; all values are taken to be equal.
The number of P; terms to be re-estimated is K, the number of terms used to approximate
the integral (1). We might expect that the density is smoothly varying and governed by a
much fewer number of parameters (not dependent on K).
The re-estimation equation for the D x M matrix W = [w 11 ... IWM] is
= TT RT +[+TG+]-l
W(m+l)
(5)
where G is the K x K diagonal matrix with
N
Gjj = LRjn
n=l
=
and TT = [tIl .. . ltN], +T
[tfJ(Zl)I .. . ltfJ(ZK)]. The term
l/J3(m) = 1/(ND) E~l E~l Rjilti - w(m+l)tfJ(Zj) 12.
J3
is re-estimated as
Once we have determined parameters of the transformation, we may invert the model by
asking for the distribution of Z given a measurement ti. That is, we require
p(Zlti)
=
p(tilz)p(z)
f p(tilz)p(z)dz
(6)
For example, we may plot the position of the peak of the distribution p( Ziti) for each data
sample ti.
3 APPLYING A CONSTRAINT
One way to retain structure is to impose a condition that ensures that a unit step in the
latent space corresponds to a unit step in the data space (more or less). For a single latent
variable, Xl, we may impose the constraints that
l ay 12 =
1
aXl
which may be written, in terms of W as
ifwTWil
1
532
A. D. Marrs and A. R. Webb
8c/J.
where;l = ~
The derivative of the data space variable with respect to the latent variable has unit magnitude. The derivative is of course a function of Xl and imposing such a condition at each
sample point in latent space would not be possible owing to the smoothness of the RBF
model. However, we may average over the latent space,
where ( .) denotes average over the latent space.
In general, for L latent variables we may impose a constraint JTWTW J = 1 L leading
to the penalty term
Tr {A(JTWTW J - IL)}
where J is an M x L matrix with jth column 8?/8xj and A is a symmetric matrix of
Lagrange multipliers. This is very similar to regularisation terms. It is a condition on
the norm of W; it incorporates the Jacobian matrix J and a symmetric L x L matrix of
Lagrange multipliers, A. The re-estimation solution for W may be written
(7)
with A chosen so that the constraint JT W T W J = 1 L is satisfied.
We may also use the derivatives of the transformation to define a distortion measure or
magnification factor,
M(Zj W) = IIJTWTW J - 1112
which is a function of the latent variables and the model parameters. A value of zero shows
that there is no distortion 1?
An alternative to the constraint approach above is to introduce a prior on the allowable
transformations using the magnification factor; for example,
P(W)
~
exp(-AM(zj W))
(8)
where A is a regularisation parameter. This leads to a modification to the M-step reestimation equation for W, providing a maximum a posteriori estimate. Equation (8) provides a natural generalisation of PCA since for the special case of a linear transformation
(Pi = Xi, M = L), the solution for W is the PCA space as A ~ 00.
4
RESAMPLING THE LATENT SPACE
Having obtained a mapping from latent space to data space using the above constraint, we
seek a better estimate to the posterior pdf of the latent samples. Current versions of GTM
require the latent samples to be uniformly distributed in the latent space which leads to
distortions when the data of interest are projected into the latent space for visualisation.
Since the responsibility matrix R can be used to determine a weight for each of the latent
samples it is possible to update these samples using a resampling scheme.
We propose to use a resampling scheme based upon adaptive kernel density estimation. The
basic procedure places a Gaussian kernel on each latent sample. This results in a Gaussian
1Note that this differs from the measure in the paper by Bishop et aI, where a rati()-()f-areas
criterion is used, a factor which is unity for zero distortion, but may also be unity for some distortions.
Exploratory Data Analysis Using Radial Basis Function Latent Variable Models
533
mixture representation of the pdf of the latent samples p( x It),
K
p(xlt) = ~PiN(lLi' E i ),
(9)
i=l
where each mixture component is weighted according to the latent sample weight Pi. Initially, the Ei'S are all equal, taking their value from the standard formula of Silverman
(1986),
Ei = hLy,
(10)
where matrix Y is an estimate of the covariance of p( x )and,
(11)
If the kernels are centered exactly on the latent samples, this model artificially inflates the
variance of the latent samples. Following West (1993) we perform kernel shrinkage by
making the lLi take the values
(12)
where jL is the mean of the latent samples. This ensures that there is no artificial inflation
of the variance.
To reduce the redundancy in our initially large number of mixture components, we propose
a kernel reduction scheme in a similar manner to West. However, the scheme used here
differs from that of West and follows a scheme proposed by Salmond (1990). Essentially,
we chose the component with the smallest weight and its nearest neighbour, denoting these
with subscripts 1 and 2 respectively. These components are then combined into a single
component denoted with subscript c as follows,
Ec = Pl[El
+ (lLc
Pc = Pl + P2
(13)
PllLl + P21L2
IL = --=---=
c
Pc
(14)
-lLl)(lL c -lLl)T]
+ P2[E 2 + (lLc
-1L2)(lLc -1L2)T].
(15)
Pc
This procedure is repeated until some stopping criterion is met. The stopping criterion
could be a simple limit upon the number of mixture components ie; smaller than K but
sufficiently large to model the data structure. Alternatively, the average kernel covariance
and between kernel covariance can be monitored and the reduction stopped before some
multiple (eg. 10) of the average kernel covariance exceeds the between kernel covariance.
Once a final mixture density estimate is obtained, a new set of equally weighted latent
samples can be drawn from it. The new latent samples represent a better estimate of the
posterior pdf of the latent samples and can be used, along with the existing RBF mapping,
to calculate a new responsibility matrix R. This procedure can be repeated to obtain a
further improved estimate of the posterior pdf which, after only a couple of iterations can
lead to good estimates of the posterior pdf which further iterations fail to improve upon.
5 RESULTS
A latent variable model based oil a spherically-symmetric Gaussian RBF has been implemented. The weights and the centres of the RBF were initialised so that the solution
best approximated the zer<rdistortion principal components solution for tw<rdimensional
projection.
A. D. Marrs and A. R. Webb
534
For our example we chose to construct a simulated data set with easily identifiable structure.
Four hundred points lying on the letters "NIPS" were sampled and projected onto a sphere
of radius 50 such that the points lay between 25 0 and 175 0 longitude and 750 and 125 0
latitude with Gaussian noise of variance 4.0 on the radius of each point. The resulting data
are shown in figure 1.
ToY dataset
Figure 1: Simulated data.
I ..,:.' .:\ ~.. ~.
"I:
.i,~ ?
.> I .~.,......
I
:: I
I
r...1 I :., ......
?~:
-to.,.
-u
l'
Figure 2: Results for standard GTM model.
Figure 3: Results for regularisedlresampled model.
Figure 2 shows results for the standard GTM (uniform grid of latent samples) projection of
the data to two dimensions. The central figure shows the projection onto the latent space,
exhibiting significant distortion. The left figure shows the projection of the regular grid of
latent samples (red points) into the data space. Distortion of this grid can be easily seen.
The right figure is a plot of the magnification factor as defined in section 3, with mean value
of 4.577. For this data set most stretching occurs at the edges of the latent variable space.
Figure 3 shows results for the regularisedlresampled version of the latent variable model
for A = 1.0. Again the central figure shows the projection onto the latent space after 2
iterations of the resampling procedure. The left-hand figure shows the projection of the
initial regular grid of latent samples into the data space. The effect of regularisation is
evident by the lack of severe distortions. Finally the magnification factors can be seen in
the right-hand figure to be lower, with a mean value of 0.976.
Exploratory Data Analysis Using Radial Basis Function Latent Variable Models
535
6 DISCUSSION
We have considered two developments of the GTM latent variable model: the incorporation
of priors on the allowable model and a resampling approach to the maximum likelihood parameter estimation. Results have been presented for this regularisedlresampling approach
and magnification factors lower than the standard model achieved, using the same RBF
model. However, further reduction in magnification factor is possible with different RBF
models, but the example illustrates that resampling offers a more robust approach. Current
work is aimed at assessing the approach on realistic data sets.
References
Bishop, C.M. and Svensen, M. and Williams, C.K.1. (1997). Magnification factors for the
GTM algorithm. lEE International Conference on Artificial Neural Networks, 465-471.
Bishop, C.M. and Svensen, M. and Williams, C.K.1. (1998). GTM: the generative topographic mapping. Neural Computation, 10,215-234.
Hastie, T. and Stuetzle, W. (1989). Principal curves, Journal of the American Statistical
Association, 84, 502-516.
Hotelling, H. (1933). Analysis of a complex of statistical variables into principal components. Journal of Educational Psychology, 24, 417-441,498-520.
Kramer, M.A. (1991). Nonlinear principal component analysis using autoassociative neural
networks. American Institute of Chemical Engineers Journal, 37(2),233-243.
LeBlanc, M. and Tibshirani, R. (1994). Adaptive principal surfaces. Journal of the American Statistical Association, 89(425), 53-664.
Lowe, D. and Tipping, M. (1996). Feed-forward neural networks and topographic mappings for exploratory data analysis. Neural Computing and Applications, 4, 83-95.
Pearson, K. (1901). On lines and planes of closest fit. Philosophical Magazine, 6, 559-572.
Salmond, D.J. (1990). Mixture reduction algorithms for target tracking in clutter. Signal &
Data processing of small targets, edited by O. Drummond, SPlE, 1305.
Silverman, B.W. (1986). Density Estimation for Statistics and Data Analysis. Chapman &
Hall,1986.
Tibshirani, R. (1992). Principal curves revisited. Statistics and Computing, 2(4), 183-190.
Webb, A.R. (1995). Multidimensional scaling by iterative majorisation using radial basis
functions. Pattern Recognition, 28(5), 753-759.
Webb, A.R. (1996). An approach to nonlinear principal components analysis using
radially-symmetric kernel functions. Statistics and Computing, 6, 159-168.
Webb, A.R. (1997). Radial basis functions for exploratory data analysis: an iterative majorisation approach for Minkowski distances based on multidimensional scaling. Journal
of Classification, 14(2),249-267.
Webb, A.R. (1998). Supervised nonlinear principal components analysis. (submitted for
publication ).
West, M. (1993). Approximating posterior distributions by mixtures. J. R. Statist. Soc B,
55(2), 409-422.
| 1544 |@word version:2 briefly:1 norm:1 nd:1 seek:1 covariance:6 tiw:2 tr:1 reduction:5 initial:1 denoting:1 existing:1 current:2 written:3 realistic:1 plot:2 update:1 resampling:7 discrimination:1 generative:3 fewer:1 plane:2 maximised:1 provides:1 revisited:1 along:1 manner:1 introduce:1 gov:1 lll:2 provided:2 notation:1 finding:1 transformation:8 multidimensional:3 ti:4 axl:1 exactly:1 uk:1 zl:1 normally:1 ly:1 unit:3 before:1 limit:2 subscript:2 might:1 chose:2 range:1 practical:1 maximisation:1 differs:2 silverman:2 procedure:4 stuetzle:2 area:1 projection:6 radial:9 road:1 regular:2 onto:3 context:1 applying:1 dz:2 williams:2 educational:1 regarded:2 exploratory:7 target:3 magazine:1 recognition:3 magnification:8 approximated:1 lay:1 majorisation:2 visualising:1 calculate:1 ensures:2 edited:1 ui:1 radar:1 upon:3 basis:11 easily:2 various:1 gtm:10 effective:1 artificial:2 pearson:2 distortion:9 statistic:3 topographic:3 final:1 xlt:1 leblanc:2 propose:2 relevant:1 zer:1 drummond:1 p:1 assessing:1 andrew:2 svensen:2 nearest:1 p2:2 longitude:1 implemented:1 soc:1 met:1 exhibiting:1 radius:2 correct:1 owing:1 centered:1 tlz:2 require:2 pl:2 lying:1 inflation:1 considered:2 sufficiently:1 hall:1 exp:2 great:1 mapping:6 vary:1 smallest:1 purpose:1 estimation:6 iw:1 weighted:2 gaussian:4 shrinkage:1 varying:1 tll:1 publication:1 modelling:1 likelihood:3 greatly:1 am:1 posteriori:1 dependent:1 el:1 stopping:2 initially:2 hidden:1 visualisation:3 mth:1 transformed:1 classification:1 denoted:1 retaining:1 development:3 special:1 equal:2 once:2 construct:1 having:1 sampling:1 chapman:1 inflates:1 neighbour:1 interest:1 severe:1 mixture:8 pc:3 copyright:1 integral:2 edge:1 re:8 stopped:1 column:1 asking:1 tg:1 wr14:1 hundred:1 uniform:1 too:1 combined:1 st:1 density:4 peak:1 international:1 ie:1 ltn:1 retain:1 lee:1 again:1 central:2 satisfied:1 american:3 derivative:3 til:1 leading:1 toy:1 depends:1 lowe:2 responsibility:2 red:1 hly:1 rbfs:1 il:2 variance:4 eip:1 stretching:1 modelled:2 handwritten:1 lli:2 submitted:1 definition:2 initialised:1 monitored:1 couple:1 sampled:1 dataset:1 radially:1 dimensionality:1 feed:1 tipping:2 supervised:2 improved:1 until:1 hand:2 gjj:1 ei:2 nonlinear:9 lack:1 oil:1 effect:1 multiplier:2 chemical:1 symmetric:4 spherically:1 deal:1 eg:1 ll:1 criterion:4 allowable:3 stress:2 pdf:6 tt:2 ay:1 evident:1 crown:1 jl:1 discussed:2 association:2 measurement:2 significant:1 imposing:1 ai:2 smoothness:1 grid:4 centre:1 surface:2 closest:3 posterior:6 termed:2 preserving:1 seen:2 impose:3 determine:1 signal:2 multiple:1 alan:1 exceeds:1 minimising:1 sphere:1 offer:1 equally:1 j3:6 regression:1 basic:2 optimisation:2 metric:1 essentially:1 iteration:3 represent:3 kernel:10 invert:1 achieved:1 background:2 spacing:1 permissible:1 subject:1 incorporates:1 xj:1 fit:3 psychology:1 hastie:2 reduce:1 pca:5 penalty:1 speech:1 autoassociative:1 aimed:1 characterise:1 clutter:1 extensively:1 statist:1 zj:3 estimated:2 tibshirani:4 shall:2 redundancy:1 four:1 acknowledged:1 drawn:1 pj:2 year:1 sum:1 letter:1 place:1 scaling:3 identifiable:1 constraint:10 deficiency:1 incorporation:1 minkowski:1 according:1 describes:1 smaller:1 em:1 character:1 unity:2 wi:2 tw:1 modification:1 making:1 taken:1 equation:3 pin:1 fail:1 hotelling:2 alternative:2 rp:1 dera:2 denotes:2 approximating:2 occurs:1 strategy:1 marrs:5 rt:1 diagonal:1 distance:1 simulated:2 manifold:1 considers:1 assuming:1 illustration:1 providing:1 webb:12 perform:1 finite:1 situation:1 introduced:2 philosophical:1 nip:1 pattern:1 latitude:1 including:1 natural:1 normality:1 scheme:5 improve:1 prior:3 l2:2 regularisation:3 expect:1 imposes:1 pi:4 course:1 jth:1 salmond:2 institute:1 wide:1 taking:2 distributed:2 curve:2 dimension:1 llc:3 evaluating:1 author:1 commonly:1 adaptive:2 projected:2 forward:1 ec:1 approximate:1 reestimation:1 generalising:1 xi:1 alternatively:1 latent:46 iterative:2 zk:1 robust:1 obtaining:1 investigated:1 complex:1 artificially:1 noise:1 repeated:2 complementary:1 malvern:1 west:4 position:1 xl:2 governed:1 jacobian:1 british:1 formula:1 bishop:8 jt:1 explored:1 magnitude:1 illustrates:1 smoothly:1 tfj:2 lagrange:2 tracking:1 corresponds:1 kramer:2 rbf:10 generalisation:4 determined:3 characterised:1 specifically:1 uniformly:1 principal:9 engineer:1 highdimensional:1 arises:1 |
596 | 1,545 | Independent Component Analysis of
Intracellular Calcium Spike Data
Klaus Prank, Julia Borger, Alexander von zur Miihlen,
Georg Brabant, Christof Schoil
Department of Clinical Endocrinology
Medical School Hannover
D-30625 Hannover
Germany
Abstract
Calcium (Ca2 +)is an ubiquitous intracellular messenger which regulates cellular processes, such as secretion, contraction, and cell
proliferation. A number of different cell types respond to hormonal
stimuli with periodic oscillations of the intracellular free calcium
concentration ([Ca 2 +]i). These Ca2+ signals are often organized
in complex temporal and spatial patterns even under conditions
of sustained stimulation. Here we study the spatio-temporal aspects of intracellular calcium ([Ca 2+]i) oscillations in clonal J3-cells
(hamster insulin secreting cells, HIT) under pharmacological stimulation (Schofi et al., 1996). We use a novel fast fixed-point algorithm (Hyvarinen and Oja, 1997) for Independent Component
Analysis (ICA) to blind source separation of the spatio-temporal
dynamics of [Ca 2 +]i in a HIT-cell. Using this approach we find two
significant independent components out of five differently mixed input signals: one [Ca2+]i signal with a mean oscillatory period of
68s and a high frequency signal with a broadband power spectrum
with considerable spectral density. This results is in good agreement with a study on high-frequency [Ca 2+]j oscillations (Palus
et al., 1998) Further theoretical and experimental studies have to
be performed to resolve the question on the functional impact of
intracellular signaling of these independent [Ca 2 +]i signals.
K. Prank et al.
932
1
INTRODUCTION
Independent component analysis (ICA) (Comon, 1994; Jutten and Herault, 1991)
has recently received much attention as a signal processing method which has been
successfully applied to blind source separation and feature extraction. The goal of
ICA is to find independent sources in an unknown linear mixture of measured sensory data. This goal is obtained by reducing 2nd-order and higher order statistical
dependencies to make the signals as independent as possible. Mainly three different
approaches for ICA exist. The first approach is based on batch computations minimizing or maximizing some relevant criterion functions (Cardoso, 1992; Comon,
1994). The second category contains adaptive algorithms often based on stochastic
gradient methods, which may have implementations in neural networks (Amari et
al., 1996; Bell and Sejnowski, 1995; Delfosse and Loubaton, 1995; Hyvarinen and
Oja, 1996; Jutten and Herault, 1991; Moreau and Macchi, 1993; Oja and Karhunen,
1995). The third class of algorithms is based on a fixed-point iteration scheme for
finding the local extrema of the kurtosis of a linear combination of the observed
variables which is equivalent to estimating the non-Gaussian independent companents (Hyvarinen and Oja 1997). Here we use the fast fixed-point algorithm for
independent component analysis proposed by Hyvarinen and Oja (1997) to analyze
the spatia-temporal dynamics of intracellular free calcium ([Ca2+]i) in a hamster
insulin secreting cell (HIT).
Oscillations of [Ca 2+]i have been reported in a number of electrically excitable and
non-excitable cells and the hypotheses of frequency coding were proposed a decade
ago (Berridge and Galione, 1988). Recent experimental results clearly demonstrate
that [Ca 2 +]i oscillations and their frequency can be specific for gene activation concerning the efficiency as well as the selectivity (Dolmetsch et al., 1998). Cells are
highly compartmentalized structures which can not be regarded as homogenous entities. Thus, [Ca 2+]i oscillations do not occur uniformly throughout the cell but
are initiated at specific sites which are distributed in a functional and nonunifortm
manner. These [Ca 2 +]i oscillations spread across individual cells in the form of
Ca2+ waves. [Ca2+]i gradients within cells have been proposed to initiate cell migration, exocytosis, lymphocyte, killer cell activity, acid secretion, transcellular ion
transport, neurotransmitter release, gap junction regulation, and numerous other
functions (Tsien and Tsien, 1990). Due to this fact it is of major importance to
study the spatia-temporal aspects of [Ca 2 +]i signaling in small sub compartments
using calcium-specific fluorescent reporter dyes and digital videomicroscopy rather
than studying the cell as a uniform entity. The aim of this study was to define the
independent components of the spatia-temporal [Ca 2 +]i signal.
2
2.1
METHODS
FAST FIXED-POINT ALGORITHM USING KURTOSIS FOR
INDEPENDENT COMPONENT ANALYSIS
In Independent Component Analysis (ICA) the original independent sources are unknown. In this study we have recorded the [Ca 2 +]i signal in single HIT-cells under
pharmacological stimulation at different subcellular regions (m = 5) in parallel.
The [Ca 2+]i signals (mixtures of sources) are denoted as Xl, X2, ?? ? , X m . Each Xi
is expressed as the weighted sum of n unknown statistically independent compa-
Independent Component Analysis ofIntracellular Calcium Spike Data
933
nents (ICs), denoted as SI, S2, ?.? , Sn. The components are assumed to be mutually
statistically independent and zero-mean. The measured signals Xi as well as the independent component variables can be arranged into vectors x = (Xl, X2, ?.. ,XIIl )
and 8 = (81,82, ... , 8 n ) respectively. The linear relationship is given by:
X=A8
(I)
Here A is a constant mixing matrix whose elements aij are the unknown coefficients
of the mixtures. The basic problem of ICA is to estimate both the mixing matrix
A and the realizations of the Si using only observations of the mixtures X j. In
order to perform ICA, it is necessary to have at least as many mixtures as there
are independent sources (m 2: n). The assumption of zero mean of the ICs is no
restriction, as this can always be accomplished by subtracting the mean from the
random vector x. The ICs and the columns of A can only be estimated up to a
mUltiplicative constant, because any constant multiplying an IC in eq. 1 could be
cancelled by dividing the corresponding column of the mixing matrix A by the same
constant. For mathematical convenience, the ICs are defined to have unit variance
making the (non-Gaussian) ICs unique, up to their signs (Comon, 1994). Here we
use a novel fixed-point algorithm for ICA estimation which is based on 'contrast'
functions whose extrema are closely connected to the estimation of ICs (Hyvarinen
and OJ a, 1997). This method denoted as fast fixed-point algorithm has a number
of desirable properties. First, it is easy to use, since there are no user-defined
parameters. Furthermore, the convergence is fast, conventionally in less than 15
steps and for an appropriate contrast function, the fixed-point algorithm is much
more robust against outliers than most ICA algorithms.
Most solutions to the ICA problem use the fourth-order cumulant or kurtosis of the
signals, defined for a zero-mean random variable x as:
(2)
where E{ x} denotes the mathematical expectation of x. The kurtosis is negative for
source signals whose amplitude has sub-Gaussian probability densitites (distribution
flatter than Gaussian, positive for super Gaussian) sharper than Gaussian, and zero
for Gausssian densities. Kurtosis is a contrast function for ICA in the following
sense. Consider a linear combination of the measured mixtures x, say wTx, where
the vector w is constrained so that E{(w T X}2} = 1. When w T x = ?Si, for some i,
i.e. when the linear combination equals, up to the sign, one of the ICs, the kurtosis
of w T x is locally minimized or maximized. This property is widely used in ICA
algorithms and forms the basis of the fixed-point algorithm used in this study which
finds the relevant extrema of kurtosis also for non-whitened data. Based on this fact,
Hyvarinen and Oja (1997) introduced a very simple and highly efficient fixed-point
algorithm for computing ICA, calculated over sphered zero-mean vectors x, that is
able to find the rows of the separation matrix (denoted as w) and so identify one
independent source at a time. The algorithm which computes a gradient descent
over the kurtosis is defined as follows:
1. Take a random initial vector
Wo
of unit norm. Let 1 = 1.
2. Let WI = E{V(Wf-1V}3} - 3WI-l. The expectation can be estimated using
a large sample of Vk vectors.
K. Prank et al.
934
3. Divide WI by its norm (e.g. the Euclidean norm
II W 11=
.J~i
wn?
4. If 1WfWI-l 1is not close enough to 1, let 1 = 1 + 1 and go back to step 2.
Otherwise, output the vector WI.
To calculate more than one solution, the algorithm may be run as many times as
required. It is nevertheless, necessary to remove the information contained in the
solutions already found, to estimate each time a different independent component.
This can be achieved, after the fourth step of the algorithm, by simply subtracting
the estimated solution 8 = w T v from the unsphered data x.
In the first step of analysis we determined the eigenvalues of the covariance matrix
of the measured [Ca 2+]i signals Si to reduce the dimensionality of the system.
Then the fast fixed-point algorithm was run using the experimental [Ca 2 +]i data to
determine the lOs. The resulting lOs were analyzed in respect to their frequency
content by computing the Fourier power spectrum.
2.2
MEASUREMENT OF INTRACELLULAR CALCIUM IN
HIT-CELLS
To measure [Ca2+]i' HIT (hamster insulin secreting tumor)-cells were loaded with
the fluorescent indicator Fura-2/ AM and Fura-2 fluorescence was recorded at five
different subcellular regions in parallel using a dual excitation spectrofluorometer
videoimaging system. The emission wavelength was 510 nm and the excitation
wavelengths were 340 nm and 380 nm respectively. The ration between the excitation wavelength (F340nm/ F38onm) which correlates to [Ca2+]i was sampled at a rate
of 1 Hz over 360 s. [Ca 2 +]i spikes in this cell were induced by the administration
of 1 nM arginine vasopressin (AVP).
3
RESULTS
From the five experimental [Ca 2 +]i signals (Fig. 1) we determined two significant
eigenvalues of the covariance matrix. The fixed-point algorithm converged in less
than 15 steps and yielded two different lOs, one slowly oscillating component with
a mean period of 68 s and one component with fast irregular oscillations with a flat
broadband power spectrum (Fig. 2). The spectral density of the second component
was considerably larger than that for the high-frequency content of the first slowly
oscillating component.
4
CONCLUSIONS
Ohanges in [Ca 2 +]i associated with Ca 2+ oscillations generally do not occur uniformly throughout the cell but are initiated at specific sites and are able to spread
across individual cells in the form of intracellular Ca2+ waves. Furthermore, Ca 2+
signaling is not limited to single cells but occurs between adjacent cells in the form of
intercellular Ca 2 + waves. The reasons for these spatio-temporal patterns of [Ca2+]i
are not yet fully understood. It has been suggested that information is encoded in
the frequency, rather than the amplitude, of Ca2+ oscillations, which has the advantage of avoiding prolonged exposures to high [Ca2+]i. Another advantage of
935
Independent Component Analysis ofIntracellular Calcium Spike Data
=}0~~J~~~j
j~~~~2i
o
50
100
150
200
250
300
-4
{k~~g
LL
0
50
100
150
o
50
100
150
-~
-4
=~~:
o
50
100
200
250
300
: : : :~j
150
200
250
300
200
250
300
rim. (5)
Figure 1: Experimental time series of [Ca2+]i in a ,B-cell (insulin secreting cell from
a hamster, HIT-cell) determined in five subcellular regions. The data are given as
the ratio between both exciation wavelengths of 340 nm and 380 nm respectively
which correspond to [Ca2+k [Ca2+]i can be calculated from this ratio. The plotted
time series are whitened.
frequency modulated signaling is its high signal-to-noise ratio. In the spatial domain, the spreading of a Ca 2+ oscillation as a Ca2+ wave provides a mechanism
by which the regulatory signal can be distributed throughout the cell. The extension of Ca2+ waves to adjacent cells by intercellular communication provides one
mechanism by which multicellular systems can effect coordinated and cooperative
cell responses to localized stimuli. In this study we demonstrated that the [Ca2+]i
signal in clonal ,B-cells (HIT cells) is composed of two independent components
using spatio-temporal [Ca 2 +]i data for analysis. One component can be described
as large amplitude slow frequency oscillations whereas the other one is a high frequency component which exhibits a broadband power spectrum. These results are
in good agreement with a previous study where only the temporal dynamics of
[Ca 2 +]i in HIT cells has been studied. Using coarse-grained entropy rates computed from information-theoretic functionals we could demonstrate in that study
that a fast oscillatory component of the [Ca2+]i signal can be modulated pharmacologically suggesting deterministic structure in the temporal dynamics (Palu8
et al., 1998). Since Ca2+ is central to the stimulation of insulin secretion from
pancreatic ,B-cells future experimental and theoretical studies should evaluate the
impact of the different oscillatory components of [Ca 2+]i onto the secretory process as well as gene transcription. One possibility to resolve that question is to
use a recently proposed mathematical model which allows for the on-line decoding
of the [Ca 2+]i into the cellular response represented by the activation (phospho-
K. Prank et al.
936
-0 '----- - -- - -----'
a
100
200
sao
100
'D.
C
'd.
.00
0
10- 6 L - -_
10 -?0?'------:0,,'-0::-":
,2""'
0,':---:"
0,-:0 --=-'
0 .?
frequency (Hz)
200
limets )
rime (S)
a
_
0.1
_
0 .2
_ _ _- - - '
O. S
0 .4
0 .5
frequency (Hz)
Figure 2: Results from the independent component analysis by the fast fixed-point
algorithm. Two independent components of [Ca 2+]i were found. A: slowlyoscillating [Ca 2+]i signal, B: fast oscillating [Ca 2 +]i signal. Fourier power spectra of the
independent components. C: the major [Ca 2+]i oscillatory period is 68 s, D: flat
broadband power spectrum.
rylation) of target proteins (Prank et al., 1998). Very recent experimental data
clearly demonstrate that specificty is encoded in the frequency of [Ca2+]i oscillations. Rapid oscillations of [Ca2+]j are able to stimulate a set of transcription
factors in T-Iymphocytes whereas slow oscillations activate only one transcription
factor (Dolmetsch et al., 1998). Frequency-dependent gene expression is likely to
be a widespread phenomenon and oscillations of [Ca 2+]i can occur with periods
of seconds to hours. The technique of independent component analyis should be
able to extract the spatio-temporal features of the [Ca2+]i signal in a variety of
cells and should help to understand the differential regulation of [Ca2+]i-dependent
intracellular processes such as gene transcription or secretion.
Acknowledgements
This study was supported by Deutsche Forschungsgemeinschaft under grants
Scho 466/1-3 and Br 915/4-4.
Independent Component Analysis ofIntracellular Calcium Spike Data
937
References
Amari, S., Cichocki, A. & Yang, H. (1996) A new learning algorithm for blind source
separation. In Touretzky, D.S., Mozer, M. C. & Hasselmo, M. E., (eds.), Advances
in Neural Information Processing 8, pp. 757-763. Cambridge, MA: MIT Press.
Bell, A. & Sejnowski, T. (1995) An information-maximization approach to blind
separation and blind deconvolution. Neural Computation 7:1129-1159.
Berridge, M. & Galione, A. (1988) Cytosolic calcium oscillators. FASEB 2:30743082.
Cardoso, J. F. (1992) Iterative techniques for blind source separation using only
fourth-order cumulants. In Proc. EUSIPCO (pp. 739-742) . Brussels.
Comon, P. (1994) Independent component analysis - a new concept? Signal Procesing 36:287-314.
Delfosse, N. & Loubaton, P. (1995) Adaptive blind separation of independent
sources: a deflation approach. Signal Processing 45:59-83.
Dolmetsch, R E., Xu, K. & Lewis, R S. (1998) Calcium oscillations increase the
efficiency and specificity of gene expression. Nature 392:933-936.
Hyvarinen, A. & Oja, E. (1996) A neuron that learns to separate one independent
component from linear mixtures. In Proc. IEEE Int. Conf. on Neural Networks,
pp. 62-67, Washington, D.C.
Hyvarinen, A. & Oja, E. (1997) A fast fixed-point algorithm for independent component analysis. Neural Computation 9:1483-1492.
Jutten, C. & Herault, J. (1991) Blind separation of sources, part I: An adaptive
algorithm based on neuromimetic architecture. Signal Processing 24:1-10.
Moureau, E., & Macchi, O. (1993) New self-adaptive algorithms for source separation based on contrast functions. In Proc. IEEE Signal Processing Workshop on
Higher Order Statistics, pp. 215-219, Lake Tahoe, USA.
OJ a, E. & Karhunen, J. (1995) Signal separation by nonlinear hebbian learning.
In Palaniswami, M., Attikiouzel, Y., Marks, R, Fogel, D. & Fukuda, T. (eds.)
Computational Intelligence - a Dynamic System Perspective pp. 83-97. IEEE Press,
New York.
Palus, M., Schafl, C., von zur Miihlen, A., Brabant, G. & Prank, K. (1998) Coarsegrained entropy rates quantify fast Ca 2 + dynamics modulated by pharmacological
stimulation. Pacific Symposium on Biocomputing 1998:645-656.
Prank, K., Laer, L., Wagner, M., von zur Miihlen, A., Brabant, G. & Schafl, C.
(1998) Decoding of intracellular calcium spike trains. Europhys. Lett. 42:143-147.
Schafl, C., ROssig, L., Leitolf, H., Mader, T., von zur Miihlen, A. & Brabant, G.
(1996) Generation of repetitive Ca 2+ transients by bombesin requires intracellular release and influx of C a2+ through voltage-dependent and voltage independent
channels in single HIT cells. Cell Calcium 19(6):485-493.
Tsien, R W . & Tsien, R. Y. (1990) Calcium channels, stores, and oscillations.
Annu. Rev. Cell BioI. 6:715-760.
| 1545 |@word norm:3 nd:1 pancreatic:1 contraction:1 covariance:2 initial:1 contains:1 series:2 activation:2 si:4 yet:1 remove:1 intelligence:1 coarse:1 provides:2 tahoe:1 five:4 mathematical:3 differential:1 symposium:1 sustained:1 manner:1 pharmacologically:1 ica:13 rapid:1 secretory:1 proliferation:1 resolve:2 prolonged:1 estimating:1 deutsche:1 killer:1 finding:1 extremum:3 temporal:11 hit:10 unit:2 medical:1 grant:1 christof:1 positive:1 understood:1 local:1 eusipco:1 initiated:2 studied:1 limited:1 statistically:2 unique:1 intercellular:2 signaling:4 bell:2 specificity:1 protein:1 convenience:1 close:1 onto:1 restriction:1 equivalent:1 deterministic:1 demonstrated:1 maximizing:1 go:1 attention:1 exposure:1 regarded:1 target:1 user:1 hypothesis:1 agreement:2 element:1 cooperative:1 observed:1 calculate:1 region:3 connected:1 mozer:1 ration:1 dynamic:6 efficiency:2 basis:1 differently:1 represented:1 neurotransmitter:1 train:1 fast:12 activate:1 sejnowski:2 klaus:1 europhys:1 whose:3 encoded:2 widely:1 larger:1 say:1 amari:2 otherwise:1 hormonal:1 statistic:1 insulin:5 advantage:2 eigenvalue:2 kurtosis:8 subtracting:2 relevant:2 realization:1 mixing:3 subcellular:3 los:3 convergence:1 oscillating:3 unsphered:1 help:1 measured:4 school:1 received:1 wtx:1 eq:1 dividing:1 quantify:1 closely:1 stochastic:1 transient:1 rime:1 extension:1 ic:8 major:2 nents:1 a2:1 delfosse:2 estimation:2 proc:3 spreading:1 fluorescence:1 hasselmo:1 successfully:1 weighted:1 mit:1 clearly:2 gaussian:6 always:1 aim:1 super:1 hannover:2 rather:2 voltage:2 release:2 emission:1 vk:1 mainly:1 contrast:4 sense:1 wf:1 am:1 dependent:3 germany:1 dual:1 denoted:4 herault:3 spatial:2 constrained:1 homogenous:1 equal:1 extraction:1 washington:1 future:1 minimized:1 stimulus:2 loubaton:2 oja:8 composed:1 individual:2 hamster:4 highly:2 possibility:1 mixture:7 analyzed:1 secretion:4 palus:2 necessary:2 divide:1 euclidean:1 plotted:1 theoretical:2 column:2 cumulants:1 maximization:1 lymphocyte:1 uniform:1 reported:1 dependency:1 periodic:1 considerably:1 migration:1 density:3 decoding:2 von:4 central:1 recorded:2 nm:6 slowly:2 berridge:2 conf:1 suggesting:1 coding:1 flatter:1 coefficient:1 int:1 coordinated:1 blind:8 performed:1 multiplicative:1 analyze:1 wave:5 parallel:2 palaniswami:1 compartment:1 acid:1 variance:1 loaded:1 maximized:1 correspond:1 identify:1 multiplying:1 ago:1 converged:1 oscillatory:4 messenger:1 touretzky:1 ed:2 against:1 frequency:14 pp:5 associated:1 sampled:1 dimensionality:1 ubiquitous:1 organized:1 amplitude:3 rim:1 back:1 higher:2 compartmentalized:1 response:2 arranged:1 furthermore:2 transport:1 nonlinear:1 widespread:1 jutten:3 stimulate:1 usa:1 effect:1 concept:1 adjacent:2 pharmacological:3 ll:1 self:1 excitation:3 criterion:1 theoretic:1 demonstrate:3 julia:1 novel:2 recently:2 scho:1 stimulation:5 functional:2 regulates:1 significant:2 measurement:1 compa:1 cambridge:1 spatia:3 recent:2 dye:1 perspective:1 store:1 selectivity:1 accomplished:1 determine:1 period:4 signal:27 ii:1 desirable:1 hebbian:1 clinical:1 concerning:1 impact:2 j3:1 basic:1 whitened:2 expectation:2 iteration:1 repetitive:1 achieved:1 cell:36 zur:4 irregular:1 whereas:2 ion:1 source:13 hz:3 induced:1 yang:1 forschungsgemeinschaft:1 easy:1 wn:1 enough:1 variety:1 architecture:1 reduce:1 br:1 administration:1 expression:2 wo:1 york:1 generally:1 cardoso:2 locally:1 category:1 exist:1 sign:2 estimated:3 georg:1 nevertheless:1 sum:1 run:2 fourth:3 ca2:24 respond:1 throughout:3 separation:10 oscillation:18 lake:1 yielded:1 activity:1 occur:3 x2:2 flat:2 influx:1 aspect:2 fourier:2 department:1 pacific:1 brussels:1 combination:3 electrically:1 across:2 wi:4 rev:1 making:1 comon:4 outlier:1 macchi:2 mutually:1 mechanism:2 deflation:1 initiate:1 neuromimetic:1 studying:1 junction:1 coarsegrained:1 spectral:2 appropriate:1 cancelled:1 batch:1 original:1 denotes:1 fukuda:1 question:2 already:1 spike:6 occurs:1 concentration:1 exhibit:1 gradient:3 separate:1 reporter:1 entity:2 procesing:1 cellular:2 reason:1 relationship:1 ratio:3 minimizing:1 regulation:2 sharper:1 negative:1 implementation:1 calcium:15 unknown:4 perform:1 observation:1 neuron:1 descent:1 prank:7 communication:1 introduced:1 required:1 fogel:1 hour:1 able:4 suggested:1 pattern:2 oj:2 power:6 indicator:1 scheme:1 numerous:1 conventionally:1 excitable:2 extract:1 cichocki:1 sn:1 acknowledgement:1 fully:1 mixed:1 generation:1 fluorescent:2 localized:1 digital:1 sao:1 row:1 supported:1 free:2 aij:1 understand:1 wagner:1 moreau:1 distributed:2 calculated:2 lett:1 computes:1 sensory:1 adaptive:4 hyvarinen:8 correlate:1 functionals:1 arginine:1 transcription:4 gene:5 assumed:1 spatio:5 xi:2 spectrum:6 regulatory:1 iterative:1 decade:1 nature:1 channel:2 robust:1 ca:33 complex:1 domain:1 spread:2 intracellular:11 s2:1 noise:1 xu:1 site:2 broadband:4 fig:2 slow:2 sub:2 xl:2 third:1 learns:1 grained:1 annu:1 specific:4 deconvolution:1 workshop:1 importance:1 multicellular:1 faseb:1 karhunen:2 gap:1 endocrinology:1 entropy:2 tsien:4 simply:1 wavelength:4 likely:1 gausssian:1 expressed:1 contained:1 a8:1 lewis:1 ma:1 bioi:1 goal:2 oscillator:1 considerable:1 content:2 determined:3 reducing:1 uniformly:2 tumor:1 experimental:7 mark:1 sphered:1 modulated:3 alexander:1 cumulant:1 biocomputing:1 avoiding:1 evaluate:1 phenomenon:1 |
597 | 1,546 | Vertex Identification in High Energy
Physics Experiments
Gideon Dror*
Department of Computer Science
The Academic College of Tel-Aviv-Yaffo, Tel Aviv 64044 , Israel
Halina Abramowicz t David Horn t
School of Physics and Astronomy
Raymond and Beverly Sackler Faculty of Exact Sciences
Tel-Aviv University, Tel Aviv 69978 , Israel
Abstract
In High Energy Physics experiments one has to sort through a high
flux of events , at a rate of tens of MHz, and select the few that are
of interest. One of the key factors in making this decision is the
location of the vertex where the interaction , that led to the event ,
took place. Here we present a novel solution to the problem of
finding the location of the vertex , based on two feedforward neural networks with fixed architectures , whose parameters are chosen
so as to obtain a high accuracy. The system is tested on simulated data sets , and is shown to perform better than conventional
algorithms.
1
Introduction
An event in High Energy Physics (HEP) is the experimental result of an interaction
during the collision of particles in an accelerator . The result of this interaction is
the production of tens of particles, each of which is ejected in a different direction
and energy. Due to the quantum mechanical effects involved, the events differ from
one another in the number of particles produced , the types of particles, and their
energies. The trajectories of produced particles are detected by a very large and
sophisticated detector .
? gideon@server.mta.ac.il
thalina@Dost.tau.ac.i1
*hom@n;uron.tau.ac.il
869
Vertex Identification in High Energy Physics Experiments
Events are typically produced at a rate of 10 MHz, in conjunction with a data
volume of up to 500 kBytes per event. The signal is very small, and is selected from
the background by multilevel triggers that perform filtering either through hardware
or software. In the present paper we confront one problem that is of interest in these
experiments and is part of the triggering consideration. This is the location of the
vertex of the interaction. To be specific we will use a simulation of data collected by
the central tracking detector [1] of the ZEUS experiment [2] at the HEP laboratory
DESY in Hamburg, Germany. This detector, placed in a magnetic field , surrounds
the interaction point and is sensitive to the path of charged particles. It has a
cylindrical shape around the axis, z, where the interaction between the incoming
particles takes place. The challenge is to find an efficient and fast method to extract
the exact location of the vertex along this axis.
2
The Input Data
An example of an event, projected onto the z = 0 plane, is shown in Figure 1. Only
the information relevant to triggering is used and displayed. The relevant points,
which denote hits by the outgoing particles on wires in the detector , form five rings
due to the concentric structure of the detector. Several slightly curved particle
tracks emanating from the origin, which is marked with a + sign, and crossing
all five rings, can easily be seen. Each track is made of 30-40 data points. All
tracks appear in this projection as arcs, and indeed, when viewed in 3 dimensions,
every particle follows a helical trajectory due to the solenoidal magnetic field in the
detector.
. "1:.-
60
40
20
Eo
.?
-20
-40
-60
-60 -40 -20
0
x[cml
20
40
60
Figure 1: A typical event projected onto the z = 0 plane. The dots, or hits , have
a two-fold ambiguity in the determination of the xy coordinates through which the
particle has moved. The correct solutions lie on curved tracks that emanate from
the origin.
Each physical hit is represented twice in Fig. 1 due to an inherent two-fold ambiguity
in the determination of its xy coordinates. The correct solutions form curved tracks
emanating from the origin. Some of those can be readily seen in the data. Due to
the limited time available for decision making at the trigger level, the z coordinate
is obtained from the difference in arrival times of a pulse at both ends of the CTD
and is available for only a fraction of these points. The hit resolution in xy is
'" 230 J.lm , while that of z-by-timing is ::: 4 cm. The quality of the z coordinate
G. Dror. H. Abramowicz and D. Hom
870
information is exemplified in figure 2. Figure 2(a) shows points forming a track of
a single particle on the z = 0 projection. Since the corresponding track forms a
helix with small curvature, one expects a linear dependence of the z coordinate of
the hits on their radial position, r = J x 2 + y2. Figure 2(b) compares the values of
r with the measured z values for these points. The scatter of the data around the
linear regression fit is considerable.
10~~-~-,..--~-~-~~-,
35,--,--,-,1---,1-...,--,..-----.1--1.----,
301-
a)
90
:-r. -
-
25f-
~20f-
-
>-
b)
80
70
E
~60
N
_
...;
101-
-
50
40
30
-
51I I I
10
20
30
40
50
x [cm)
'I
60
70
80
20
1~5
20
25
30
35
40
r[cm)
45
50
55
Figure 2: A typical example of uncertainties in the measured z values: (a) a single
track taken from the event shown in figure 1, (b) the z coordinate vs r
Jx 2 + y2
the distance from the z axis for the data points shown in (a). The full line is a
linear regression fit.
=
3
The Network
Our network is based on step-wise changes in the representation of the data, moving
from the input points, to local line segments and to global arcs. The nature of
the data and the problem suggest it is best to separate the treatment of the xy
coordinates from that of the z coordinate. Two parallel networks which perform
entirely different computations, form our final system. The first network, which
handles the xy information is responsible for constructing arcs that correctly identify
some of the particle tracks in the event. The second network uses this information
to evaluate the z location of the point where all tracks meet .
3.1
Arc Identification Network
The arc identification network processes information in a fashion akin to the method
visual information is processed by the primary visual system [3].
The input layer for this network is made of a large number of neurons (several tens
of thousands) and corresponds to the function of the retina. Each input neuron
has its distinct receptive field. The sum of all fields covers completely the relevant
domain in the xy plane. This domain has 5 concentric rings, which show up in
figure 1. The total area of the rings is about 5000 cm 2 , and covering it with 100000
input neurons leads to satisfactory resolution. A neuron in the input level fires
when a hit is present in its receptive field. We shall label each input neuron by the
(xy) coordinates of the center of its receptive field.
Neurons of the second layer are line segment detectors. Each second layer neuron
is labeled by (XY a), where (X, Y) are the coordinates of the center of the segment
Vertex Identification in High Energy Physics Experiments
and
0'
871
denotes its orientation. The activation of second layer neurons is given by
VXYa
= g(2:: J XY a ,xy V xy
- ( 2) ,
(1)
xy
where
lxY a ,ry
={
~1
ifr.L < O.5cmArll < 2cm
ifO.5cm< r.L < 1cmArii < 2cm
otherwise
(2)
and g( x) is the standard Heaviside step function . rll and r.L are the parallel and
perpendicular distances between (X , Y) and (x, y) with respect to the axis of the
line segment , defined by 0' . It is important to note that at this level , values of the
threshold 82 which are slightly lower than optimum are preferable, taking the risk
of obtaining superfluous line segments in order to reduce the probability of missing
one . Superfluous line segments are filtered out very efficiently in higher layers.
Figure 3 represents the output of the second layer neurons for the input illustrated
by the event of figure 1. An active second layer neuron (XY 0') is represented in
this figure by a line segment centered at the point (X , Y) making an angle 0' with
the x axis. The length of the line segments is immaterial and was chosen only for
the purpose of visual clarity.
60
~
40
...
...
"Z
~
20
E
~
0
>-
-20
-40
-60
#- .>~~
1'1(
-t!-
s
~
~
,
.
~.::.. ~
~
.
"\0
~ ~.
~
~1t
..,. ~, .
~
'
"'" "
i-!.
'"
;J
I'
l-J..
-60 -40 -20
.~
0
xfcml
20
40
60
Figure 3: Representation of the activity of second layer neurons XY 0' for the input
of figure 1 taken by plotting the appropriate line segments in the xy plane . At some
XY locations several line segments with different directions occur due to the rather
low threshold parameter used , 82 = 4.
Neurons of the third layer transform the representation of local line segments into
local arc segments. An arc which passes through the origin is uniquely defined by
its radius of curvature R and its slope at the origin . Thus , each third layer neuron
is labeled by '" 8 i , where 1"'1 = 1/ R is the curvature and the sign of '" determines
the orient ation of the arc . 1 < i < 5 is an index which relates each arc segment to
the ring it belongs to.
-The mapping between second and third layers is based on a winner-take-all mechanism. Namely, for a given local arc segment, we take the arc segment which is
closest to being tangent to the local arc segment.
Denoting the average radius of the ring i ( i=1 ,2, ...5) by rj and using f3i = sin -1 (y)
G. Dror. H. Abramowicz and D. Horn
872
the final expression for the activation of the third layer neurons is
V",lIi
_0
= maxe
0<3
2
2
cos (() - 2f3i - 0:),
(3)
where 6 = 6(X , Y, "', (), i) = J(X - ri cos((} - f3d)2 + (Y - ri sin(() - f3d)2 is simply
the distance of the center of the receptive field of the (XY 0:) neuron to the ("'(})
arc.
The fourth layer is the last one in the arc identification network . Neurons belonging
to this layer are global arc detectors. In other words, they detect projected tracks
on the z 0 plane. A fourth level neuron is denoted by "'(} , where", and () have the
previous meaning , now describing global arcs. Fourth layer neurons are connected
to third layer neurons in a simple fashion ,
=
Vd
= g( L
6""",,611 ,11' V""II'i - (}4) .
(4)
",'II' i
Figure 4 represents the activity of fourth layer neurons . Each active neuron "'(} is
equivalent in the xy plane to one arc appearing in the figure .
. ~60
40
20
E
~o
>-
~
-20
-40
-60
-60 -40 -20
x f<em] 20
40
60
Figure 4: Representation of the activity of fourth layer neurons "'(} for the input of
figure 1 taken by plotting the appropriate arcs in t he xy plane. The arcs are not
precisely congruent to the activity of the input layer which is also shown , due to
the finite widths which were used, il", = 0.004 and il(} = 7r/20. This figure was
produced with (}4 = 3.
3.2
z Location Network
The architecture of the second network has a structure which is identical to the first
one, although its computational task is different. We will use an identical labeling
system for its neurons , but denote their activities by v xy . The latter will assume
continuous values in this network.
A first layer neuron of the z-location network receives its input from the same
receptive field as its corresponding neuron in the first network. Its value , v xy , is the
mean value of the z values of the points within its receptive field . If no z values are
available for these points , a null value is assigned to it.
The second layer neurons compute the mean value vXY a = (v xy ) of the z coordinate
of the first layer neurons in their receptive field , averaging over all neurons within
873
Vertex Identification in High Energy Physics Experiments
the section
II(x -
X) sina - (y - Y) cosal < 0.5cm/\ (x - X)2 + (y - y)2 < 4cm2} ,
which corresponds to the excitatory part of the synaptic connections of equation
(2). If null values appear within that section they are disregarded by the averaging
procedure. If all values are null , VXYa is assigned a null value too . This Z averaging
procedure is similarly propagated to the third layer neurons.
{xy
The fourth layer neurons evaluate the Z value of the origin of each arc identified
by the first network. This is performed by a simple linear extrapolation. The final
z estimate of the vertex, Znet , which should be the common origin of all arcs , IS
calculated by averaging the outputs of all active fourth layer neurons.
4
Results
In order to test the network , we ran it over a set of 1000 events generated by a
Monte-Carlo simulator as well as over a sample of physical events taken from the
ZEUS experiment at the HEP laboratory DESY in Hamburg . For the former set
we compared the estimate of the net Znet with the nominal location of the vertex z ,
whereas for the real events in the latter set , we compared it with an estimate Zrec
obtained by full reconstruction algorithm , which runs off-line and uses all available
data. Results of the two tests can be compared since it is well established that the
result of the full reconstruction algorithm is within 1 mm from the exact location
of the vertex.
z
z
Network
<Az>=-2.7?O.2
(1 = 6.1 ?O.2
140
140
120
120
100
100
80
80
60
60
40
40
20
0
-40
J ~~
-20
0
20
Histogrom
<Az>= 1.9?O.3
(1 = 8.4?O.3
20
40
0
Az [em]
Figure 5: Distribution of ~ z = Ze8timate - Zexact values for two types of estimates,
(a) the one proposed in this paper and (b) the one based on a commonly used
histogram method .
We also compared our results with those of an algorithmic method used for triggering at ZEUS [4]. We shall refer to this method as the 'histogram method '. The
performance of the two methods was compared on a sample of 1000 Monte-Carlo
events. The network was unable to get an estimate for 16 events from the set ,
as compared with 15 for the histogram method (15 of those events were common
G. Dror, H. Abramowicz and D. Horn
874
failures). In Figure 5 we compare the distributions of ~z = Znet - Zexact and
~Z = Zhist - Zexact for the sample of Monte-Carlo events , where Zexact is the generated location of the vertex. Both methods lead to small biases, -2.7 cm for Znet
and 1.9 cm for Zhist . The resolution, as obtained from a Gaussian fit , was found
to be better for the network approach (06.1 cm) as compared to the histogram
method (08.4cm). In addition, it should be noted that the histogram method
yields discrete results, with a step of 10 cm, whereas the current method gives continuous values. This can be of great advantage for further processing. Note that
off-line, after using the whole CTD information, the resolution is better than 1 mm.
=
5
=
Discussion
We have described a feed forward double neural network that performs a task of
pattern identification by thresholding and selecting subsets of data on which a simple
computation can lead to the final answer. The network uses a fixed architecture,
which allows for its implementation in hardware, crucial for fast triggering purposes.
The basic idea of using a fixed architecture that is inspired by the way our brain
processes visual information, is similar to the the raison d 'etre of the orientation
selective neural network employed by [5]. The latter was based on orientation selective cells only, which were sufficient to select linear tracks that are of interest in
HEP experiments. Here we develop an arc identification method, following similar
steps. Both methods can also be viewed as generalizations of the Hough transform [6] that was originally proposed for straight line identification and may be
regarded as a basic element of pattern recognition problems [7]. Neither [5] nor
our present proposal were considered by previous neural network analyses of HEP
data [8] . The results that we have obtained are very promising. We hope that they
open the possibility for a new type of neural network implementation in triggering
devices of HEP experiments.
Acknowledgments
We are indebted to the ZEUS Collaboration whose data were used for this study.
This research was partially supported by the Israel National Science Foundation .
References
[1] B. Foster et al. , Nuclear Instrum. and Methods in Phys. Res. A338 (1994) 254.
[2] ZEUS Collab., The ZEUS Detector, Status Report 1993, DESY 1993; M.
Derrick et al. , Phys. Lett. B 293 (1992) 465 .
[3] D. H. Hubel and T . N. Wiesel, J. Physiol. 195 (1968) 215.
[4] A. Quadt , MSc thesis, University of Oxford (1997) .
[5] H. Abramowicz , D. Horn , U. Naftaly and C . Sahar-Pikielny, Nuclear Instrum.
and Methods in Phys. Res. A378 (1996) 305; Advances in Neural Information
Processing Systems 9, eds. M. C . Mozer , M. J. Jordan and T. Petsche, MIT
Press 1997, pp. 925- 931.
[6] P. V. Hough , "Methods and means to recognize complex patterns", U.S. patent
3.069.654.
[7] R. O. Duda and P. E. Hart , "Pattern classification and scene analysis" , Wiley,
New York , 1973.
[8] B. Denby, Neural Computation, 5 (1993) 505.
| 1546 |@word cylindrical:1 faculty:1 wiesel:1 duda:1 open:1 cm2:1 simulation:1 cml:1 pulse:1 denby:1 selecting:1 denoting:1 current:1 activation:2 scatter:1 readily:1 physiol:1 shape:1 v:1 selected:1 device:1 plane:7 kbytes:1 filtered:1 location:11 five:2 along:1 abramowicz:5 indeed:1 nor:1 ry:1 simulator:1 brain:1 inspired:1 null:4 israel:3 cm:13 dror:4 astronomy:1 finding:1 every:1 preferable:1 hit:6 appear:2 timing:1 local:5 oxford:1 meet:1 path:1 twice:1 co:2 limited:1 perpendicular:1 horn:4 responsible:1 acknowledgment:1 procedure:2 emanate:1 area:1 projection:2 word:1 radial:1 suggest:1 get:1 onto:2 risk:1 conventional:1 equivalent:1 charged:1 center:3 rll:1 missing:1 resolution:4 regarded:1 nuclear:2 handle:1 coordinate:11 trigger:2 nominal:1 exact:3 us:3 origin:7 crossing:1 element:1 recognition:1 labeled:2 thousand:1 connected:1 ran:1 mozer:1 immaterial:1 segment:16 completely:1 easily:1 represented:2 distinct:1 fast:2 monte:3 detected:1 emanating:2 labeling:1 whose:2 otherwise:1 transform:2 final:4 advantage:1 net:1 took:1 reconstruction:2 interaction:6 relevant:3 helical:1 moved:1 az:3 double:1 optimum:1 congruent:1 ring:6 develop:1 ac:3 measured:2 school:1 differ:1 direction:2 radius:2 correct:2 centered:1 multilevel:1 generalization:1 mm:2 around:2 considered:1 great:1 mapping:1 algorithmic:1 lm:1 jx:1 purpose:2 label:1 sensitive:1 hope:1 mit:1 gaussian:1 rather:1 conjunction:1 detect:1 typically:1 selective:2 i1:1 germany:1 classification:1 orientation:3 denoted:1 field:10 identical:2 represents:2 report:1 inherent:1 few:1 retina:1 national:1 recognize:1 fire:1 interest:3 possibility:1 superfluous:2 xy:23 hough:2 re:2 hep:6 cover:1 mhz:2 vertex:12 subset:1 expects:1 too:1 answer:1 physic:7 off:2 thesis:1 central:1 ambiguity:2 lii:1 ifo:1 performed:1 extrapolation:1 sort:1 parallel:2 slope:1 il:4 accuracy:1 beverly:1 efficiently:1 yield:1 identify:1 identification:10 produced:4 carlo:3 trajectory:2 straight:1 indebted:1 detector:9 phys:3 synaptic:1 ed:1 failure:1 energy:8 pp:1 involved:1 propagated:1 derrick:1 treatment:1 zeus:6 sophisticated:1 feed:1 higher:1 originally:1 msc:1 outgoing:1 receives:1 quality:1 aviv:4 effect:1 y2:2 lxy:1 former:1 assigned:2 laboratory:2 satisfactory:1 illustrated:1 sin:2 during:1 width:1 uniquely:1 covering:1 noted:1 performs:1 meaning:1 wise:1 consideration:1 novel:1 common:2 physical:2 patent:1 winner:1 volume:1 yaffo:1 he:1 refer:1 surround:1 vxy:1 similarly:1 particle:13 dot:1 moving:1 curvature:3 closest:1 belongs:1 hamburg:2 collab:1 server:1 pikielny:1 seen:2 eo:1 employed:1 signal:1 ii:3 relates:1 full:3 rj:1 academic:1 determination:2 hart:1 regression:2 basic:2 confront:1 histogram:5 cell:1 proposal:1 background:1 whereas:2 addition:1 crucial:1 pass:1 jordan:1 feedforward:1 fit:3 architecture:4 identified:1 triggering:5 f3i:2 reduce:1 idea:1 expression:1 akin:1 york:1 collision:1 ten:3 hardware:2 processed:1 ctd:2 sign:2 per:1 track:12 correctly:1 discrete:1 shall:2 key:1 threshold:2 clarity:1 neither:1 fraction:1 sum:1 orient:1 angle:1 run:1 uncertainty:1 fourth:7 place:2 decision:2 entirely:1 layer:25 hom:2 sackler:1 fold:2 activity:5 occur:1 precisely:1 ri:2 software:1 scene:1 department:1 mta:1 belonging:1 slightly:2 em:2 naftaly:1 making:3 taken:4 equation:1 describing:1 mechanism:1 end:1 available:4 appropriate:2 petsche:1 magnetic:2 appearing:1 denotes:1 ifr:1 receptive:7 primary:1 dependence:1 distance:3 separate:1 unable:1 simulated:1 vd:1 collected:1 length:1 index:1 znet:4 implementation:2 perform:3 wire:1 neuron:31 arc:22 finite:1 curved:3 displayed:1 concentric:2 david:1 namely:1 mechanical:1 connection:1 established:1 exemplified:1 pattern:4 gideon:2 challenge:1 tau:2 event:18 ation:1 axis:5 extract:1 raymond:1 tangent:1 f3d:2 accelerator:1 sahar:1 filtering:1 foundation:1 sufficient:1 plotting:2 thresholding:1 foster:1 helix:1 collaboration:1 production:1 excitatory:1 placed:1 last:1 supported:1 bias:1 taking:1 dimension:1 calculated:1 lett:1 quantum:1 forward:1 made:2 commonly:1 projected:3 flux:1 status:1 global:3 active:3 incoming:1 hubel:1 continuous:2 promising:1 nature:1 tel:4 obtaining:1 complex:1 constructing:1 domain:2 whole:1 arrival:1 fig:1 fashion:2 wiley:1 position:1 lie:1 uron:1 third:6 specific:1 disregarded:1 led:1 desy:3 simply:1 forming:1 visual:4 tracking:1 partially:1 corresponds:2 determines:1 marked:1 viewed:2 considerable:1 change:1 instrum:2 typical:2 averaging:4 total:1 experimental:1 maxe:1 select:2 college:1 latter:3 sina:1 evaluate:2 heaviside:1 tested:1 |
598 | 1,547 | Attentional Modulation of Human Pattern
Discrimination Psychophysics Reproduced
by a Quantitative Model
Laurent Itti, Jochen Braun, Dale K. Lee and Christof Koch
{itti, achim, jjwen, koch}Oklab.caltech.edu
Computation & Neural Systems, MSC 139-74
California Institute of Technology, Pasadena, CA 91125, U.S.A.
Abstract
We previously proposed a quantitative model of early visual processing in primates, based on non-linearly interacting visual filters
and statistically efficient decision. We now use this model to interpret the observed modulation of a range of human psychophysical
thresholds with and without focal visual attention. Our model calibrated by an automatic fitting procedure - simultaneously reproduces thresholds for four classical pattern discrimination tasks,
performed while attention was engaged by another concurrent task.
Our model then predicts that the seemingly complex improvements
of certain thresholds, which we observed when attention was fully
available for the discrimination tasks, can best be explained by a
strengthening of competition among early visual filters.
1
INTRODUCTION
What happens when we voluntarily focus our attention to a restricted part of our
visual field? Focal attention is often thought as a gating mechanism, which selectively allows a certain spatial location and and certain types of visual features to
reach higher visual processes. We here investigate the possibility that attention
might have a specific computational modulatory effect on early visual processing.
We and others have observed that focal visual attention can modulate human psychophysical thresholds for simple pattern discrimination tasks [7, 8, 5] When attention is drawn away from a task, for example by "cueing" [12] to another location
of the display, or by a second, concurrent task [1, 7, 8], an apparently complex
pattern of performance degradation is observed: For some tasks, attention has little or no effect on performance (e.g., detection of luminance increments), while for
790
L. ltti, J. Braun, D. K. Lee and C. Koch
other tasks, attention dramatically improves performance (e.g., discrimination of
orientation). Our specific findings with dual-task psychophysics are detailed below.
These observations have been paralleled by electrophysiological studies of attention.
In the awake macaque, neuronal responses to attended stimuli can be 20% to 100%
higher than to otherwise identical unattended stimuli. This has been demonstrated
in visual cortical areas VI, V2, and V4 [16, 11, 10,9] when the animal discriminates
stimulus orientation, and in areas MT and MST when the animal discriminates the
speed of stimulus motion [17]. Even spontaneous firing rates are 40% larger when
attention is directed at a neuron's receptive field [9]. Whether neuronal responses
to attended stimuli are merely enhanced [17] or whether they are also more sharply
tuned for certain stimulus dimensions [16] remains controversial. Very recently,
fMRI studies have shown similar enhancement (as measured with BOLD contrast)
in area VI of humans, specifically at the retinotopic location where subjects had
been instructed to focus their attention to [2, 14].
All of these observations directly address the issue of the "top-down" computational
effect of attentional focusing onto early visual processing stages. This issue should
be distinguished from that of the "bottom-up" control of visual attention [6], which
studies which visual features are likely to attract the attention focusing mechanism (e.g., pop-out phenomena and studies of visual search). Top-down attentional
modulation happens after attention has been focused to a location of the visual
field, and most probably involves the massive feedback circuits which anatomically
project from higher cortical areas back to early visual processing areas.
In the present study, we quantify the modulatory effect of attention observed in
human psychophysics using a model of early visual processing. The model is based
on non-linearly interacting visual filters and statistically efficient decision [4, 5].
Although attention could modulate virtually any visual processing stage (e.g ., the
decision stage, which compares internal responses from different stimuli), our basic
hypothesis here - supported by electrophysiology and fMRI [16,11,10,17,9,2, 14]is that this modulation might happen very early in the visual processing hierarchy.
Given this basic hypothesis, we investigate how attention should affect early visual
processing in order to quantitatively reproduce the psychophysical results.
2
PSYCHOPHYSICAL EXPERIMENTS
We measured attentional modulation
of spatial vision thresholds using a
Central task:
dual-task paradigm [15, 7]: At the
center of the visual field, a letter discrimination task is presented, while
a pattern discrimination task is simultaneously presented at a random
peripheral location (4 0 eccentricity).
The central task consists of discriminating between five letters "T" or
four "T" and one "L". It has been
shown to efficiently engage attention
[7]. The peripheral task is chosen
from a battery of a classical pattern
threshold measurement
discrimination tasks, and is the task
of interest for this study. Psychophysical thresholds are measured for two distinct conditions: In the "fully attended"
condition, observers are asked to devote their entire attention to the peripheral
791
Quantitative Modeling ofAttentional Modulation
task, and to ignore the central task (while still fixating the center of the screen).
In the "poorly attended" condition, observers are asked to pay full attention to
the central task (and the blocks of trials for which performance for the central task
falls below a certain cut-off are discarded).
Four classical pattern discrimination tasks were investigated, each with two volunteer subjects (average shown in Figure 1), similarly to our previous experiments
[7, 8]. Screen luminance resolution was 0.2%. Screen luminance varied from 1 to
90cd/m 2 (mean 45cd/m 2 ), room illumination was 5cd/m 2 and viewing distance
80cm. The Yes/No (present/absent) paradigm was used (one stimulus presentation
per trial). Threshold (75% correct peformance) was reached using a staircase procedure , and computed through a maximum-likelihood fit of a Weibull function with
two degrees of freedom to the psychometric curves.
Exp. 2: Orientation discrimination
f
20
J
15
., 10
c:
,g
S 5
~ ~~~~~
0.6
0.8
Contrast
Mask contrast
0.4
:2
:2
.B
0
I:
e
II)
II>
?;
-6
!0.2
i.to
0.2
c:
c:
8
0
()
??
?
Figure 1: Psychophysical data and model fits using the parameters from Table 1
(P=poorly and F=fully attended). Gray curves: Model predictions for fully attended
data, using the poorly attended parameters, except for -y = 2.9 and {) = 2.1 (see Results).
Expo 1 measured increment contrast discrimination threshold: The observer discriminates between a 4cpd (cycles per degree) stochastic oriented mask [7] at fixed
contrast , and the same mask plus a low-contrast sixth-derivative-of-Gaussian (D6G)
bar; threshold is measured for bar contrast [8]. Expo 2 measured orientation discrimination thresholds: The observer discriminates between a vertical and tilted
grating at 4cpd; threshold for the angle difference is measured. In addition, two
contrast masking tasks were investigated for their sensitivity to non-linearities in
visual processing. A 4cpd stochastic mask (50% contrast) was always present, and
threshold was measured for the contrast of a vertical superimposed D6G bar. In
Expo 3, the orientation of the masker was varied and its spatial frequency fixed
(4cpd), while in Expo 4 the spatial period of the masker was varied and its orientation vertical. Our aim was to investigate very dissimilar tasks, in particular with
respect to the decision strategy used by the observer.
Using the dual-task paradigm, we found mixed attentional effects on psychophysical
thresholds, including the appearance of a more pronounced contrast discrimination
L. ltti, J. Braun, D. K. Lee and C. Koch
792
"dipper" in Exp. 1, substantial improvement of orientation thresholds in Exp. 2,
and reduced contrast elevations due to masking in Exps. 3-4 (also see [7, 8]).
3
MODEL
The model consists of three
successive stages [4, 5]. In
the first stage, a bank of
Gabor-like linear filters analyzes a fixed location of the
visual scene. Here, a singlescale model composed of 12
pairs of filters in quadrature
phase, tuned for orientations
o E e evenly spanning 1800 ,
was sufficient to account for
the data (although a multiscale model may account for
a wider range of psychophysical thresholds). The linear filters take values between
0.0 and 100.0, then multiplied by a gain factor A (one of the ten free parameters of
the model), and to which a small background activity f. is added.
In the second stage, filters non-linearly interact as follows: (1) Each unit receives
non-linear self-excitation, and (2) each unit receives non-linear divisive inhibition
from a pool of similarly-tuned units: With E8 being the linear response from a unit
tuned for orientation 0, the pooled response R8 is given by:
where
W8(O')
=e-
(/1'_/1)2
2E~
is a Gaussian weighting function centered around 0, and 1J a positive constant to
account for background activity in the pooling stage. This stage is inspired from
Heeger's model of gain control in cat VI [3, 4]. Our formulation, in which none of
the parameters is given a particular value, however allows for multiple outcomes,
to be determined by fitting the model to our psychophysical data: A sigmoidal
(S > 0, I > d') as well as simple power-law (S = 0) or even linear (! = 1, d' = 0)
contrast response characteristic could emerge, the responses could be saturating
d') or not (, i= d'), and the inhibitory pool size (~8) could be broad or narrow.
Because striate neurons are noisy, physiological noise is assumed in the model at
the outputs of the second stage. The noise level is chosen close to what is typically
observed in cortical pyramidal cells, and modeled by Gaussian noise with variance
equal to mean taken to some power a determined by fitting.
(, =
Because the decision stage - which quantitatively relates activity in the population
of pooled noisy units to behavioral discrimination performance - is not fully characterized in humans, we are not in a position to model it in any detail. Instead,
we trained our subjects (for 2-3 hours on each task), and assume that they perform close to an "optimal detector". Such optimal detector may be characterized
in a formal manner, using Statistical Estimation Theory [4, 5]. We assume that a
brain mechanism exists, which, for a given stimulus presentation, builds an internal estimate of some stimulus attribute ( (e.g., contrast, orientation, period). The
central assumption of our decision stage is that this brain mechanism will perform
close to an unbiased efficient statistic T, which is the best possible estimator of (
Quantitative Modeling ofAttentional Modulation
793
given the noisy population response from the second stage. The accuracy (variance) with which T estimates ( can be computed formally, and is the inverse of
the Fisher Information with respect to ( [13, 4]. Simply put, this means that, from
the first two stages of the model alone, we have a means of computing the best
possible estimation performance for (, and consequently, the best possible discrimination performance between two stimuli with parameters (1 and (2 [4, 5]. Such
statistically efficient decision stage is implementable as a neural network [13].
This decision stage provides a unified framework for optimal discrimination in any
behavioral situation, and eliminates the need for task-dependent assumptions about
the strategy used by the observers to perform the task in a near optimal manner.
Our model allows for a quantitative prediction of human psychophysical thresholds,
based on a crude simulation of the physiology of primary visual cortex (area VI).
4
RESULTS
All parameters in the model were automatically adjusted in order to best fit the psychophysical data from all experiments. A multidimensional downhill simplex with
simulated annealing overhead was used to minimize the root-mean-square distance
between the quantitative predictions of the model and the human data [4]. The
best-fit parameters obtained independently for the "fully attended" and "poorly
attended" conditions are reported in Table 1. The model's simultaneous fits to our
entire dataset are plotted in Figure 1 for both conditions. After convergence of
the fitting procedure, a measure of how well constrained each parameter was by the
data was computed as follows: Each parameter was systematically varied around its
best-fit value, in 0.5% steps, and the fitting error was recomputed; the amplitude
by which each parameter could be varied before the fitting error increased by more
than 10% of its optimum is noted as a standard deviation in Table 1. A lower
deviation indicates that the parameter is more strongly constrained by the dataset.
Table 1. Model parameters for both attentional conditions.
Name
Symbol fully attended
poorly attended
Linear gaint
A
l.7 ? 0.2
8.2 ? 0.9
Activity-independent inhibition t
S
14.1 ? 2.3
10l.5 ? 16.6
Excitatory exponent
'Y
3.36 ? 0.02
2.09 ? 0.01
Inhibitory exponent
6
2.48 ? 0.02
l.51 ? 0.02
Noise exponent
a
l.34 ? 0.07
1.39 ? 0.08
Background activity, linear stage
f
l.13 ? 0.35
1.25 ? 0.60
Background activity, pooling stage
7]
0.18 ? 0.05
0.77 ? 0.11
Spatial period tuning width X
(r>.
0.85 ? 0.06 oct. 0.85 ? 0.09 oct .
Orientation tuning width X
(r8
26? ? 2.4?
38? ? 5.5?
Orientation pooling width X
~8
48? ? 25?
50? ? 26?
t Dynamic range of linear filters is [? ... 100.0 X A + 4
x For clarity, FWHM values are given rather than 17 values (FWHM = 2I7J2ln(2?.
Although no human bias was introduced during the fitting procedure, interestingly,
all of the model's internal parameters reached physiologically plausible best-fit values, such as, for example, slightly supra-Poisson noise level (a ~ 1.35), ~ 30?
orientation tuning FWHM (full-width at half-maximum), and ~ 0.85 octave spatial period tuning FWHM. Some of the internal characteristics of the model which
more closely relate to the putative underlying physiological mechanisms are shown
in Figure 2.
794
a
L. ltti, J. Braun, D. K. Lee and C. Koch
b
Transducer function
Orientation tuning
C
0.8
p
0.4
0.8
0.8
5.c: o.s
..c:
en
Ci5
~0.4
~o.s
e 0.4
0.2
0.8
Contrast
Orientation pooling
F
0 -40
-20
0
20
40
Orientation (deg)
Figure 2: Internals of the model. (a) The response function of individual units to contrast
was sigmoidal under full (F) and almost linear under poor (P) attention. (b) Native linear
orientation tuning was broader under poor (NP) than full (NF) attention, but it was
sharpened in both cases by pooling (PP=pooled poor, and PF=pooled full attention). (c)
There was no difference in orientation pooling width under poor (P) or full (F) attention.
Using poorly attended parameters, except for -y = 2.9 and ~ = 2.1 (grey curves), yielded
steep non-linear contrast response, and intermediary tuning (same width as NF).
In Table 1, attention had the following significant effects on the model's parameters: 1) Both pooling exponents (-y, d) were higher; 2) the tuning width (0"/1) was
narrower; 3) the linear gain (A) and associated activity-independent inhibition (5)
were lower; and 4) the background activity of the pooling stage was lower. This
yielded increased competition between filters: The network behaved more like a
winner-take-all under full attention, and more like a linear network of independent
units under poor attention. While the attentional modulation of "d and 0"/1 are
easy to interpret, its effect on the A, 5 and 'fJ is more difficult to understand.
Consequently, we conducted a further automatic fit, which, starting from the
"poorly attended" parameters, was only allowed to alter, and d to fit the "fully
attended" data. The motivation for not varying 0"/1 was that we observed significant
sharpening of the tuning induced by higher exponents "d (Figure 2) . Also, slight
changes in the difference , - d can easily produce large changes in the overall gain
of the system, hence compensating for changes in A, 5 and 'fJ . (We however do not
imply here that 0"/1, A, 5 and 'fJ are redundant parameters; there is only a small range
around the best-fit point over which, and d can compensate for variations in the
other parameters, without dramatically impairing the quality of fit) .
Although the new fit was not as accurate as that obtained with all parameters
allowed to vary, it appeared that a simple modification of the pooling exponents
well captured the effect of attention (Figure 1). Hence, the "poorly attended"
parameters of Table 1 well described the "poorly attended" data, and the same
parameters except for, = 2.9 and d = 2.1 well described the "fully attended" data.
A variety of other simple parameter modifications were also tested, but none except
for the pooling exponents (-y,o) could fully account for the attentional modulation.
These modifications include: Changes in gain (obtained by modifying A only, ,
only, or d only), in tuning (0"/1), in the extent ofthe inhibitory pool (E/I), and in the
noise level (a). A more systematic study, in which all possible parameter subsets
are successively examined, is currently in progress in our laboratory.
5
DISCUSSION and CONCLUSION
At the basis of our results is the hypothesis that attention might modulate the
earlier rather than the later stages of visual processing. We found that a very
Quantitative Modeling ofAttentional Modulation
795
simple, prototypical, task-independent enhancement of the amount of competition
between early visual filters accounts well for the human data. This enhancement
resulted from increases in parameters 'Y and 5 in the model, and was paralleled by
an increase in contrast gain and a sharpening in orientation tuning. Although it is
not possible from our data to rule out any attentional modulation at later stages,
our hypothesis has recently received experimental support that attention indeed
modulates early visual processing in humans [2, 14].
More psychophysical experiments are needed to investigate attentional modulation
at later processing stages. For example, it might be possible to study the effect
of attention on the decision stage by manipulating attention during experiments
involving decision uncertainty. In the absence of such results, we have attempted
in our experiments to minimize the possible impact of attention on later stages,
by using only simple stimulus patterns devoid of conceptual or emotional meaning,
such as to involve as little as possible the more cognitive stages of visual processing.
Our finding that attention may increase the amount of competition between early
visual filters is accompanied by an enhancement of the gain and sensitivity of the
filters, and by a sharpening of their tuning properties. The existence of two such
processing states - one, more sensitive and selective inside the focus of attention,
and the other, more broadly-tuned and non-specific outside - can be justified by
at least two observations: First, the higher level of activity in attended neurons
consumes more energy, which may not be desirable over the entire extent of visual
cortices. Second, although less efficient for fine discriminations, the broadly-tuned
and non-specific state may have greater ability at catching unexpected, non-specific
visual events. In this perspective, this state would be desirable as an input to
bottom-up, visual alerting mechanisms, which monitor the rest of our visual world
while we are focusing on a specific task requiring high focal accuracy.
Acknowledgements
This research was supported by ONR and NSF (Caltech ERG).
References
[1] Bonnel AM, Stein JF, Bertucci P. Q J Exp Psychol fA} 1992;44(4):601-26
[2] Gandhi SP, Heeger DJ, Boynton GM. Inv Opht Vis Sci (ARVO'98) 1998;39(4):5194
[3] Heeger DJ. Vis Neurosci 1992;9:181-97
[4] Itti L, Braun J, Lee DK, Koch C. Proc NIPS *97 {in press)
[5] Itti L, Koch C, Braun J. Inv Opht Vis Sci (Proc ARVO'98) 1998;39(4):2934
[6] Koch C, Ullman S. Hum NeurobioI1985;4:219-27
[7] Lee DK, Koch C, Braun J. Vis Res 1997:37(17):2409-18
[8] Lee DK, Koch C, Itti L, Braun J . Inv Opht Vis Sci (Proc ARVO'98) 1998;39(4):2938
[9] Luck SJ, Chelazzi L, Hillyard SA, Desimone R. J Neurophysio/1997;77{I):24-42
[10] Maunsell JH. Science 1995;270(5237)764-9
[11] Motter BC. J NeurophysioiI993;70(3):909-19
[12] Nakayama K, Mackeben M. Vis Res 1989;29(11):1631-47
[13] Pouget A, Zhang K, Deneve S, Latham PE. Neur Comp 1998;10:373-401
[14] Somers DC, et al. lnv Opht Vis Sci (Proc ARVO'98) 1998;39(4):5192
[15] Sperling G, Melchner MJ. Science 1978;202:315-8
[16] Spitzer H, Desimone R, Moran J. Science 1988;240(4850):338-40
[17] Treue S, Maunsell JH. Nature 1996;382(6591):539-41
| 1547 |@word trial:2 grey:1 simulation:1 attended:18 tuned:6 bc:1 interestingly:1 tilted:1 mst:1 happen:1 discrimination:17 alone:1 half:1 provides:1 location:6 successive:1 sigmoidal:2 zhang:1 five:1 transducer:1 consists:2 fitting:7 overhead:1 behavioral:2 inside:1 manner:2 mask:4 indeed:1 brain:2 compensating:1 inspired:1 automatically:1 little:2 pf:1 project:1 retinotopic:1 linearity:1 underlying:1 circuit:1 spitzer:1 what:2 cm:1 weibull:1 unified:1 finding:2 sharpening:3 quantitative:7 w8:1 multidimensional:1 nf:2 braun:8 control:2 unit:7 maunsell:2 christof:1 positive:1 before:1 laurent:1 firing:1 modulation:12 might:4 plus:1 examined:1 range:4 statistically:3 fwhm:4 directed:1 internals:1 block:1 procedure:4 area:6 thought:1 gabor:1 physiology:1 onto:1 close:3 put:1 unattended:1 demonstrated:1 center:2 attention:37 starting:1 independently:1 focused:1 resolution:1 pouget:1 estimator:1 rule:1 erg:1 population:2 variation:1 increment:2 hierarchy:1 spontaneous:1 enhanced:1 massive:1 engage:1 gandhi:1 gm:1 hypothesis:4 cut:1 predicts:1 native:1 observed:7 bottom:2 cycle:1 e8:1 luck:1 consumes:1 voluntarily:1 discriminates:4 substantial:1 asked:2 battery:1 exps:1 dynamic:1 trained:1 basis:1 easily:1 cat:1 distinct:1 outcome:1 outside:1 larger:1 plausible:1 arvo:4 otherwise:1 ability:1 statistic:1 noisy:3 reproduced:1 seemingly:1 strengthening:1 poorly:9 pronounced:1 competition:4 convergence:1 enhancement:4 eccentricity:1 optimum:1 supra:1 produce:1 wider:1 measured:8 received:1 progress:1 sa:1 grating:1 involves:1 quantify:1 closely:1 correct:1 attribute:1 filter:12 stochastic:2 modifying:1 centered:1 human:11 viewing:1 elevation:1 adjusted:1 koch:10 around:3 exp:4 vary:1 early:11 estimation:2 proc:4 intermediary:1 currently:1 sensitive:1 concurrent:2 gaussian:3 always:1 aim:1 rather:2 varying:1 broader:1 treue:1 focus:3 improvement:2 likelihood:1 superimposed:1 indicates:1 contrast:18 am:1 dependent:1 attract:1 entire:3 typically:1 achim:1 pasadena:1 manipulating:1 reproduce:1 selective:1 issue:2 among:1 orientation:19 dual:3 overall:1 exponent:7 animal:2 spatial:6 constrained:2 psychophysics:3 singlescale:1 field:4 equal:1 identical:1 broad:1 alter:1 jochen:1 fmri:2 simplex:1 others:1 stimulus:12 quantitatively:2 cpd:4 np:1 oriented:1 melchner:1 composed:1 simultaneously:2 resulted:1 individual:1 phase:1 freedom:1 detection:1 interest:1 investigate:4 possibility:1 accurate:1 desimone:2 re:2 plotted:1 catching:1 increased:2 modeling:3 earlier:1 deviation:2 subset:1 conducted:1 reported:1 calibrated:1 devoid:1 sensitivity:2 discriminating:1 lee:7 v4:1 off:1 systematic:1 pool:3 central:6 sharpened:1 successively:1 bonnel:1 cognitive:1 derivative:1 itti:5 ullman:1 fixating:1 account:5 accompanied:1 bold:1 pooled:4 vi:11 performed:1 root:1 observer:6 later:4 apparently:1 reached:2 masking:2 minimize:2 square:1 accuracy:2 variance:2 characteristic:2 efficiently:1 ofthe:1 yes:1 none:2 comp:1 detector:2 simultaneous:1 reach:1 sixth:1 energy:1 frequency:1 pp:1 associated:1 gain:7 cueing:1 dataset:2 improves:1 electrophysiological:1 amplitude:1 back:1 focusing:3 higher:6 response:10 formulation:1 strongly:1 stage:24 msc:1 receives:2 multiscale:1 quality:1 gray:1 behaved:1 name:1 effect:9 staircase:1 unbiased:1 requiring:1 hence:2 laboratory:1 during:2 self:1 width:7 noted:1 excitation:1 octave:1 latham:1 motion:1 fj:3 meaning:1 recently:2 mt:1 winner:1 slight:1 interpret:2 measurement:1 significant:2 automatic:2 tuning:12 focal:4 similarly:2 dj:2 had:2 hillyard:1 cortex:2 inhibition:3 perspective:1 certain:5 onr:1 caltech:2 captured:1 analyzes:1 greater:1 paradigm:3 period:4 redundant:1 ii:2 relates:1 full:7 multiple:1 desirable:2 characterized:2 compensate:1 impact:1 prediction:3 involving:1 basic:2 vision:1 poisson:1 volunteer:1 cell:1 justified:1 addition:1 background:5 fine:1 annealing:1 pyramidal:1 eliminates:1 rest:1 probably:1 subject:3 dipper:1 virtually:1 pooling:10 induced:1 near:1 easy:1 peformance:1 variety:1 affect:1 fit:12 absent:1 whether:2 dramatically:2 modulatory:2 detailed:1 involve:1 amount:2 stein:1 ten:1 reduced:1 nsf:1 inhibitory:3 per:2 broadly:2 motter:1 recomputed:1 four:3 threshold:17 monitor:1 drawn:1 clarity:1 luminance:3 deneve:1 merely:1 angle:1 letter:2 inverse:1 uncertainty:1 somers:1 almost:1 putative:1 decision:10 pay:1 display:1 chelazzi:1 yielded:2 activity:9 sharply:1 awake:1 expo:4 scene:1 speed:1 neur:1 peripheral:3 poor:5 slightly:1 primate:1 happens:2 modification:3 explained:1 restricted:1 anatomically:1 taken:1 previously:1 remains:1 sperling:1 mechanism:6 needed:1 available:1 multiplied:1 away:1 v2:1 distinguished:1 existence:1 top:2 include:1 emotional:1 build:1 classical:3 psychophysical:12 added:1 hum:1 receptive:1 strategy:2 primary:1 striate:1 fa:1 devote:1 distance:2 attentional:10 simulated:1 sci:4 evenly:1 extent:2 spanning:1 modeled:1 difficult:1 steep:1 relate:1 perform:3 vertical:3 observation:3 neuron:3 discarded:1 implementable:1 situation:1 dc:1 interacting:2 varied:5 inv:3 introduced:1 pair:1 alerting:1 california:1 narrow:1 pop:1 hour:1 macaque:1 nip:1 address:1 bar:3 below:2 pattern:8 appeared:1 including:1 power:2 event:1 technology:1 imply:1 psychol:1 acknowledgement:1 law:1 fully:10 mixed:1 prototypical:1 degree:2 controversial:1 sufficient:1 boynton:1 bank:1 systematically:1 cd:3 excitatory:1 supported:2 free:1 formal:1 bias:1 understand:1 jh:2 institute:1 fall:1 emerge:1 feedback:1 dimension:1 cortical:3 curve:3 world:1 dale:1 instructed:1 sj:1 ignore:1 deg:1 reproduces:1 conceptual:1 assumed:1 masker:2 search:1 physiologically:1 table:6 mj:1 nature:1 ca:1 nakayama:1 interact:1 investigated:2 complex:2 sp:1 linearly:3 neurosci:1 motivation:1 noise:6 allowed:2 quadrature:1 neuronal:2 psychometric:1 en:1 screen:3 position:1 downhill:1 heeger:3 crude:1 pe:1 weighting:1 down:2 specific:6 gating:1 r8:2 symbol:1 dk:3 physiological:2 moran:1 exists:1 modulates:1 illumination:1 electrophysiology:1 simply:1 likely:1 appearance:1 visual:34 unexpected:1 saturating:1 oct:2 modulate:3 presentation:2 narrower:1 consequently:2 room:1 jf:1 fisher:1 absence:1 change:4 specifically:1 except:4 determined:2 degradation:1 engaged:1 divisive:1 experimental:1 attempted:1 selectively:1 formally:1 internal:4 support:1 paralleled:2 dissimilar:1 tested:1 phenomenon:1 |
599 | 1,548 | Familiarity Discrimination of Radar
Pulses
Eric Grangerl, Stephen Grossberg 2
Mark A. RUbin2 , William W. Streilein 2
1 Department
of Electrical and Computer Engineering
Ecole Polytechnique de Montreal
Montreal, Qc. H3C 3A 7 CAN ADA
2Department of Cognitive and Neural Systems, Boston University
Boston, MA 02215 USA
Abstract
The ARTMAP-FD neural network performs both identification
(placing test patterns in classes encountered during training) and
familiarity discrimination (judging whether a test pattern belongs
to any of the classes encountered during training). The performance of ARTMAP-FD is tested on radar pulse data obtained in
the field, and compared to that of the nearest-neighbor-based NEN
algorithm and to a k > 1 extension of NEN.
1
Introduction
The recognition process involves both identification and familiarity discrimination.
Consider, for example, a neural network designed to identify aircraft based on their
radar reflections and trained on sample reflections from ten types of aircraft A . . . J.
After training, the network should correctly classify radar reflections belonging to
the familiar classes A . .. J, but it should also abstain from making a meaningless
guess when presented with a radar reflection from an object belonging to a different,
unfamiliar class. Familiarity discrimination is also referred to as "novelty detection,"
a "reject option," and "recognition in partially exposed environments."
ARTMAP-FD, an extension of fuzzy ARTMAP that performs familiarity discrimination, has shown its effectiveness on datasets consisting of simulated radar range
profiles from aircraft targets [1, 2]. In the present paper we examine the performance of ARTMAP-FD on radar pulse data obtained in the field , and compare it
876
E. Granger, S. Grossberg, M. A. Rubin and W. W. Streilein
to that of NEN, a nearest-neighbor-based familiarity discrimination algorithm, and
to a k > 1 extension of NEN.
2
Fuzzy ARTMAP
Fuzzy ARTMAP [3] is a self-organizing neural network for learning, recognition,
and prediction. Each input a learns to predict an output class K. During training,
the network creates internal recognition categories, with the number of categories
determined on-line by predictive success. Components of the vector a are scaled
so that each ai E [0,1] (i = 1 ... M). Complement coding [4] doubles the number
of components in the input vector, which becomes A
(a, a C ) , where the ith
component of a C is ai (I-ad. With fast learning, the weight vector w) records the
largest and smallest component values of input vectors placed in the /h category.
The 2M-dimensional vector Wj may be visualized as the hyperbox R j that just
encloses all the vectors a that selected category j during training.
=
=
Activation of the coding field F2 is determined by the Weber law choice function
Tj(A) =1 A 1\ Wj 1 /(0:+ 1 Wj I), where (P 1\ Q)i = min(Pi , Qj) and 1 P 1=
L;~ 1 Pi I? With winner-take-all coding, the F2 node J that receives the largest
Fl -+ F2 input T j becomes active. Node J remains active if it satisfies the matching
criterion: 1Al\wj 1/ 1A 1= 1Al\wj 1/M > p, where p E [0,1] is the dimensionless
vigilance parameter. Otherwise, the network resets the active F2 node and searches
until J satisfies the matching criterion. If node J then makes an incorrect class
prediction, a match tracking signal raises vigilance just enough to induce a search,
which continues until either some F2 node becomes active for the first time, in
which case J learns the correct output class label k( J) = K; or a node J that has
previously learned to predict K becomes active. During testing, a pattern a that
activates node J is predicted to belong to the class K = k( J).
3
ARTMAP-FD
Familiarity measure. During testing, an input pattern a is defined as familiar
when a familiarity function ?(A) is greater than a decision threshold T Once a
category choice has been made by the winner-take-all rule, fuzzy ARTMAP ignores
the size of the input TJ. In contrast, ARTMAP-FD uses TJ to define familiarity,
taking
?(A) = TJ(A) = 1 A 1\ WJ 1
(1)
TjlAX
1 WJ 1
'
where TjlAX =1 WJ 1/(0:+ 1WJ I)? This maximal value of T J is attained by each
input a that lies in the hyperbox RJ, since 1 A 1\ W J 1= 1 W J 1 for these points.
An input that chooses category J during testing is then assigned the maximum
familiarity value 1 if and only if a lies within RJ.
Familiarity discrimination algorithm. ARTMAP-FD is identical to fuzzy
ARTMAP during training. During testing, ?(A) is computed after fuzzy ARTMAP
has yielded a winning node J and a predicted class K = k(J). If ?(A) > I,
ARTMAP-FD predicts class K for the input a. If ?(A) ::; I, a is regarded as
belonging to an unfamiliar class and the network makes no prediction.
Note that fuzzy ARTMAP can also abstain from classification, when the baseline
vigilance parameter 15 is greater than zero during testing. Typically 15 = during
training, to maximize code compression. In radar range profile simulations such
?
Familiarity Discrimination of Radar Pulses
877
as those described below, fuzzy ARTMAP can perform familiarity discrimination
when p > 0 during both training and testing. However, accurate discrimination
requires that p be close to 1, which causes category proliferation during training.
Range profile simulations have also set p = 0 during both training and testing, but
with the familiarity measure set equal to the fuzzy ARTMAP match function:
(2)
This approach is essentially equivalent to taking p = 0 during training and p > 0
during testing, with p
However, for a test set input a E RJ, the function
defined by (2) sets ?(A) =1 w J 1/ M, which may be large or small although a is
familiar. Thus this function does not provide as good familiarity discrimination as
the one defined by (1), which always sets ?(A) = 1 when a E RJ. Except as noted ,
all the simulations below employ the function (1), with p = O.
=,.
Sequential evidence accumulation. ART-EMAP (Stage 3) [5] identifies a test
set object's class after exposure to a sequence of input patterns, such as differing
views, all identified with that one object. Training is identical to that of fuzzy
ART MAP, with winner-take-all coding at F2 . ART-EMAP generally employs distributed F2 coding during testing. With winner-take-all coding during testing as
well as training, ART-EMAP predicts the object's class to be the one selected by the
largest number of inputs in the sequence. Extending this approach, ARTMAP-FD
accumulates familiarity measures for each predicted class K as the test set sequence
is presented. Once the winning class is determined, the object's familiarity is defined as the average accumulated familiarity measure of the predicted class during
the test sequence.
4
Familiarity discrimination simulations
Since familiarity discrimination involves placing an input into one of two sets, familiar and unfamiliar, the receiver operating characteristic (ROC) formalism can
be used to evaluate the effectiveness of ARTMAP-FD on this task. The hit rate
H is the fraction of familiar targets the network correctly identifies as familiar and
the false alarm rate F is the fraction of unfamiliar targets the network incorrectly
identifies as familiar. An ROC curve is a plot of H vs. F, parameterized by the
threshold'Y (i.e., it is equivalent to the two curves Fh) and Hh)) . The area under
the ROC curve is the c-index, a measure of predictive accuracy that is independent
of both the fraction of positive (familiar) cases in the test set and the positive-case
decision threshold 'Y.
An ARTMAP-FD network was trained on simulated radar range profiles from 18
targets out of a 36-target set (Fig. la). Simulations tested sequential evidence
accumulation performance for 1, 3, and 100 observations, corresponding to 0.05,
0.15, and 5.0 sec. of observation (smooth curves, Fig. Ib) . As in the case of
identification [6], a combination of multiwavelength range profiles and sequential
evidence accumulation produces good familiarity discrimination, with the c-index
approaching 1 as the number of sequential observations grows.
Fig. Ib also demonstrates the importance of the proper choice of familiarity measure. The jagged ROC curve was produced by a familiarity discrimination simulation identical to that which resulted in the IOO-sequential-view smooth curve, but
using the match function (2) instead of ? as given by (1).
E. Granger, S. Grossberg, M A. Rubin and W. W. Streilein
878
IO , -
- - - - ----r
I
'F
~_~~~II
?""'\"MA
'-"- ..
o
o
0 .2
0.4
0.6
F
(b)
08
'Y
T.
(c)
Figure l:(a) 36 simulation targets with 6 wing positions and 6 wing lengths, and 100
scattering centers per target. Boxes indicate randomly selected familiar targets. (b) ROC
curves from ARTMAP-FD simulations, with multiwavelength range profiles having 40
center frequencies. Sequential evidence accumulation for 1, 3 and 100 views uses familiarity
measure (1) (smooth curves); and for 100 views uses the match function (2) (jagged curve).
(c) Training and test curves of miss rate M = (1- H) and false alarm rate F vs threshold
1', for 36 targets and one view, Training curves intersect at the point where "y = r p
(predicted); and test curves intersect near the point where l' = ra (optimal). The training
curves are based on data from the first training epoch, the test curves on data from 3
training epochs.
5
Familiarity threshold selection
When a system is placed in operation, one particular decision threshold 'Y = r must
be chosen . In a given application, selection of r depends upon the relative cost
of errors due to missed targets and false alarms. The optimal r corresponds to a
point on the parameterized ROC curve that is typically close to the upper left-hand
corner of the unit square, to maximize correct selection of familiar targets (H) while
minimizing incorrect selection of unfamiliar tar gets (F) .
Validation set method. To determine a predicted threshold r p , the training
data is partitioned into a training subset and a validation subset. The network
is trained on the training subset, and an ROC curve (F(r) , H(r)) is calculated
for the validation subset. r p is then taken to be the point on the curve that
maximizes [H(r) - F(r)]. (For ease of computation the symmetry point on the
curve, where 1 - H('y) = F(r), can yield a good approximation.) For a familiarity
discrimination task the validation set must include examples of classes not present
in the training set. Once rp is determined , the training subset and validation subset
should be recombined and the network retrained on the complete training set. The
retrained network and the predicted threshold r p are then employed for familiarity
discrimination on the test set.
On-line threshold determination. During ARTMAP-FD training, category
nodes compete for new patterns as they are presented. When a node J wins the
competition, learning expands the category hyperbox RJ enough to enclose the
training pattern a. The familiarity measure ? for each training set input then becomes equal to 1. However, before this learning takes place, ? can be less than 1,
and the degree to which this initial value of ? is less than 1 reflects the distance
from the training pattern to RJ. An event of this type- a training pattern successfully coded by a category node-is taken to be representative of familiar test-set
patterns. The corresponding initial values of ? are thus used to generate a training
Familiarity Discrimination of Radar Pulses
879
hit rate curve, where H("() equals the fraction of training inputs with cp > ,.
What about false alarms? By definition, all patterns presented during training are
familiar. However, a reset event during training (Sec. 2) resembles the arrival of
an unfamiliar pattern during testing. Recall that a reset occurs when a category
node that predicts class K wins the competition for a pattern that actually belongs
to a different class k. The corresponding values of cp for these events can thus be
used to generate a training false-alarm rate curve, where F("() equals the fraction
of match-tracking inputs with initial cp > "(.
Predictive accuracy is improved by use of a reduced set of cp values in the trainingset ROC curve construction process. Namely, training patterns that fall inside
RJ, where cp = I, are not used because these exemplars tend to distort the miss
rate curve. In addition, the first incorrect response to a training input is the best
predictor of the network's response to an unfamiliar testing input, since sequential
search will not be available during testing. Finally, giving more weight to events
occurring later in the training process improves accuracy. This can be accomplished
by first computing training curves H("() and F("() and a preliminary predicted
threshold r p using the reduced training set; then recomputing the curves and r p
from data presented only after the system has activated the final category node
of the training process (Fig. Ic). The final predicted threshold r p averages these
values. This calculation can still be made on-line, by taking the "final" node to be
the last one activated.
Table I shows that applying on-line threshold determination to simulated radar
range profile data gives good predictions for the actual hit and false alarm rates, H A
and FA. Furthermore, the HA and FA so obtained are close to optimal, particularly
when the ROC curve has a c-index close to one. The method is effective even when
testing involves sequential evidence accumulation, despite the fact that the training
curves use only single views of each target.
6
NEN
Near-enough-neighbor (NEN) [7, 8] is a familiarity discrimination algorithm based
on the single nearest neighbor classifier. For each familiar class K, the familiarity
threshold t:l.K is the largest distance between any training pattern of class K and
its nearest neighbor also of class K. During testing, a test pattern is declared
unfamiliar if the distance to its nearest neighbor is greater than the threshold t:l.K
corresponding to the class K of that nearest neighbor.
We have extended NEN to k > I by retaining the above definition of the t:l.K's,
while taking the comparison during testing to be between t:l.K and the distance
between the test pattern and the closest of its k nearest neighbors which is of the
class K to which the test pattern is deemed to belong.
7
Radar pulse data
Identifying the type of emitter from which a radar signal was transmitted is an
important task for radar electronic support measures (ESM) systems. Familiarity
discrimination is a key component of this task, particularly as the continual proliferation of new emitters outstrips the ability of emitter libraries to document every
sort of emitter which may be encountered.
The data analyzed here, gathered by Defense Research Establishment Ottawa, con-
880
E. Granger, S. Grossberg, M. A. Rubin and W W Streilein
hit rate
false alarm rate
accuracy
actual
0.81
0.11
0.95
3x3
optimal
0.86
0.14
1.00
actual
0.77
0.24
0.93
6x6
optimal
0.77
0.23
1.00
actual
0.99
0.06
1.00
6x6*
optimal
0.98
0.02
1.00
Table 1: Familiarity discrimination, using ARTMAP-FD with on-line threshold prediction, of simulated radar range profile data. Training on half the target classes (boxed
"aircraft" in Fig. la) , testing on all target classes. (In 3x3 case, 4 classes out of 9 total used for training.) Accuracy equals the fraction of correctly-classified targets out of
familiar targets selected by the network as familiar. The results for the 6x6' dataset involve sequential evidence accumulation, with 100 observations (5 sec.) per test target.
Radar range profile simulations use 40 center frequencies evenly spaced between 18GHz
and 22GHz, and wp x wl simulated targets, where wp =number of wing positions and
wi =number of wing lengths. The number of range bins (2/3 m. per bin) is 60 , so each
pattern vector has (60 range bins) x (40 center frequencies) = 2400 components. Training
patterns are at 21 evenly spaced aspects in a 10? angular range and, for each viewing
angle, at 15 downrange shifts evenly spaced within a single bin width. Testing patterns
are at random aspects and downrange shifts within the angular range and half the total
range profile extent of (60 bins) x (2/3 m.) =40 m.
method
ARTMAP-FD
hit rate
f. a. rate
accuracy
[memory
0.95
0.02
1.00
21
[I
II
NEN
city-block metric
Euclidean metric
k-l k-5 k - 25
k-l k-5 k - 25
0.94
0.94
0.93
0.94
0.93
0.92
0.02
0.02
0.04
0.14
0.05
0.13
1.00
0.99
1.00
1.00
1.00
1.00
446
Table 2: Familiarity discrimination of radar pulse data set, using ARTMAP-FD and NEN
with different metrics and values of k. Figure given for memory is twice number of F2
nodes (due to complement coding) for ARTMAP-FD, number of training patterns for
NEN. Training (single epoch) on first three quarters of data in classes 1-9, testing on other
quarter of data in classes 1-9 and all data in classes 10-12. (Values given are averages
over four cyclic permutations of the the 12 classes.) ARTMAP-FD familiarity threshold
determined by validation-set method with retraining.
sist of radar pulses from 12 ship borne navigation radars [9]. Fifty pulses were
collected from each radar, with the exception of radars #7 (100 pulses) and #8
(200 pulses). The pulses were preprocessed to yield 800 I5-component vectors. with
the components taking values between a and l.
8
Results
From Table 2, ARTMAP-FD is seen to perform effective familiarity discrimination
on the radar pulse data. NEN (k = 1) performs comparatively poorly. Extensions
of NEN to k > 1 perform well. During fielded operation these would incur the
cost of the additional computation required to find the k nearest neighbors of the
current test pattern , as well as the cost of higher memory requirements] relative to
ARTMAP-FD. The combination of low hit rate with low false alarm rate obtained
by NEN on the simulated radar range profile datasets (Table 3) suggests that the
algorithm performs poorly here because it selects a familiarity threshold which is
1The memory requirements of kNN pattern classifiers can be reduced by editing
techniques[8], but how the use of these methods affects performance of kNN-based familiarity discrimination methods is an open question.
881
Familiarity Discrimination ofRadar Pulses
method
I
dataset
hit rate
false alarm rate
accuracy
memory
II__
II
II
ARTMAP -FD
3x3
0.81
0.11
0.95
I
12
I
6x6
0.77
0.24
0.93
88
Ill-rk
-----.-1.,.....,_...-,--,-N_E......N....,..-...--.-_..---r-.----..-i
__
k - 5 k - 99 k - 1 I k - 5
II
3x3
6x6
0.11
0.00
1.00
II
0.11
0.00
1.00
1260
0.11
0.00
1.00
0.14
0.14
0.00
0.00
1.00
1.00
5670
Table 3: Familiarity discrimination of simulated radar range profiles using ARTMAP-FD
and NEN with different values of k. Training and testing as in Table 1. ARTMAP-FD
familiarity threshold determined by on-line method. City-block metric used with NEN;
results with Euclidean metric were slighlty poorer.
too high. ARTMAP-FD on-line threshold selection, on the other hand, yields a
value for the familiarity threshold which balances the desiderata of high hit rate
and low false alarm rate.
This research was supported in part by grants from the Office of Naval Research, ONR
NOOOI4-95-1-0657 (S . G.) and ONR NOOOI4-96-1-0659 (M. A. R ., W. W. S.) , and by a grant
from the Defense Advanced Research Projects Agency and the Office of Naval Research,
ONR NOOOI4-95-1-0409 (S. G. , M. A. R. , W W. S.). E. G. was supported in part by
the Defense Research Establishment Ottawa and the Natural Sciences and Engineering
Research Council of Canada.
References
[1] Carpenter, G. A., Rubin, M. A. , & Streilein, W . W ., ARTMAP-FD: Familiarity
discrimination applied to radar target recognition , in ICNN'97: Proceedings of the
IEEE International Conference on Neural N etworks, Houston, June 1997;
[2] Carpenter, G. A., Rubin, M. A., & Streilein, W. W ., Threshold Determination for
ARTMAP-FD Familiarity Discrimination, in C . H. Dagli et al., eds., Intelligent Engineering Systems Through Artificial Neural Networks, 1, 23-28, ASME, New York,
1997.
[3] Carpenter, G . A., Grossberg, S. , Markuzon, N., Reynolds, J . H., & Rosen, D. E .,
Fuzzy ARTMAP: A neural network architecture for incremental supervised learning
of analog multidimensional maps, IEEE Transactions on N eural Networks, 3, 698-713,
1992.
[4] Carpenter, G. A., Grossberg, S., & Rosen . D . B. , Fuzzy ART: Fast stable learning and
categorization of analog patterns by an adaptive resonance system, Neural Networks,
4,759-771, 1991.
[5] Carpenter, G. A., & Ross, W . D. , ART-EMAP : A neural network architecture for
object recognition by evidence accumulation , IEEE Transactions on Neural Networks,
6, 805-818, 1995.
[6] Rubin, M. A., Application of fuzzy ARTMAP and ART-EMAP to automatic target
recognition using radar range profiles, Neural Networks , 8, 1109-1116, 1995.
[7] Dasarathy, E. V.,.Is your nearest neighbor near enough a neighbor?, in Lainious, D. G.
and Tzannes, N. S., eds. Applications and Research in Informations Systems and
Sciences, 1, 114-117, Hemisphere Publishing Corp. , Washington, 1977.
[8] Dasarathy, B. V., ed., Nearest Neighbor(NN) Norm : NN Pattern Classification Techniques, IEEE Computer Society Press, Los Alamitos, CA, 1991.
[9] Granger, E. , Savaria, Y, Lavoie, P., & Cantin, M.-A ., A comparison of self-organizing
neural networks for fast clustering of radar pulses, Signal Processing , 64, 249-269,
1998.
| 1548 |@word aircraft:4 compression:1 norm:1 retraining:1 open:1 pulse:15 simulation:9 initial:3 cyclic:1 ecole:1 document:1 reynolds:1 current:1 activation:1 must:2 designed:1 plot:1 discrimination:28 v:2 half:2 selected:4 guess:1 ith:1 record:1 node:15 incorrect:3 inside:1 ra:1 proliferation:2 examine:1 actual:4 becomes:5 project:1 maximizes:1 what:1 sist:1 fuzzy:13 differing:1 every:1 multidimensional:1 expands:1 continual:1 scaled:1 hit:8 demonstrates:1 classifier:2 unit:1 grant:2 positive:2 before:1 engineering:3 io:1 despite:1 accumulates:1 dasarathy:2 twice:1 resembles:1 suggests:1 ease:1 emitter:4 range:17 grossberg:6 testing:20 block:2 x3:4 area:1 intersect:2 reject:1 matching:2 induce:1 get:1 close:4 encloses:1 selection:5 dimensionless:1 applying:1 accumulation:7 equivalent:2 map:2 center:4 exposure:1 artmap:37 qc:1 identifying:1 rule:1 regarded:1 target:20 construction:1 us:3 recognition:7 particularly:2 continues:1 predicts:3 electrical:1 wj:9 environment:1 agency:1 radar:28 trained:3 raise:1 exposed:1 predictive:3 incur:1 creates:1 upon:1 eric:1 f2:8 fast:3 effective:2 artificial:1 otherwise:1 ability:1 knn:2 h3c:1 final:3 sequence:4 maximal:1 reset:3 organizing:2 poorly:2 competition:2 los:1 double:1 requirement:2 extending:1 produce:1 categorization:1 incremental:1 object:6 montreal:2 exemplar:1 nearest:10 predicted:9 involves:3 indicate:1 enclose:1 slighlty:1 correct:2 viewing:1 bin:5 icnn:1 preliminary:1 recombined:1 extension:4 ic:1 predict:2 smallest:1 fh:1 label:1 ross:1 council:1 largest:4 wl:1 successfully:1 city:2 reflects:1 activates:1 always:1 establishment:2 tar:1 office:2 june:1 naval:2 contrast:1 baseline:1 accumulated:1 nn:2 typically:2 selects:1 classification:2 ill:1 retaining:1 resonance:1 art:7 field:3 once:3 equal:5 having:1 washington:1 identical:3 placing:2 rosen:2 intelligent:1 employ:2 randomly:1 resulted:1 familiar:15 consisting:1 william:1 detection:1 fd:26 analyzed:1 navigation:1 activated:2 tj:4 accurate:1 poorer:1 euclidean:2 recomputing:1 classify:1 formalism:1 ada:1 cost:3 ottawa:2 subset:6 predictor:1 too:1 chooses:1 international:1 outstrips:1 vigilance:3 borne:1 cognitive:1 corner:1 wing:4 de:1 coding:7 sec:3 jagged:2 ad:1 depends:1 later:1 view:6 sort:1 option:1 square:1 accuracy:7 characteristic:1 yield:3 identify:1 gathered:1 spaced:3 identification:3 produced:1 classified:1 ed:3 definition:2 distort:1 frequency:3 con:1 dataset:2 noooi4:3 recall:1 improves:1 actually:1 scattering:1 attained:1 higher:1 supervised:1 x6:5 response:2 improved:1 editing:1 box:1 cantin:1 furthermore:1 just:2 stage:1 angular:2 until:2 hand:2 receives:1 grows:1 usa:1 assigned:1 wp:2 during:27 self:2 width:1 noted:1 criterion:2 asme:1 complete:1 polytechnique:1 performs:4 cp:5 reflection:4 weber:1 abstain:2 quarter:2 winner:4 belong:2 analog:2 unfamiliar:8 ai:2 automatic:1 stable:1 operating:1 closest:1 belongs:2 hemisphere:1 ship:1 corp:1 onr:3 success:1 accomplished:1 transmitted:1 seen:1 greater:3 additional:1 houston:1 employed:1 novelty:1 maximize:2 determine:1 signal:3 stephen:1 ii:6 rj:7 smooth:3 match:5 determination:3 calculation:1 coded:1 prediction:5 desideratum:1 essentially:1 metric:5 addition:1 fifty:1 meaningless:1 tend:1 effectiveness:2 near:3 enough:4 trainingset:1 affect:1 architecture:2 identified:1 approaching:1 shift:2 qj:1 whether:1 defense:3 york:1 cause:1 generally:1 involve:1 ten:1 visualized:1 category:12 reduced:3 generate:2 judging:1 correctly:3 per:3 key:1 four:1 threshold:21 preprocessed:1 lavoie:1 fraction:6 compete:1 angle:1 parameterized:2 i5:1 place:1 electronic:1 missed:1 decision:3 fl:1 encountered:3 yielded:1 your:1 declared:1 aspect:2 min:1 department:2 combination:2 belonging:3 partitioned:1 wi:1 making:1 taken:2 remains:1 previously:1 etworks:1 granger:4 hh:1 available:1 operation:2 nen:15 esm:1 rp:1 clustering:1 include:1 publishing:1 giving:1 society:1 comparatively:1 question:1 alamitos:1 occurs:1 fa:2 win:2 distance:4 simulated:7 evenly:3 extent:1 collected:1 code:1 length:2 index:3 minimizing:1 balance:1 proper:1 perform:3 upper:1 observation:4 datasets:2 incorrectly:1 extended:1 retrained:2 canada:1 complement:2 namely:1 required:1 fielded:1 learned:1 below:2 pattern:26 memory:5 event:4 natural:1 advanced:1 library:1 identifies:3 deemed:1 epoch:3 relative:2 law:1 permutation:1 validation:6 degree:1 rubin:6 pi:2 placed:2 last:1 supported:2 neighbor:12 fall:1 taking:5 distributed:1 ghz:2 curve:26 calculated:1 emap:5 ignores:1 made:2 adaptive:1 transaction:2 active:5 receiver:1 search:3 table:7 ca:1 symmetry:1 ioo:1 boxed:1 alarm:10 profile:13 arrival:1 carpenter:5 eural:1 fig:5 referred:1 representative:1 roc:9 position:2 winning:2 lie:2 ib:2 learns:2 rk:1 familiarity:44 evidence:7 false:10 sequential:9 importance:1 occurring:1 boston:2 tracking:2 partially:1 corresponds:1 satisfies:2 ma:2 determined:6 except:1 miss:2 total:2 la:2 exception:1 internal:1 mark:1 support:1 evaluate:1 tested:2 |